Repository navigation

#

pretrained-language-model

YAYI 2 是中科闻歌研发的新一代开源大语言模型,采用了超过 2 万亿 Tokens 的高质量、多语言语料进行预训练。(Repo for YaYi 2 Chinese LLMs)

Python
3424
1 年前

An optimized deep prompt tuning strategy comparable to fine-tuning across scales and tasks

Python
2055
2 年前

A plug-and-play library for parameter-efficient-tuning (Delta Tuning)

Python
1032
1 年前

word2vec, sentence2vec, machine reading comprehension, dialog system, text classification, pretrained language model (i.e., XLNet, BERT, ELMo, GPT), sequence labeling, information retrieval, information extraction (i.e., entity, relation and event extraction), knowledge graph, text generation, network embedding

OpenEdge ABL
553
5 年前

Code associated with the Don't Stop Pretraining ACL 2020 paper

Python
533
4 年前

The official repository of the dots.llm1 base and instruct models proposed by rednote-hilab.

455
9 小时前

ACL'2023: DiffusionBERT: Improving Generative Masked Language Models with Diffusion Models

Python
324
2 年前

MWPToolkit is an open-source framework for math word problem(MWP) solvers.

Python
163
3 年前

[NeurIPS 2023] This is the code for the paper `Large Language Model as Attributed Training Data Generator: A Tale of Diversity and Bias`.

Python
153
2 年前

[ACM Computing Surveys 2025] This repository collects awesome survey, resource, and paper for Lifelong Learning with Large Language Models. (Updated Regularly)

143
3 个月前

EMNLP'23 survey: a curation of awesome papers and resources on refreshing large language models (LLMs) without expensive retraining.

134
2 年前

Worth-reading papers and related resources on attention mechanism, Transformer and pretrained language model (PLM) such as BERT. 值得一读的注意力机制、Transformer和预训练语言模型论文与相关资源集合

131
4 年前

The official code for "TEMPO: Prompt-based Generative Pre-trained Transformer for Time Series Forecasting (ICLR 2024)". TEMPO is one of the very first open source Time Series Foundation Models for forecasting task v1.0 version.

Python
114
6 个月前