YAYI 2 是中科闻歌研发的新一代开源大语言模型,采用了超过 2 万亿 Tokens 的...
Pretrained language model and its related optimization techniques develo...
Foundation Architecture for (M)LLMs
A curated list of pretrained sentence and word embedding models
An optimized deep prompt tuning strategy comparable to fine-tuning acros...
A plug-and-play library for parameter-efficient-tuning (Delta Tuning)
Code associated with the Don't Stop Pretraining ACL 2020 paper
Live Training for Open-source Big Models
word2vec, sentence2vec, machine reading comprehension, dialog system, te...
ACL'2023: DiffusionBERT: Improving Generative Masked Language Models wit...
MWPToolkit is an open-source framework for math word problem(MWP) solvers.
[NeurIPS 2021] COCO-LM: Correcting and Contrasting Text Sequences for La...
Bamboo-7B Large Language Model
EMNLP'23 survey: a curation of awesome papers and resources on refreshin...
Implementation of ICLR 21 paper: Probing BERT in Hyperbolic Spaces