Pretrained Language Models (PLMs) have been achieved great success over many NLP tasks. And with the rapid development of PLMs, there is a need to make a list to show these succesful and remarkable language models.
Group by companies and organizations.
- PLUG (Pre-training for Language Understanding and Generation))
- Alexa Teacher Model
- Alexa Teacher Model: Pretraining and Distilling Multi-Billion-Parameter Encoders for Natural Language Understanding Systems
- AlexaTM 20B
- github repo (model hasn't been publicly released yet, stay tuned!)
- ERNIE family: github repo
- ERNIE 1.0
- ERNIE 2.0
- ERNIE 3.0
- ERNIE 3.0 TiTan
- ERNIE 3.0 Zeus
- ERNIE-GEN
- ERNIE-ViL
- ERNIE-ViLG
- ERNIE-Gram
- ERNIE-M
- ERNIE-Doc
- ERNIE-Tiny
- ERNIE-GeoL
- ERNIE-SPARSE
- ERNIE-Search
- PLATO family: github repo
- PLATO-1
- PLATO-2
- PLATO-XL
- PLATO-KAG
- T0 series
- BLOOM
- Gopher
- Chinchilla
- Retro
- Improving language models by retrieving from trillions of tokens
- github repo (Non-offical implementation)
- GPT-Neo
- GPT-J
- GPT-NeoX
- XGLM
- XLM-R
- M2M-100
- Blender
- Recipes for building an open-domain chatbot
- Internet-Augmented Dialogue Generation
- Beyond Goldfish Memory: Long-Term Open-Domain Conversation
- BlenderBot 3: a deployed conversational agent that continually learns to responsibly engage
- github repo
- BlenderBot A CONVERSATIONAL AI PROTOTYPE
- Hugging Face 🤗
- OPT
- NLLB
- Altas
- PEER
- CPT
- T5 family
- T5 series
- ByT5
- ExT5
- LongT5
- LaMDA
- LaMDA: Language Models for Dialog Applications
- github repo (Non-offical)
- FLAN
- GShard
- Meena
- Towards a Human-like Open-Domain Chatbot
- github repo (Non-offical)
- PaLM
- PaLM: Scaling Language Modeling with Pathways
- github repo (Non-offical)
- UL2
- Minerva
- Flan-PaLM
哈工大讯飞联合实验室: Joint Laboratory of HIT and iFLYTEK Research (HFL).
- PanGu (盘古) family
- PanGu-Alpha
- PanGu-Bot
- PanGu-Coder
- NEZHA (哪吒)
Huawei-Noah offical github repo for Pretrained-Language Model.
- Yuan 1.0
- 封神榜——中文语言预训练模型开源计划
- Erlangshen (二郎神)
- Wenzhong (闻仲)
- Rangdeng (燃灯)
- Yuyuan (余元)
- Bigan (比干)
- Zhouwenwang (周文王)
- Taiyi (太乙)
IDEA refers to 粤港澳大湾区数字经济研究院.
- 织女 Vega v1
- Megatron-Turning NLG family
- DeBERTa series
- DeepNMT
- DeepNet
- GODEL
- METRO
- Megatron-LM
- GPT-family
- GPT, GPT-2, GPT-3
- WebGPT
- InstructGPT
- Codex
- OBERT
- Motian (摩天)
- ShenZhou (神舟)
- ShenNonG (神农)
- HunYuan (混元)
- THUDM
- Wenhui (文汇)
- GLM-130B
THUDM refers to "Data Mining Research Group at Tsinghua University".
-
Tsinghua AI
-
THU-CoAI
THU-CoAI refers to "Conversational AI groups from Tsinghua University".
- BAAI
- Chinese-Transformer-XL
- GLM
🔥Coming soon!🔥
Thanks for all these companies and orgnizations paying a lot of money and efforts to build and train these large models for the benefit of all human beings.
This repo is inspired by awesome pretrained Chinese NLP models.