大语言模型(Large Language Models)
OPT: OPT : Open Pre-trained Transformer Language Models
开放预训练的 Transformer 语言模型
GPT-v1:Improving Language Understanding by Generative Pre-Training
GPT&LLM
GPT-v2:Language Models are Unsupervised Multitask Learners
GPT&LLM
GPT-v3:Language Models are Few-Shot Learners
GPT&LLM
GPT-v4:GPT-4 Technical Report
GPT&LLM
2017-06
Transformers
NeurIPS
2018-06
GPT 1.0
OpenAI
2018-10
BERT
NAACL
2019-02
GPT 2.0
OpenAI
2019-09
Megatron-LM
NVIDIA
2019-10
T5
JMLR
2019-10
ZeRO
Microsoft
SC
2020-01
Scaling Law
OpenAI
2020-05
GPT 3.0
OpenAI
NeurIPS
2021-01
Switch Transformers
JMLR
2021-08
Codex
OpenAI
2021-08
Foundation Models
Stanford
2021-09
FLAN
ICLR
2021-10
T0
HuggingFace et al.
ICLR
2021-12
GLaM
ICML
2021-12
WebGPT
OpenAI
2021-12
Retro
DeepMind
ICML
2021-12
Gopher
DeepMind
2022-01
COT
NeurIPS
2022-01
LaMDA
2022-01
Minerva
NeurIPS
2022-01
Megatron-Turing NLG
Microsoft&NVIDIA
2022-03
InstructGPT
OpenAI
2022-04
PaLM
2022-04
Chinchilla
DeepMind
NeurIPS
2022-05
OPT
Meta
2022-06
Emergent Abilities
TMLR
2022-06
BIG-bench
2022-06
METALM
Microsoft
2022-09
Sparrow
DeepMind
2022-10
Flan-T5/PaLM
2022-10
GLM-130B
Tsinghua
ICLR
2022-11
HELM
Stanford
2022-11
BLOOM
BigScience
2022-11
Galactica
Meta
2022-12
OPT-IML
Meta
2023-01
Flan 2022 Collection
2023-02
LLaMA
Meta
2023-02
Kosmos-1
Microsoft
2023-03
PaLM-E
2023-03
GPT 4
OpenAI
Last updated