language-model

1.[Transformer] Attention Is All You Need

post-thumbnail

2.[GPT] Improving Language Understanding by Generative Pre-Training

post-thumbnail

3.BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

post-thumbnail

4.RoBERTa: A Robustly Optimized BERT Pretraining Approach

post-thumbnail

5.BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension

post-thumbnail

6.[T0] Multitask Prompted Training Enables Zero-Shot Task Generalization

post-thumbnail

7.LLaMA: Open and Efficient Foundation Language Models

post-thumbnail

8.Sparks of Artificial General Intelligence: Early experiments with GPT-4

post-thumbnail

9.OPT: Open Pre-trained Transformer Language Models

post-thumbnail