Attention Is All You Need[Transformer]
BERT 논문 리뷰
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
GPT 논문 리뷰
NEURAL MACHINE TRANSLATION BY JOINTLY LEARNING TO ALIGN AND TRANSLATE
RoBERTa: A Robustly Optimized BERT Pretraining Approach
Efficient Estimation of Word Representations in Vector Space
Sequence to Sequence Learning with Neural Networks
LLaMA: Open and Efficient Foundation Language Models
EDA: Easy Data Augmentation Techniques for Boosting Performance on Text Classification Tasks
GloVe: Global Vectors for Word Representation