Attention Is All You Need[Transformer]
BERT 논문 리뷰
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension
GPT 논문 리뷰
NEURAL MACHINE TRANSLATION BY JOINTLY LEARNING TO ALIGN AND TRANSLATE
RoBERTa: A Robustly Optimized BERT Pretraining Approach
Efficient Estimation of Word Representations in Vector Space
Sequence to Sequence Learning with Neural Networks
LLaMA: Open and Efficient Foundation Language Models
EDA: Easy Data Augmentation Techniques for Boosting Performance on Text Classification Tasks
GloVe: Global Vectors for Word Representation GloVe: Global Vectors for Word Representation는 2014년에 스탠포드 대학 연구진들이 발표한 논문입니다. 요약 Glove는 co-occurrence(단어 - 단어 공동 등장 빈도)를 활용하여 단어의 의미를 다차원 벡터에 대입을 시키는 방식...
ELECTRA: PRE-TRAINING TEXT ENCODERS AS DISCRIMINATORS RATHER THAN GENERATORS