Attention Is All You Need[Transformer]
BERT 논문 리뷰
BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation, Translation, and Comprehension