Language Model: BERT

Ko Hyejung·2021년 12월 3일
0

2021 SKT AI

목록 보기
9/12

Bidirectional Encoder Representation Transformer (BERT)

Pre-trained language representation for language understanding
Large scale architecture
BASE: L=12, A=12, ~110M
LARGE: L=24, A=16, ~340M

Pre-Training: Masked LM & Sentence Prediction

Task Fine-Tuning

GLUE Results

0개의 댓글