자연어 논문리뷰

1.[논문리뷰] Attention is All you need

post-thumbnail

2.Word2Vec

post-thumbnail

3.[논문 리뷰] GPT-1(2018) : Improving Language Understanding by Generative Pre-Training

post-thumbnail

4.[논문 리뷰] GPT-2: Language Models are Unsupervised Multitask Learners

post-thumbnail

5.[논문 리뷰]Deep contextualized word representations(ELMo)

post-thumbnail

6.Why are Sensitive Functions Hard for Transformers?(2024)

post-thumbnail

7.Mission Impossible Language Models(2024)

post-thumbnail

8.Theoretical Limitations of Self-Attention in Neural Sequence Models(2020)

post-thumbnail

9.Quantized Side Tuning: Fast and Memory-Efficient Tuning of Quantized Large Language Models (2024)

post-thumbnail

10.DOLA: DECODING BY CONTRASTING LAYERS IMPROVES FACTUALITY IN LARGE LANGUAGE MODELS(2024)

post-thumbnail

11.Contrastive Decoding: Open-ended Text Generation as Optimization(2023)

post-thumbnail

12.Dense Passage Retrieval for Open-Domain Question Answering(2020)

post-thumbnail

13.MuGI: Enhancing Information Retrieval through Multi-Text Generation Intergration with Large Language Models(2024)

post-thumbnail

14.HiPPO: Recurrent Memory with Optimal Polynomial Projections(2020)

post-thumbnail

15.Combining Recurrent, Convolutional, and Continuous-time Models with Linear State-Space Layers(2021)

post-thumbnail

16.S4: Efficiently Modeling Long Sequences with Structured State Spaces(2022)

post-thumbnail

17.Mamba: Linear-Time Sequence Modeling with Selective State Spaces(2024)

post-thumbnail

18.Byte Latent Transformer(BLT): Patches Scale Better Than Tokens(2024)

post-thumbnail