review

1.[논문 리뷰] DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning

post-thumbnail

2.[논문 리뷰] Reflexion: Language Agents with Verbal Reinforcement Learning

post-thumbnail

3.[논문 리뷰] SWE-Agent: Agent-Computer Interfaces Enable Automated Software Engineering

post-thumbnail

4.[논문 리뷰] DPO: Direct Preference Optimization

post-thumbnail

5.[논문 리뷰] Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection

post-thumbnail

6.[논문 리뷰] LoRA: Low-Rank Adaptation of Large Language Models

post-thumbnail

7.[논문 리뷰] FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness

post-thumbnail

8.[논문 리뷰] AWQ: Activation-Aware Weight Quantization for LLM Compression & Acceleration

post-thumbnail

9.[논문 리뷰] Mamba: Linear-Time Sequence Modeling with Selective State Spaces

post-thumbnail

10.[논문 리뷰] MetaGPT: Meta Programming for A Multi-Agent Collaborative Framework

post-thumbnail

11.[논문 리뷰] Voyager: An Open-Ended Embodied Agent with Large Language Models

post-thumbnail

12.[논문 리뷰] Toolformer: Language Models Can Teach Themselves to Use Tools

post-thumbnail

13.[논문 리뷰] Fast Inference from Transformers via Speculative Decoding

post-thumbnail

14.[논문 리뷰] LLaVA: Visual Instruction Tuning

post-thumbnail

15.[논문 리뷰] DSPy: Compiling Declarative Language Model Calls into Self-Improving Pipelines

post-thumbnail

16.[논문 리뷰] Orca: Progressive Learning from Complex Explanation Traces of GPT-4

post-thumbnail

17.[논문 리뷰] SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot

post-thumbnail

18.[논문 리뷰] BitNet b1.58: The Era of 1-bit LLMs — Every LLM Parameter is Ternary {-1, 0, 1}

post-thumbnail

19.[논문 리뷰] vLLM: Efficient Memory Management for LLM Serving with PagedAttention

post-thumbnail

20.[논문 리뷰] Constitutional AI: Harmlessness from AI Feedback

post-thumbnail

21.[논문 리뷰] Chain-of-Thought Prompting Elicits Reasoning in Large Language Models

post-thumbnail

22.[논문 리뷰] ReAct: Synergizing Reasoning and Acting in Language Models

post-thumbnail

23.[논문 리뷰] Attention Is All You Need

post-thumbnail

24.[논문 리뷰] Training Compute-Optimal Large Language Models (Chinchilla)

post-thumbnail

25.[논문 리뷰] Mixtral of Experts

post-thumbnail

26.[논문 리뷰] QLoRA: Efficient Finetuning of Quantized LLMs

post-thumbnail

27.[논문 리뷰] Corrective Retrieval Augmented Generation (CRAG)

post-thumbnail

28.[논문 리뷰] Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks

post-thumbnail

29.[논문 리뷰] Wanda: A Simple and Effective Pruning Approach for Large Language Models

post-thumbnail

30.[논문 리뷰] InstructGPT: Training Language Models to Follow Instructions with Human Feedback

post-thumbnail