논문 리뷰

1.On-Device Training Under 256KB Memory

post-thumbnail

2.Edge AI: On-Demand Accelerating Deep Neural Network Inference via Edge Computing

post-thumbnail

3.Group Normalization

post-thumbnail

4.TinyTL

post-thumbnail

5.LLM-QAT: Data-Free Quantization Aware Training for Large Language Models

post-thumbnail

6.SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models

post-thumbnail

7.A Survey of Quantization Methods for Efficient Neural Network Inference

post-thumbnail

8.BinaryViT: Pushing Binary Vision Transformers Towards Convolutional Models

post-thumbnail

9.SpecEE: Accelerating Large Language Model Inference with Speculative Early Exiting

post-thumbnail

10.Mix-QViT: Mixed-Precision Vision Transformer Quantization Driven by Layer Importance and Quantization Sensitivity

post-thumbnail

11.Explaining NonLinear Classification Decisions with Deep Taylor Decomposition

post-thumbnail

12.MPQ via Learned Layer-wise Importance

post-thumbnail

14.Post-Training Quantization for Vision Transformer

post-thumbnail

15.PTQ4ViT

post-thumbnail

16.FQ-ViT: Post-Training Quantization for Fully Quantized Vision Transformer

post-thumbnail

17.APQ-ViT: Towards Accurate Post-Training Quantization for Vision Transformer

post-thumbnail

18.RepQ-ViT: Scale Reparameterization for Post-Training Quantization of Vision Transformers

post-thumbnail

19.Hierarchical Reasoning Model

post-thumbnail

20.Less is More: Recursive Reasoning with Tiny Networks

post-thumbnail

21.Early-Exit Deep Neural Network - A Comprehensive Survey

post-thumbnail

22.Relaxed Recursive Transformers: Effective Parameter Sharing with Layer-wise LoRA

post-thumbnail

23.Mixture-of-Recursions: Learning Dynamic Recursive Depths for Adaptive Token-Level Computation

post-thumbnail

24.DINOv3

post-thumbnail