논문 리뷰

1.On-Device Training Under 256KB Memory

post-thumbnail

2.Edge AI: On-Demand Accelerating Deep Neural Network Inference via Edge Computing

post-thumbnail

3.Group Normalization

post-thumbnail

4.TinyTL

post-thumbnail

5.LLM-QAT: Data-Free Quantization Aware Training for Large Language Models

post-thumbnail

6.SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models

post-thumbnail

7.A Survey of Quantization Methods for Efficient Neural Network Inference

post-thumbnail