이 논문을 읽게 된 이유는 object detector의 knowledge distillation 관련한 연구를 진행하고 있는데 도움이 필요하기 때문...teacher model의 bbox prediction을student model이 유사하게 학습하도록 KL Dive
https://link.springer.com/article/10.1007/s44267-024-00040-3Jia, Z., Sun, S., Liu, G. et al. MSSD: multi-scale self-distillation for object detec
https://arxiv.org/abs/2109.11496Zhang, Peizhen, et al. "LGD: Label-guided self-distillation for object detection." Proceedings of the AAAI confer
https://arxiv.org/abs/2303.05015Deng, Jieren, et al. "Smooth and Stepwise Self-Distillation for Object Detection." 2023 IEEE International Confer
https://papers.neurips.cc/paper_files/paper/2022/file/631ad9ae3174bf4d6c0f6fdca77335a4-Paper-Conference.pdfCao, Weihan, et al. "Pkd: General dist
https://openaccess.thecvf.com/content/CVPR2024/papers/Wang_CrossKD_Cross-Head_Knowledge_Distillation_for_Object_Detection_CVPR_2024_paper.pdfWang
Chen, Xiaokang, et al. "D $^ 3$ ETR: Decoder Distillation for Detection Transformer." arXiv preprint arXiv:2211.09768 (2022).CNN-based detectors에서 다양한
https://openaccess.thecvf.com/content/ICCV2023/papers/ChangDETRDistillAUniversalKnowledgeDistillationFrameworkforDETR-familiesICCV2023_paper.pdf Abst
https://openaccess.thecvf.com/content/CVPR2024/papers/Wang_KD-DETR_Knowledge_Distillation_for_Detection_Transformer_with_Consistent_Distillation_
Beyer, Lucas, et al. "Knowledge distillation: A good teacher is patient and consistent." Proceedings of the IEEE/CVF conference on computer vision and
KD는 model compression을 위한 가장 있는 방법 중 하나이다.key idea는 teacher model(T)의 knowledge를 student model(S)로 전달하는 것이다.그러나 기존 방법들은 S와 T의 학습 능력 간의 큰 차이로 인해 degrad