TY - GEN
T1 - Improve Fine-grained Visual Classification Accuracy by Controllable Location Knowledge Distillation
AU - Tsai, You Lin
AU - Lin, Cheng Hung
AU - Chou, Po Yung
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - The current state-of-the-art network models have achieved remarkable performance. However, they often face an issue of having excessively large architectures, making them challenging to deploy on edge devices. In response to this challenge, a groundbreaking solution known as knowledge distillation has been introduced. The concept of knowledge distillation involves transferring information from a complex teacher model to a simpler student model, effectively reducing the model's complexity. Prior approach has demonstrated promising transfer effects through this technique. Nevertheless, in the field of fine-grained image classification, there has been limited exploration of distillation methods custom-tailored specifically for this domain. In this paper, we focus on knowledge distillation specifically designed for fine-grained image recognition. Notably, this strategy is inspired by the Class Activation Maps (CAM). We first train a complex model and use it generated feature maps with spatial information, which is call hint maps. Furthermore, we propose an adjustment strategy for this hint map, which can control local distribution of information. Referring to it as Controllable Size CAM (CTRLS-CAM). Use it as a guide allowing the student model to effectively learn from the teacher model's behavior and concentrate on discriminate details. In contrast to conventional distillation models, this strategy proves particularly advantageous for fine-grained recognition, enhancing learning outcomes and enabling the student model to achieve superior performance. In CTRLS-CAM method, we refine the hint maps value distribution, redefining the relative relationships between primary and secondary feature areas. We conducted experiments using the CUB200-2011 dataset, and our results demonstrated a significant accuracy improvement of about 5% compared to the original non-distilled student model. Moreover, our approach achieved a 1.23% enhancement over traditional knowledge distillation methods.
AB - The current state-of-the-art network models have achieved remarkable performance. However, they often face an issue of having excessively large architectures, making them challenging to deploy on edge devices. In response to this challenge, a groundbreaking solution known as knowledge distillation has been introduced. The concept of knowledge distillation involves transferring information from a complex teacher model to a simpler student model, effectively reducing the model's complexity. Prior approach has demonstrated promising transfer effects through this technique. Nevertheless, in the field of fine-grained image classification, there has been limited exploration of distillation methods custom-tailored specifically for this domain. In this paper, we focus on knowledge distillation specifically designed for fine-grained image recognition. Notably, this strategy is inspired by the Class Activation Maps (CAM). We first train a complex model and use it generated feature maps with spatial information, which is call hint maps. Furthermore, we propose an adjustment strategy for this hint map, which can control local distribution of information. Referring to it as Controllable Size CAM (CTRLS-CAM). Use it as a guide allowing the student model to effectively learn from the teacher model's behavior and concentrate on discriminate details. In contrast to conventional distillation models, this strategy proves particularly advantageous for fine-grained recognition, enhancing learning outcomes and enabling the student model to achieve superior performance. In CTRLS-CAM method, we refine the hint maps value distribution, redefining the relative relationships between primary and secondary feature areas. We conducted experiments using the CUB200-2011 dataset, and our results demonstrated a significant accuracy improvement of about 5% compared to the original non-distilled student model. Moreover, our approach achieved a 1.23% enhancement over traditional knowledge distillation methods.
KW - CAM
KW - Deep Learning
KW - Fine-grained Visual Classification
KW - Knowledge Distillation
UR - http://www.scopus.com/inward/record.url?scp=85186992329&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85186992329&partnerID=8YFLogxK
U2 - 10.1109/ICCE59016.2024.10444242
DO - 10.1109/ICCE59016.2024.10444242
M3 - Conference contribution
AN - SCOPUS:85186992329
T3 - Digest of Technical Papers - IEEE International Conference on Consumer Electronics
BT - 2024 IEEE International Conference on Consumer Electronics, ICCE 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2024 IEEE International Conference on Consumer Electronics, ICCE 2024
Y2 - 6 January 2024 through 8 January 2024
ER -