GhostNetV3-Small: A Tailored Architecture and Comparative Study of Distillation Strategies for Tiny Images
By: Florian Zager, Hamza A. A. Gardi
Potential Business Impact:
Makes smart phone cameras see better with less power.
Deep neural networks have achieved remarkable success across a range of tasks, however their computational demands often make them unsuitable for deployment on resource-constrained edge devices. This paper explores strategies for compressing and adapting models to enable efficient inference in such environments. We focus on GhostNetV3, a state-of-the-art architecture for mobile applications, and propose GhostNetV3-Small, a modified variant designed to perform better on low-resolution inputs such as those in the CIFAR-10 dataset. In addition to architectural adaptation, we provide a comparative evaluation of knowledge distillation techniques, including traditional knowledge distillation, teacher assistants, and teacher ensembles. Experimental results show that GhostNetV3-Small significantly outperforms the original GhostNetV3 on CIFAR-10, achieving an accuracy of 93.94%. Contrary to expectations, all examined distillation strategies led to reduced accuracy compared to baseline training. These findings indicate that architectural adaptation can be more impactful than distillation in small-scale image classification tasks, highlighting the need for further research on effective model design and advanced distillation techniques for low-resolution domains.
Similar Papers
Comparative Analysis of Lightweight Deep Learning Models for Memory-Constrained Devices
CV and Pattern Recognition
Makes smart computer vision work on small phones.
Hybrid Knowledge Transfer through Attention and Logit Distillation for On-Device Vision Systems in Agricultural IoT
CV and Pattern Recognition
Helps phones find plant sickness faster, cheaper.
CNN and ViT Efficiency Study on Tiny ImageNet and DermaMNIST Datasets
CV and Pattern Recognition
Makes AI see pictures faster and with less power.