Score: 0

InstantFT: An FPGA-Based Runtime Subsecond Fine-tuning of CNN Models

Published: June 6, 2025 | arXiv ID: 2506.06505v1

By: Keisuke Sugiura, Hiroki Matsutani

Potential Business Impact:

Makes smart devices learn new things super fast.

Business Areas:
Field-Programmable Gate Array (FPGA) Hardware

Training deep neural networks (DNNs) requires significantly more computation and memory than inference, making runtime adaptation of DNNs challenging on resource-limited IoT platforms. We propose InstantFT, an FPGA-based method for ultra-fast CNN fine-tuning on IoT devices, by optimizing the forward and backward computations in parameter-efficient fine-tuning (PEFT). Experiments on datasets with concept drift demonstrate that InstantFT fine-tunes a pre-trained CNN 17.4x faster than existing Low-Rank Adaptation (LoRA)-based approaches, while achieving comparable accuracy. Our FPGA-based InstantFT reduces the fine-tuning time to just 0.36s and improves energy-efficiency by 16.3x, enabling on-the-fly adaptation of CNNs to non-stationary data distributions.

Country of Origin
🇯🇵 Japan

Page Count
8 pages

Category
Computer Science:
Machine Learning (CS)