Cache and Reuse: Rethinking the Efficiency of On-device Transfer Learning

Computer Vision and Pattern Recognition(2024)

引用 0|浏览4
摘要
Training only the last few layers in deep neural networks has been considered an effective strategy for enhancing the efficiency of on-device training. Prior work has adopted this approach and focused on accelerating backpropagation. However, by conducting a thorough system-wide analysis, we discover that the primary bottleneck is actually the forward propagation through the frozen layers, rather than backpropagation, if only the last few layers are trained. To address this issue, we introduce the "cache and reuse" idea for on-device transfer learning and propose a two-stage training method, which consists of a cache initialization stage, where we store the output from the frozen layers, followed by a training stage. To make our approach practical, we also propose augmented feature caching and cache compression to address the challenges of non-cacheable feature maps and cache size explosion. We carry out extensive experiments on various models (e.g., convolutional neural network and vision transformers) using real edge devices to demonstrate the effectiveness of our method. As an example, on NVIDIA Jetson Orin NX with MobileNet-V2, our approach boosts the training speed by 6.6 ×, and improves the accuracy by 2.1%. For EfficientNet-b0, our method increases the training speed by 2.2 × and improves its accuracy by 1.3%. Therefore, our approach represents a significant improvement in enabling practical on-device transfer learning for edge devices with limited resources.
更多
查看译文
关键词
Transfer Learning,Neural Network,Convolutional Neural Network,Deep Neural Network,Feature Maps,Training Methods,Two-stage Method,Training Efficiency,Caching,Forward Propagation,Edge Devices,Vision Transformer,Two-stage Training,Cache Size,Source Code,Data Augmentation,ImageNet,Increase In Accuracy,Residual Block,CIFAR-100 Dataset,Memory System,Backward Propagation,Drawing Inspiration,Gain In Accuracy,Two-stage Approach,Augmentation Methods,Reduce Memory Usage,Intermediate Feature Maps,Training Approach
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
0
您的评分 :

暂无评分

数据免责声明
页面数据均来自互联网公开来源、合作出版商和通过AI技术自动分析结果,我们不对页面数据的有效性、准确性、正确性、可靠性、完整性和及时性做出任何承诺和保证。若有疑问,可以通过电子邮件方式联系我们:report@aminer.cn