Intelligent Measurement on Edge Devices Using Hardware Memory-Aware Joint Compression Enabled Neural Networks

IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT(2024)

引用 0|浏览9
摘要
Applying neural network (NN) on edge devices is crucial for intelligent measurement. However, the limited memory resource is one of the main challenges for deploying NN with high spatial complexity. To address this issue, a hardware memory-aware joint compression (MAJOR) is proposed in this article. First, a trainable joint compression framework is established for the parallel operation of pruning and quantization. Then, a partial memory-bottleneck penalty regularizer is designed to guide the compression to meet actual hardware memory demands accurately. The proposed method is evaluated on SLOC4, WHU-RS19, RSSCN7, and OPT-Aircraft v1.0 remote sensing image datasets using VGG, ResNet-20/50, MobileNetv1, EfficientNet-B0, DenseNet-121, and ResNeXt-50 for the satellite onboard processing platform, which is a typical hardware memory-limited device. The experimental results show that MAJOR could reduce the model’s partial peak memory size up to $23.19\times $ , ensuring it fits within the target platform’s capacity with less than 3% accuracy loss. Furthermore, the proposed method is evaluated on CIFAR-10 using VGG, ResNet-20, and MobileNetv1 for four edge hardware with different memory sizes. It achieves a remarkable compression ratio of up to 94.92% while ensuring that model accuracy loss remains below 3%. Compared with other compression methods, MAJOR accurately reduces model memory size to match the capacity of the various target hardware. The partial memory bottleneck in NNs could be significantly eliminated.
更多
查看译文
关键词
Edge computing,intelligent measurement,memory constrained hardware,model compression,neural network (NN),pruning,quantization
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
0
您的评分 :

暂无评分

数据免责声明
页面数据均来自互联网公开来源、合作出版商和通过AI技术自动分析结果,我们不对页面数据的有效性、准确性、正确性、可靠性、完整性和及时性做出任何承诺和保证。若有疑问,可以通过电子邮件方式联系我们:report@aminer.cn