Approximated Likelihood Ratio: A Forward-Only and Parallel Framework for Boosting Neural Network Training
CoRR(2024)
摘要
Efficient and biologically plausible alternatives to backpropagation in
neural network training remain a challenge due to issues such as high
computational complexity and additional assumptions about neural networks,
which limit scalability to deeper networks. The likelihood ratio method offers
a promising gradient estimation strategy but is constrained by significant
memory consumption, especially when deploying multiple copies of data to reduce
estimation variance. In this paper, we introduce an approximation technique for
the likelihood ratio (LR) method to alleviate computational and memory demands
in gradient estimation. By exploiting the natural parallelism during the
backward pass using LR, we further provide a high-performance training
strategy, which pipelines both the forward and backward pass, to make it more
suitable for the computation on specialized hardware. Extensive experiments
demonstrate the effectiveness of the approximation technique in neural network
training. This work underscores the potential of the likelihood ratio method in
achieving high-performance neural network training, suggesting avenues for
further exploration.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
数据免责声明
页面数据均来自互联网公开来源、合作出版商和通过AI技术自动分析结果,我们不对页面数据的有效性、准确性、正确性、可靠性、完整性和及时性做出任何承诺和保证。若有疑问,可以通过电子邮件方式联系我们:report@aminer.cn