Orca-Math: Unlocking the Potential of SLMs in Grade School Math

arXiv (Cornell University)(2024)

引用 39|浏览1371
摘要
Mathematical word problem-solving has long been recognized as a complex taskfor small language models (SLMs). A recent study hypothesized that the smallestmodel size, needed to achieve over 80billion parameters. To reach this level of performance with smaller models,researcher often train SLMs to generate Python code or use tools to help avoidcalculation errors. Additionally, they employ ensembling, where outputs of upto 100 model runs are combined to arrive at a more accurate result. Resultselection is done using consensus, majority vote or a separate a verifier modelused in conjunction with the SLM. Ensembling provides a substantial boost inaccuracy but at a significant cost increase with multiple calls to the model(e.g., Phi-GSM uses top-48 to boost the performance from 68.2 to 81.5). In this work, we present Orca-Math, a 7-billion-parameter SLM based on theMistral-7B, which achieves 86.81calls or the use of verifiers, code execution or any other external tools. Ourapproach has the following key elements: (1) A high quality synthetic datasetof 200K math problems created using a multi-agent setup where agentscollaborate to create the data, (2) An iterative learning techniques thatenables the SLM to practice solving problems, receive feedback on its solutionsand learn from preference pairs incorporating the SLM solutions and thefeedback. When trained with Supervised Fine-Tuning alone, Orca-Math achieves81.50achieves 86.81larger models such as LLAMA-2-70B, WizardMath-70B, Gemini-Pro, ChatGPT-3.5. Italso significantly outperforms other smaller models while using much smallerdata (hundreds of thousands vs. millions of problems).
更多
查看译文
关键词
Number Processing,Problem-Solving,Mathematical Literacy,Teaching Strategies,Working Memory
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
0
您的评分 :

暂无评分

数据免责声明
页面数据均来自互联网公开来源、合作出版商和通过AI技术自动分析结果,我们不对页面数据的有效性、准确性、正确性、可靠性、完整性和及时性做出任何承诺和保证。若有疑问,可以通过电子邮件方式联系我们:report@aminer.cn