HyperAIHyperAI

Command Palette

Search for a command to run...

3 months ago

Learning Multi-Step Reasoning by Solving Arithmetic Tasks

Tianduo Wang Wei Lu

Learning Multi-Step Reasoning by Solving Arithmetic Tasks

Abstract

Mathematical reasoning is regarded as a necessary ability for Language Models (LMs). Recent works demonstrate large LMs' impressive performance in solving math problems. The success is attributed to their Chain-of-Thought (CoT) reasoning abilities, i.e., the ability to decompose complex questions into step-by-step reasoning chains, but such ability seems only to emerge from models with abundant parameters. This work investigates how to incorporate relatively small LMs with the capabilities of multi-step reasoning. We propose to inject such abilities by continually pre-training LMs on a synthetic dataset MsAT which is composed of Multi-step Arithmetic Tasks. Our experiments on four math word problem datasets show the effectiveness of the proposed method in enhancing LMs' math reasoning abilities.

Code Repositories

TianduoWang/MsAT
Official
pytorch

Benchmarks

BenchmarkMethodologyMetrics
math-word-problem-solving-on-mawpsMsAT-DeductReasoner
Accuracy (%): 94.3
math-word-problem-solving-on-svampMsAT-DeductReasoner
Execution Accuracy: 48.9

Build AI with AI

From idea to launch — accelerate your AI development with free AI co-coding, out-of-the-box environment and best price of GPUs.

AI Co-coding
Ready-to-use GPUs
Best Pricing
Get Started

Hyper Newsletters

Subscribe to our latest updates
We will deliver the latest updates of the week to your inbox at nine o'clock every Monday morning
Powered by MailChimp
Learning Multi-Step Reasoning by Solving Arithmetic Tasks | Papers | HyperAI