Fixing Math Phrase Issues


We’ve skilled a system that solves grade faculty math issues with practically twice the accuracy of a fine-tuned GPT-3 mannequin. It solves about 90% as many issues as actual youngsters: a small pattern of 9-12 yr olds scored 60% on a take a look at from our dataset, whereas our system scored 55% on those self same issues. That is essential as a result of as we speak’s AI remains to be fairly weak at commonsense multistep reasoning, which is simple even for grade faculty youngsters. We achieved these outcomes by coaching our mannequin to acknowledge its errors, in order that it may attempt repeatedly till it finds an answer that works.

Learn paperBrowse samplesObtain dataset


Massive language fashions like GPT-3 have many spectacular abilities, together with their potential to mimic many writing types, and their in depth factual data. Nonetheless, they wrestle to carry out duties that require correct multistep reasoning, like fixing grade faculty math phrase issues. Though the mannequin can mimic the cadence of right options, it commonly produces vital errors in logic.

To match human efficiency in advanced logical domains, our fashions should study to acknowledge their errors and to decide on their steps rigorously. To that finish, we prepare verifiers to guage whether or not or not a proposed answer is right. To unravel a brand new drawback, we use verifiers to pick out the very best amongst many proposed options. We collected the brand new GSM8K dataset to guage our strategies, and we’re releasing this dataset to facilitate analysis.

Within the ten examples under, we present options generated by our new technique, verification, and our baseline technique, fine-tuning.

GSM8K Dataset

GSM8K consists of 8.5K top quality grade faculty math phrase issues. Every drawback takes between 2 and eight steps to resolve, and options primarily contain performing a sequence of elementary calculations utilizing fundamental arithmetic operations (+ − × ÷) to achieve the ultimate reply. High quality-tuned state-of-the-art language fashions carry out poorly on this dataset, primarily because of the excessive variety of issues. On the identical time, GSM8K options rely solely on elementary ideas, so attaining excessive take a look at efficiency is a tractable objective.

Options in GSM8K are written as pure language reasonably than as pure math expressions. By sticking to pure language, model-generated options are extra readily interpretable by people, and our strategies stay comparatively area agnostic.

Coaching Verifiers: Fashions that Study from their Errors

One important problem in mathematical reasoning is the excessive sensitivity to particular person errors. Autoregressive fashions, which generate every answer token by token, don’t have any mechanism to right their very own errors. Options that veer off-course rapidly turn out to be unrecoverable, as might be seen within the examples offered.

We tackle this drawback by coaching verifiers to guage the correctness of model-generated options. Verifiers are given many potential options, all written by the mannequin itself, and they’re skilled to resolve which of them, if any, are right.

To unravel a brand new drawback at take a look at time, we generate 100 candidate options after which choose the answer that’s ranked highest by the verifier. Verifiers profit from this inherent optionality, in addition to from the truth that verification is usually a less complicated activity than era.

We discover that we get a powerful enhance in efficiency from verification, so long as the dataset is giant sufficient. With datasets which might be too small, we imagine that the verifiers overfit by memorizing the ultimate solutions within the coaching set, reasonably than studying any extra helpful properties of mathematical reasoning.

On the total coaching set, 6B parameter verification barely outperforms a fine-tuned 175B parameter mannequin, giving a efficiency enhance that’s roughly equal to a 30x mannequin measurement improve. Furthermore, verification seems to scale extra successfully with further knowledge, if we extrapolate primarily based on present outcomes.


Producing right arguments and recognizing incorrect ones are key challenges in creating extra basic AI. Grade faculty math is a perfect testbed for these capabilities. The issues in GSM8K are conceptually easy, but one refined mistake is sufficient to derail a complete answer. Figuring out and avoiding such errors is a vital ability for our fashions to develop. By coaching verifiers, we educate our fashions to separate the great options from those that didn’t fairly work out. We count on these abilities to turn out to be more and more related as we try to use our fashions to extra logically advanced domains.


Leave a Reply

Your email address will not be published. Required fields are marked *