Evaluating Mathematical Reasoning of Large Language Models: A Focus on
Error Identification and Correction
Evaluating Mathematical Reasoning of Large Language Models: A Focus on
Error Identification and Correction
The rapid advancement of Large Language Models (LLMs) in the realm of mathematical reasoning necessitates comprehensive evaluations to gauge progress and inspire future directions. Existing assessments predominantly focus on problem-solving from the examinee perspective, overlooking a dual perspective of examiner regarding error identification and correction. From the examiner perspective, we …