Document detail
ID

oai:arXiv.org:2411.05407

Topic
Computer Science - Computation and...
Author
Mohammadkhani, Mohammad Ghiasvand
Category

Computer Science

Year

2024

listing date

11/13/2024

Keywords
reasoning mathematical language solution
Metrics

Abstract

Despite the strong performance of large language models (LLMs) in tasks like mathematical reasoning, their practical use is limited by high computational demands and proprietary restrictions.

Chain-of-thought (CoT) and program-of-thought (PoT) fine-tuning are common methods to transfer LLM knowledge to small language models (SLMs).

However, CoT often leads to calculation errors in SLMs, while PoT has shown more promise.

While most PoT-based approaches focus on direct problem-to-code conversion or extracting only the key information from questions and then providing code solution for it, this work emphasizes filling the gaps in the question to clearly illustrate the solution path, which can be challenging for an SLM to understand when such information is not explicitly provided.

Therefore, this paper introduces Gap-Filling Prompting (GFP), a novel two-step prompting strategy designed to enhance the problem-solving process for SLMs.

The first step identifies these gaps and provides hints for filling them, while the second step adds the hints to the question to generate a final code solution.

Experimental results on two benchmark datasets demonstrate that GFP significantly improves the mathematical reasoning abilities of SLMs.

Mohammadkhani, Mohammad Ghiasvand, 2024, Gap-Filling Prompting Enhances Code-Assisted Mathematical Reasoning

Document

Open

Share

Source

Articles recommended by ES/IODE AI

Clinical Relevance of Plaque Distribution for Basilar Artery Stenosis
study endovascular imaging wall basilar complications plaque postoperative artery plaques stenosis