Skip to main navigation Skip to search Skip to main content

Improving cross-task generalization with step-by-step instructions

  • Harbin Institute of Technology
  • Huawei Technologies Co., Ltd.
  • Singapore Management University

Research output: Contribution to journalArticlepeer-review

Abstract

Instruction tuning aims to improve cross-task generalization of language models. However, it is still challenging for language models to complete the target tasks following the instructions because the instructions are general and lack intermediate steps. To solve this issue, step-by-step instructions are introduced to help language models decompose tasks, which can offer detailed and specific procedures for completing the target tasks. The step-by-step instructions are obtained automatically by prompting ChatGPT, and are further merged with the original instructions to tune the language models. Extensive experiments on Sup-NatInst reveal that high-quality step-by-step instructions can enhance cross-task generalization across different model sizes. Further analysis indicates the significance of the order of steps in the proposed instruction for improvement. To promote future research, step-by-step instructions and human quality evaluation results will be released.

Original languageEnglish
Article number172102
JournalScience China Information Sciences
Volume68
Issue number7
DOIs
StatePublished - Jul 2025

Keywords

  • generalization
  • instruction tuning
  • large language model
  • step-by-step instruction
  • task decomposition

Fingerprint

Dive into the research topics of 'Improving cross-task generalization with step-by-step instructions'. Together they form a unique fingerprint.

Cite this