BriefGPT.xyz
Jun, 2024
mCoT: 多语言指导调整以提升语言模型的推理一致性
mCoT: Multilingual Instruction Tuning for Reasoning Consistency in Language Models
HTML
PDF
Huiyuan Lai, Malvina Nissim
TL;DR
通过多语言推理一致性研究,我们构建了第一个大规模的多语言数学推理数据集mCoT-MATH,然后引入多语言CoT指令调节以提高模型的推理能力和一致性。我们的7B参数模型mCoT在各种语言上展现出惊人的一致性,并且在性能上优于或与更大规模的闭源和开源模型相媲美。
Abstract
large language models
(LLMs) with
chain-of-thought
(CoT) have recently emerged as a powerful technique for eliciting reasoning to improve various downstream tasks. As most research mainly focuses on English, with
→