Vincent
(2026-01-31 17:31):
#paper https://arxiv.org/abs/2201.11903
arxiv 2022. Chain-of-Thought Prompting Elicits Reasoning in Large Language Models.
这篇文首次提出了Chain-of-Thought(CoT)的思路,通过在少样本提示中显式提供中间自然语言推理步骤,可以显著提升大语言模型在复杂推理任务上的表现。作者在多种推理任务基准测试上展示了 CoT 的显著增益,尤其在 100B+ 参数规模模型上表现为一种随规模涌现(emergent)的能力。消融实验表明,性能提升并非仅来自“多算一步”,而是顺序化、可读的推理过程本身在发挥作用。该方法无需额外训练或微调,仅通过提示即可实现,因而得以广泛运用,为大模型的可解释推理研究开辟了新方向
arXiv,
2022-01-28T02:33:07Z.
DOI: 10.48550/arXiv.2201.11903
Chain-of-Thought Prompting Elicits Reasoning in Large Language Models
翻译
Abstract:
We explore how generating a chain of thought -- a series of intermediate reasoning steps -- significantly improves the ability of large language models to perform complex reasoning. In particular, we show how such reasoning abilities emerge naturally in sufficiently large language models via a simple method called chain of thought prompting, where a few chain of thought demonstrations are provided as exemplars in prompting. Experiments on three large language models show that chain of thought prompting improves performance on a range of arithmetic, commonsense, and symbolic reasoning tasks. The empirical gains can be striking. For instance, prompting a 540B-parameter language model with just eight chain of thought exemplars achieves state of the art accuracy on the GSM8K benchmark of math word problems, surpassing even finetuned GPT-3 with a verifier.
翻译
Related Links: