响马读paper

一个要求成员每月至少读一篇文献并打卡的学术交流社群

2024, arXiv. DOI: 10.48550/arXiv.2403.16527 arXiv ID: 2403.16527
Hallucination Detection in Foundation Models for Decision-Making: A Flexible Definition and Review of the State of the Art
Neeloy Chakraborty, Melkior Ornik, Katherine Driggs-Campbell
Abstract:
Autonomous systems are soon to be ubiquitous, from manufacturing autonomy to
agricultural field robots, and from health care assistants to the entertainment
industry. The majority of these systems are developed with modular
sub-components for decision-making, planning, and control that may be
hand-engineered or learning-based. While these existing approaches have been
shown to perform well under the situations they were specifically designed for,
they can perform especially poorly in rare, out-of-distribution scenarios that
will undoubtedly arise at test-time. The rise of foundation models trained on
multiple tasks with impressively large datasets from a variety of fields has
led researchers to believe that these models may provide common sense reasoning
that existing planners are missing. Researchers posit that this common sense
reasoning will bridge the gap between algorithm development and deployment to
out-of-distribution tasks, like how humans adapt to unexpected scenarios. Large
language models have already penetrated the robotics and autonomous systems
domains as researchers are scrambling to showcase their potential use cases in
deployment. While this application direction is very promising empirically,
foundation models are known to hallucinate and generate decisions that may
sound reasonable, but are in fact poor. We argue there is a need to step back
and simultaneously design systems that can quantify the certainty of a model's
decision, and detect when it may be hallucinating. In this work, we discuss the
current use cases of foundation models for decision-making tasks, provide a
general definition for hallucinations with examples, discuss existing
approaches to hallucination detection and mitigation with a focus on decision
problems, and explore areas for further research in this exciting field.
2024-03-31 23:50:00
#paper doi.org/10.48550/arXiv.2403.16527, 2024, Hallucination Detection in Foundation Models for Decision-Making: A Flexible Definition and Review of the State of the Art. 智能控制系统能通过预训练在各场景下得到广泛应用,但在训练外场景下表现糟糕。大模型出现有希望提供现有训练方式缺乏的推理能力,但大模型会产生“幻觉”(听起来合理但很差的决策)。本文尝试定义“幻觉”,并给出检测和缓解规划中出现“幻觉”的方法分类,评估指标和数据集等
TOP