[UW, Allen AI, IBM] Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection

  Рет қаралды 50

Trend in Research

Trend in Research

Күн бұрын

arxiv.org/abs/2310.11511
Despite their remarkable capabilities, large language models (LLMs) often produce responses containing factual inaccuracies due to their sole reliance on the parametric knowledge they encapsulate. Retrieval-Augmented Generation (RAG), an ad hoc approach that augments LMs with retrieval of relevant knowledge, decreases such issues. However, indiscriminately retrieving and incorporating a fixed number of retrieved passages, regardless of whether retrieval is necessary, or passages are relevant, diminishes LM versatility or can lead to unhelpful response generation. We introduce a new framework called Self-Reflective Retrieval-Augmented Generation (Self-RAG) that enhances an LM's quality and factuality through retrieval and self-reflection. Our framework trains a single arbitrary LM that adaptively retrieves passages on-demand, and generates and reflects on retrieved passages and its own generations using special tokens, called reflection tokens. Generating reflection tokens makes the LM controllable during the inference phase, enabling it to tailor its behavior to diverse task requirements. Experiments show that Self-RAG (7B and 13B parameters) significantly outperforms state-of-the-art LLMs and retrieval-augmented models on a diverse set of tasks. Specifically, Self-RAG outperforms ChatGPT and retrieval-augmented Llama2-chat on Open-domain QA, reasoning and fact verification tasks, and it shows significant gains in improving factuality and citation accuracy for long-form generations relative to these models.

Пікірлер
The moment we stopped understanding AI [AlexNet]
17:38
Welch Labs
Рет қаралды 733 М.
What are AI Agents?
12:29
IBM Technology
Рет қаралды 59 М.
ОСКАР vs БАДАБУМЧИК БОЙ!  УВЕЗЛИ на СКОРОЙ!
13:45
Бадабумчик
Рет қаралды 6 МЛН
Smart Sigma Kid #funny #sigma #comedy
00:25
CRAZY GREAPA
Рет қаралды 38 МЛН
4 Methods of Prompt Engineering
12:42
IBM Technology
Рет қаралды 121 М.
AI Evolution: DSPy Transforms Prompting with Self-Improving Pipelines
15:58
What Creates Consciousness?
45:45
World Science Festival
Рет қаралды 105 М.
RAG: The Secret to Making AI Work for You (Not Against You)
7:56
Generative AI in a Nutshell - how to survive and thrive in the age of AI
17:57
Prompt Engineering, RAG, and Fine-tuning: Benefits and When to Use
15:21
What is Sentiment Analysis?
10:05
IBM Technology
Рет қаралды 13 М.
Large Language Models (LLMs) - Everything You NEED To Know
25:20
Matthew Berman
Рет қаралды 66 М.