Listen "Two-Turn Debate Doesn’t Help Humans Answer Hard Reading Comprehension Questions"
Episode Synopsis
Using hard multiple-choice reading comprehension questions as a testbed, we assess whether presenting humans with arguments for two competing answer options, where one is correct and the other is incorrect, allows human judges to perform more accurately, even when one of the arguments is unreliable and deceptive. If this is helpful, we may be able to increase our justified trust in language-model-based systems by asking them to produce these arguments where needed. Previous research has shown that just a single turn of arguments in this format is not helpful to humans. However, as debate settings are characterized by a back-and-forth dialogue, we follow up on previous results to test whether adding a second round of counter-arguments is helpful to humans. We find that, regardless of whether they have access to arguments or not, humans perform similarly on our task. These findings suggest that, in the case of answering reading comprehension questions, debate is not a helpful format.Source:https://arxiv.org/abs/2210.10860Narrated for AI Safety Fundamentals by Perrin Walker of TYPE III AUDIO.---A podcast by BlueDot Impact.Learn more on the AI Safety Fundamentals website.
More episodes of the podcast AI Safety Fundamentals
AI and Leviathan: Part I
29/09/2025
d/acc: One Year Later
19/09/2025
A Playbook for Securing AI Model Weights
18/09/2025
Resilience and Adaptation to Advanced AI
18/09/2025
Introduction to AI Control
18/09/2025
The Project: Situational Awareness
18/09/2025
The Intelligence Curse
18/09/2025