Skip to content

[ICLR 2026] Are Reasoning LLMs Robust to Interventions on their Chain-of-Thought?

Notifications You must be signed in to change notification settings

ExplainableML/RLLM-CoT-Robustness

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 

Repository files navigation

Are Reasoning LLMs Robust to Interventions on their Chain-of-Thought?
ICLR 2026

Paper

Alexander von Recum 1,2   Leander Girrbach1,3   Zeynep Akata1,3

1Helmholtz Munich   2 Ludwig Maximilian University of Munich 3 Technical University of Munich, MCML

Abstract

Reasoning LLMs (RLLMs) generate step-by-step chains of thought (CoTs) before giving an answer, which improves performance on complex tasks and makes reasoning transparent. But how robust are these reasoning traces to disruptions that occur within them? To address this question, we introduce a controlled evaluation framework that perturbs a model’s own CoT at fixed timesteps. We design seven interventions (benign, neutral, and adversarial) and apply them to multiple open-weight RLLMs across Math, Science, and Logic tasks. Our results show that RLLMs are generally robust, reliably recovering from diverse perturbations, with robustness improving with model size and degrading when interventions occur early. However, robustness is not style-invariant: paraphrasing suppresses doubt-like expressions and reduces performance, while other interventions trigger doubt and support recovery. Recovery also carries a cost: neutral and adversarial noise can inflate CoT length by more than 200%, whereas paraphrasing shortens traces but harms accuracy. These findings provide new evidence on how RLLMs maintain reasoning integrity, identify doubt as a central recovery mechanism, and highlight trade-offs between robustness and efficiency that future training methods should address.

About

[ICLR 2026] Are Reasoning LLMs Robust to Interventions on their Chain-of-Thought?

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published