Dancing in Chains: Strategic Persuasion in Academic Rebuttal via Theory of Mind
Abstract
RebuttalAgent is a novel framework that applies Theory of Mind to academic rebuttal, utilizing a ToM-Strategy-Response pipeline with supervised fine-tuning and reinforcement learning for improved automated evaluation.
Although artificial intelligence (AI) has become deeply integrated into various stages of the research workflow and achieved remarkable advancements, academic rebuttal remains a significant and underexplored challenge. This is because rebuttal is a complex process of strategic communication under severe information asymmetry rather than a simple technical debate. Consequently, current approaches struggle as they largely imitate surface-level linguistics, missing the essential element of perspective-taking required for effective persuasion. In this paper, we introduce RebuttalAgent, the first framework to ground academic rebuttal in Theory of Mind (ToM), operationalized through a ToM-Strategy-Response (TSR) pipeline that models reviewer mental state, formulates persuasion strategy, and generates strategy-grounded response. To train our agent, we construct RebuttalBench, a large-scale dataset synthesized via a novel critique-and-refine approach. Our training process consists of two stages, beginning with a supervised fine-tuning phase to equip the agent with ToM-based analysis and strategic planning capabilities, followed by a reinforcement learning phase leveraging the self-reward mechanism for scalable self-improvement. For reliable and efficient automated evaluation, we further develop Rebuttal-RM, a specialized evaluator trained on over 100K samples of multi-source rebuttal data, which achieves scoring consistency with human preferences surpassing powerful judge GPT-4.1. Extensive experiments show RebuttalAgent significantly outperforms the base model by an average of 18.3% on automated metrics, while also outperforming advanced proprietary models across both automated and human evaluations. Disclaimer: the generated rebuttal content is for reference only to inspire authors and assist in drafting. It is not intended to replace the author's own critical analysis and response.
Community
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Paper2Rebuttal: A Multi-Agent Framework for Transparent Author Response Assistance (2026)
- Dual-Phase LLM Reasoning: Self-Evolved Mathematical Frameworks (2026)
- CogDoc: Towards Unified thinking in Documents (2025)
- Think-Reflect-Revise: A Policy-Guided Reflective Framework for Safety Alignment in Large Vision Language Models (2025)
- Character-R1: Enhancing Role-Aware Reasoning in Role-Playing Agents via RLVR (2026)
- REVEALER: Reinforcement-Guided Visual Reasoning for Element-Level Text-Image Alignment Evaluation (2025)
- R-Debater: Retrieval-Augmented Debate Generation through Argumentative Memory (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
๐ Our work: RebuttalAgent (Accepted by ICLR 2026)
Academic rebuttal is a high-stakes game of strategic communication, yet current AI models often struggle because they lack the ability to truly understand a reviewer's perspective.
RebuttalAgent is the first framework to ground academic rebuttal in Theory of Mind (ToM). It moves beyond simple text generation by employing a ToM-Strategy-Response (TSR) pipeline to model reviewer mental states, formulate persuasion strategies, and generate grounded responses.
๐ Key Highlights
- ToM-Driven Framework: Uses a novel TSR pipeline to bridge the gap between technical debate and strategic persuasion.
- RebuttalBench: A new, large-scale dataset synthesized via a unique critique-and-refine approach.
- Two-Stage Training: Combines Supervised Fine-Tuning (SFT) for strategy planning with Reinforcement Learning (RL) for scalable self-improvement.
- Rebuttal-RM: A specialized evaluator that achieves human-level scoring consistency, surpassing even GPT-4.1.
- Superior Results: Outperforms advanced proprietary models, showing an 18.3% improvement over base models in automated metrics.
๐ Resources
- ๐ ArXiv: https://arxiv.org/pdf/2601.15715
- ๐ค Model: RebuttalAgent on Hugging Face
- ๐ป Code: GitHub Repository
Disclaimer: Generated content is for reference only to inspire authors and assist in drafting. It is not intended to replace the author's own critical analysis.
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper