Papers
arxiv:2512.14503

RecGPT-V2 Technical Report

Published on Dec 16
· Submitted by
TangJiakai
on Dec 17
Authors:
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,

Abstract

RecGPT-V2 enhances recommender systems by integrating a Hierarchical Multi-Agent System, Hybrid Representation Inference, Meta-Prompting, constrained reinforcement learning, and an Agent-as-a-Judge framework to improve efficiency, explanation diversity, generalization, and human preference alignment.

AI-generated summary

Large language models (LLMs) have demonstrated remarkable potential in transforming recommender systems from implicit behavioral pattern matching to explicit intent reasoning. While RecGPT-V1 successfully pioneered this paradigm by integrating LLM-based reasoning into user interest mining and item tag prediction, it suffers from four fundamental limitations: (1) computational inefficiency and cognitive redundancy across multiple reasoning routes; (2) insufficient explanation diversity in fixed-template generation; (3) limited generalization under supervised learning paradigms; and (4) simplistic outcome-focused evaluation that fails to match human standards. To address these challenges, we present RecGPT-V2 with four key innovations. First, a Hierarchical Multi-Agent System restructures intent reasoning through coordinated collaboration, eliminating cognitive duplication while enabling diverse intent coverage. Combined with Hybrid Representation Inference that compresses user-behavior contexts, our framework reduces GPU consumption by 60% and improves exclusive recall from 9.39% to 10.99%. Second, a Meta-Prompting framework dynamically generates contextually adaptive prompts, improving explanation diversity by +7.3%. Third, constrained reinforcement learning mitigates multi-reward conflicts, achieving +24.1% improvement in tag prediction and +13.0% in explanation acceptance. Fourth, an Agent-as-a-Judge framework decomposes assessment into multi-step reasoning, improving human preference alignment. Online A/B tests on Taobao demonstrate significant improvements: +2.98% CTR, +3.71% IPV, +2.19% TV, and +11.46% NER. RecGPT-V2 establishes both the technical feasibility and commercial viability of deploying LLM-powered intent reasoning at scale, bridging the gap between cognitive exploration and industrial utility.

Community

Paper author Paper submitter

🌟 RecGPT-V2: A Major Leap in LLM-Powered Recommendation (RecGPT-V1’s Power Upgrade!) 🌟

Thrilled to unveil RecGPT-V2—the highly anticipated successor to RecGPT-V1! This agentic framework addresses V1’s core limitations, fusing cognitive reasoning with industrial scalability for next-gen intent-centric recommendations.

🔥 Core Innovations:
Hierarchical Multi-Agent + Hybrid Representation: 60% less GPU usage, 9.39%→10.99% exclusive recall, and 32K→11K token compression (context intact).
Meta-Prompting: +7.3% explanation diversity with adaptive, non-generic prompts.
Constrained RL: Resolves multi-reward conflicts—+24.1% better tag prediction and +13.0% higher explanation acceptance vs. V1.
Agent-as-a-Judge: Human-like multi-step evaluation, closer alignment with real-world standards.

🚀 Taobao A/B Test Results:
+2.98% CTR | +3.71% IPV | +2.19% TV | +11.46% NER (Novelty Exposure Rate)

Validated for large-scale deployment—bridging cognitive AI and practical utility, with room to evolve.

🎯 Why It Matters:
Fixes V1’s pain points (computational bloat, rigid explanations, weak generalization, oversimplified evaluation) to deliver a scalable, efficient, human-aligned paradigm. Perfect for researchers and engineers—this is just a key milestone in refining intent-driven AI!

👉 Dive into the full technical report to unlock scalable intent-driven recommendations. Let’s shape personalized AI4Rec’s future!

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2512.14503 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2512.14503 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2512.14503 in a Space README.md to link it from this page.

Collections including this paper 0

No Collection including this paper

Add this paper to a collection to link it from this page.