DynaAct: Large Language Model Reasoning with Dynamic Action Spaces
Abstract
A framework named DynaAct uses large language models to construct a compact action space for sequential decision-making, improving performance and efficiency.
In modern sequential decision-making systems, the construction of an optimal candidate action space is critical to efficient inference. However, existing approaches either rely on manually defined action spaces that lack scalability or utilize unstructured spaces that render exhaustive search computationally prohibitive. In this paper, we propose a novel framework named DynaAct for automatically constructing a compact action space to enhance sequential reasoning in complex problem-solving scenarios. Our method first estimates a proxy for the complete action space by extracting general sketches observed in a corpus covering diverse complex reasoning problems using large language models. We then formulate a submodular function that jointly evaluates candidate actions based on their utility to the current state and their diversity, and employ a greedy algorithm to select an optimal candidate set. Extensive experiments on six diverse standard benchmarks demonstrate that our approach significantly improves overall performance, while maintaining efficient inference without introducing substantial latency. The implementation is available at https://github.com/zhaoxlpku/DynaAct.
Community
A new perspective on test-time scaling — instead of just “thinking longer,” DynaAct makes models think smarter by dynamically constructing compact action spaces for each reasoning step.
It introduces a submodular optimization framework balancing utility & diversity to learn effective reasoning actions.
🧩 The open-source code further integrates vLLM into MCTS, bringing major speedups in node expansion, rollout, and reward computation for large-scale reasoning research.
🔗 https://arxiv.org/abs/2511.08043
💻 https://github.com/zhaoxlpku/DynaAct
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- GPO: Learning from Critical Steps to Improve LLM Reasoning (2025)
- PlanU: Large Language Model Reasoning through Planning under Uncertainty (2025)
- MITS: Enhanced Tree Search Reasoning for LLMs via Pointwise Mutual Information (2025)
- THOR: Tool-Integrated Hierarchical Optimization via RL for Mathematical Reasoning (2025)
- Reasoning Planning for Language Models (2025)
- Adaptive Test-Time Reasoning via Reward-Guided Dual-Phase Search (2025)
- Explore-Execute Chain: Towards an Efficient Structured Reasoning Paradigm (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper