Experience-Guided Adaptation of Inference-Time Reasoning Strategies
Abstract
Experience-Guided Reasoner dynamically generates and optimizes computational strategies at inference time, adapting to problems using accumulated experience and improving accuracy and efficiency.
Enabling agentic AI systems to adapt their problem-solving approaches based on post-training interactions remains a fundamental challenge. While systems that update and maintain a memory at inference time have been proposed, existing designs only steer the system by modifying textual input to a language model or agent, which means that they cannot change sampling parameters, remove tools, modify system prompts, or switch between agentic and workflow paradigms. On the other hand, systems that adapt more flexibly require offline optimization and remain static once deployed. We present Experience-Guided Reasoner (EGuR), which generates tailored strategies -- complete computational procedures involving LLM calls, tools, sampling parameters, and control logic -- dynamically at inference time based on accumulated experience. We achieve this using an LLM-based meta-strategy -- a strategy that outputs strategies -- enabling adaptation of all strategy components (prompts, sampling parameters, tool configurations, and control logic). EGuR operates through two components: a Guide generates multiple candidate strategies conditioned on the current problem and structured memory of past experiences, while a Consolidator integrates execution feedback to improve future strategy generation. This produces complete, ready-to-run strategies optimized for each problem, which can be cached, retrieved, and executed as needed without wasting resources. Across five challenging benchmarks (AIME 2025, 3-SAT, and three Big Bench Extra Hard tasks), EGuR achieves up to 14% accuracy improvements over the strongest baselines while reducing computational costs by up to 111x, with both metrics improving as the system gains experience.
Community
We introduce Experience-Guided Reasoner (EGuR), a system that learns better reasoning strategies at inference time. Instead of using a fixed CoT/agent workflow, EGuR generates multiple candidate strategies per query, evaluates them, and updates a structured memory to improve the generation of future strategies. This enables continual adaptation of prompts, sampling parameters, tool use, and control flow, without any retraining. Across math, logic, and Big Bench Extra Hard tasks, EGUR improves accuracy by up to 14% and reduces compute cost by up to 111x, continually improving with experience.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Agentic Context Engineering: Evolving Contexts for Self-Improving Language Models (2025)
- Alita-G: Self-Evolving Generative Agent for Agent Generation (2025)
- DyFlow: Dynamic Workflow Framework for Agentic Reasoning (2025)
- A Multi-Agent Framework for Stateful Inference-Time Search (2025)
- COMPASS: Enhancing Agent Long-Horizon Reasoning with Evolving Context (2025)
- EvolveR: Self-Evolving LLM Agents through an Experience-Driven Lifecycle (2025)
- JoyAgent-JDGenie: Technical Report on the GAIA (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper