From Soliloquy to Agora: Memory-Enhanced LLM Agents with Decentralized Debate for Optimization Modeling
Optimization modeling is essential for industries like logistics and energy, but using Large Language Models (LLMs) to translate natural-language requirements into accurate mathematical models remains difficult. Current methods often rely on either training a model specifically for this task—which makes it hard to upgrade to newer, better models—or using agentic systems that lack the ability to learn from past mistakes. This paper introduces Agora-Opt, a framework that solves these issues by combining decentralized debate with a read-write memory bank, allowing multiple AI agents to collaborate and improve over time without needing constant retraining.
A Flexible Agentic Foundation
Agora-Opt treats the underlying LLM as an interchangeable component. Instead of locking the system into one specific model, the framework uses a role-structured pipeline where agents move from problem text to formulation, code generation, and solver execution. Because the system is modular, users can swap in newer or different LLM backbones without having to redesign the entire process or perform expensive parameter retuning. This design ensures the framework remains portable and adaptable as AI technology evolves.
Decentralized Debate for Better Accuracy
To overcome "single-model myopia"—where a single AI might confidently produce an incorrect solution—Agora-Opt uses a decentralized debate protocol. Rather than relying on a central moderator (which can inherit the biases of its own backbone), the framework runs multiple agent teams on different models. These teams work independently to produce solutions. The system only accepts an answer when the solver-verified outcomes align or when a set number of rounds is reached. This approach uses objective, quantitative results from solvers to reach a consensus, allowing the agents to cross-check each other and even recover correct formulations when initial attempts are flawed.
Learning Through Read-Write Memory
Unlike traditional systems that are "read-only" or static, Agora-Opt features a memory bank that actively records experiences. It consists of two parts: generation memory, which stores verified solutions and debugging steps to speed up future tasks, and debate memory, which tracks how previous disagreements were resolved. By saving these "lessons learned," the system becomes more capable with every use. This allows the agents to improve their performance between runs and ensures that valuable, solver-verified knowledge is preserved even when the underlying model is upgraded.
Proven Performance
Experiments across six public benchmarks demonstrate that Agora-Opt outperforms existing zero-shot LLMs, training-centric models, and previous agentic baselines. The research highlights that this framework is not only effective but also highly robust, showing consistent gains across different backbone choices. By combining collaborative cross-checking with a persistent memory of past successes and failures, Agora-Opt provides a practical, extensible foundation for reliable optimization modeling.
Comments (0)
to join the discussion
No comments yet
Be the first to share your thoughts!