From Intent to Execution: Composing Agentic Workflows with Agent Recommendation
Creating Multi-Agent Systems (MAS)—where multiple AI agents work together to fulfill a user's goal—is currently a manual, labor-intensive process. Developers must write plans, select specific agents, and build execution graphs by hand. This paper introduces "AutoMAS," an automated framework designed to streamline this process. By using an LLM-based planner and a sophisticated agent recommendation system, the framework can take a user's natural language intent and automatically generate, configure, and execute a multi-agent workflow with minimal human intervention.
Automating Workflow Design
The AutoMAS framework functions by breaking down a user's request into a series of subtasks. It represents these tasks as a finite state machine, which allows for flexible, non-sequential execution. To ensure the system is robust, it uses a "Variable Call Graph" that can reroute tasks if an agent becomes unavailable or if environmental conditions change. This architecture allows the system to adapt to real-time constraints, such as prioritizing lower costs or faster execution speeds based on user requirements.
The Two-Stage Agent Recommender
A central challenge in building these systems is finding the right agent for a specific job from a vast registry. AutoMAS solves this with a two-stage retrieval process:
Retriever: A fast, hybrid search (combining keywords and vector embeddings) that quickly identifies a broad list of potentially relevant agents.
Re-ranker: An LLM-based module that analyzes the candidates from the first stage to provide a more nuanced, accurate selection based on the specific task context.
To further improve accuracy, the framework uses "agent description enrichment," where it generates synthetic queries from agent documentation at the time of indexing. This creates a richer, more searchable profile for each agent, making it easier for the system to match them to user needs.
The Role of the Critique Agent
The framework includes an optional "critique" mechanism that acts as a quality control layer. Unlike the recommender, which focuses on finding the best individual agent, the critique agent evaluates the entire plan. It checks whether the selected agents are compatible with one another and whether the overall workflow adheres to user constraints like security, cost, and latency. Experiments showed that including this critique step significantly improved the system's recall scores, proving that a holistic review of the entire workflow is essential for building reliable, end-to-end multi-agent systems.
Performance and Scalability
The researchers benchmarked AutoMAS using established datasets to evaluate both planning and agent selection. The results indicate that the two-stage retrieval approach, combined with description enrichment and the critique mechanism, outperforms existing state-of-the-art methods. The system demonstrated high robustness and scalability, effectively navigating the "needle in an agentic stack" problem as the number of available agents grows. By automating the transition from intent to execution, AutoMAS provides a scalable path for developing complex, adaptive AI applications.
Comments (0)
to join the discussion
No comments yet
Be the first to share your thoughts!