FastOMOP: A Foundational Architecture for Reliable Agentic Real-World Evidence Generation on OMOP CDM data
Generating real-world evidence (RWE) from massive electronic health record databases is currently a slow, manual process that requires deep clinical and technical expertise. While AI and large language models (LLMs) offer a way to automate these tasks, they are often unreliable, prone to errors, and difficult to govern. This paper introduces FastOMOP, an open-source, multi-agent architecture designed to make RWE generation safer, more transparent, and more reliable by embedding governance directly into the system's infrastructure rather than relying on the AI model alone.
A Layered Approach to Safety
FastOMOP addresses the risks of autonomous AI by separating the system into three independent layers: orchestration, governance, and observability. The orchestration layer manages how agents work together to complete complex tasks, while the observability layer creates a permanent, detailed record of every step taken during the process. The most critical component is the governance layer, which acts as a "safety gate" at the boundary of the system. Because this layer uses deterministic, rule-based validation that operates independently of the AI’s reasoning, it ensures that even if an agent hallucinates or is compromised, it cannot execute unsafe or unauthorized database queries.
Pluggable Agent Teams
The architecture is designed to be flexible, allowing specialized "agent teams" to be plugged into the system to handle specific parts of the RWE lifecycle, such as defining patient groups (phenotyping), designing studies, or performing statistical analysis. These agents inherit the system’s safety and auditability features automatically. By using the Model Context Protocol, the system ensures that agents only have access to the specific tools and data they need to perform their assigned tasks, following the principle of least privilege to minimize potential damage from any single agent failure.
Proven Reliability in Clinical Settings
To test the architecture, the researchers built a proof-of-concept system capable of converting natural language requests into SQL queries for OMOP-standardized databases. They validated this system across three different datasets, including synthetic data, the MIMIC-IV database, and a real-world dataset from the Lancashire Teaching Hospitals NHS Foundation Trust. The results showed high reliability scores (ranging from 0.84 to 0.94). Most importantly, the system achieved perfect scores in blocking both adversarial attacks and out-of-scope queries, proving that the architectural governance successfully prevents the system from performing unauthorized or irrelevant actions.
Shifting the Focus to Architecture
The findings suggest that the current "reliability gap" in clinical AI is not necessarily a failure of the AI models themselves, but rather a lack of proper architectural infrastructure. By moving safety controls out of the agent’s internal logic and into the process boundary, FastOMOP provides a way to deploy AI in high-stakes clinical environments where transparency and regulatory compliance are essential. This approach allows for the progressive automation of the RWE lifecycle while maintaining strict control over how data is accessed and analyzed.

Comments (0)
to join the discussion
No comments yet
Be the first to share your thoughts!