Leverage Laws: A Per-Task Framework for Human-Agent Collaboration proposes a standardized way to measure the effectiveness of human-AI teamwork. While existing research has established that AI can improve output and that productivity varies across different tasks, there has been no unified unit of analysis to help practitioners decide where to invest their time to maximize the displacement of human labor. This framework provides a mathematical ratio that connects the costs of task specification, mid-run interruptions, and final reviews into a single, actionable model.
The Leverage Ratio
The core of the framework is the "per-task leverage" ratio. It is calculated by dividing the amount of human work displaced by the agent by the total human time spent managing the agent. This management time is broken down into three specific phases: planning (specifying the task), resolving interruptions (clarifying details mid-run), and reviewing (verifying the final output). By using this ratio, teams can identify exactly which part of the collaboration process is consuming the most time and prioritize engineering efforts—such as better tools or clearer workflows—to reduce those specific costs.
Information Density and Direction
The paper introduces the concept of "information density" to explain why some parts of a task are more time-consuming than others. It argues that communication between a human and an agent is asymmetric: human-to-agent flow (like typing or speaking) is bottlenecked by human output speed, while agent-to-human flow (like reading or viewing a dashboard) is bottlenecked by human visual processing. Because these two directions have different limits, improving one does not necessarily help the other. The framework suggests that by understanding the "information share" of each phase, practitioners can choose the right intervention—such as voice-to-text for high-input phases or structured dashboards for high-output phases—to improve efficiency.
Scaling and Amortization
The framework distinguishes between one-off tasks and recurring workflows. For a single task, there is a "non-zero floor" on planning time caused by the irreducible novelty of the task itself; no matter how good the AI is, some human input is always required to define what is new. However, when tasks recur or spawn subtasks, the cost of planning can be spread out, or "amortized," over many instances. Over time, as the human and agent build up shared context—or "memory"—the information density of their exchanges increases, allowing them to complete tasks faster and with fewer interruptions.
Practical Implications
The framework is designed to be testable, offering a list of open problems and a protocol for falsification. For example, it predicts that improvements to input speed will only reduce time in phases that are heavily reliant on human input, while improvements to output displays will only help in phases that are heavily reliant on reading or viewing. If these interventions do not produce the expected phase-specific time savings, the framework’s core claim regarding directional asymmetry would be refuted. By providing these clear, measurable targets, the paper aims to move human-agent collaboration from qualitative observation to rigorous engineering.
Comments (0)
to join the discussion
No comments yet
Be the first to share your thoughts!