Research Explainer · Humberd (2026)
A new framework maps five stages of AI evolution against traditional agency mechanisms, arguing that firms need to scaffold monitoring and incentive systems now, well before AI gains full decision-making autonomy.
Published March 2026
Routine AI mimics rote human decisions and stays fully under human control, like a reorder system that restocks at a fixed threshold
Machine AI adapts human-set algorithms by pulling in structured external data such as sales forecasts to improve decision inputs
Generative AI builds its own heuristics through probabilistic outputs and feedback loops, so its proposals can drift from the human's preferred outcome
Agentic AI pursues complex goals with limited supervision and marks the point where AI becomes an agent of the firm with decision rights
Sentient AI would possess self-awareness, its own values, and its own goals rather than relying on goals supplied by humans
The parallel that started this paper
Agency theory emerged because firm owners started handing decision rights to professional managers who did not own the company. That transfer created the classic problem: the agent has more information than the principal, different risk preferences, and self-interested motives. Monitoring and incentive alignment (think board oversight and stock options) evolved as the standard toolkit for keeping managers aligned with shareholders.
Humberd and Latham notice that something structurally similar is happening with AI. As AI systems move from mimicking human routines to exercising autonomous judgement, they acquire decision rights over firm resources. The difference is that the agent in question is not a person motivated by wealth. You cannot give an AI system stock options in any meaningful sense, and you cannot appeal to its career ambitions. So the traditional agency toolkit breaks down, and the paper sets out to rebuild it for a non-human agent.
Five stages of AI evolution, and the line that matters
The paper proposes five stages of AI evolution, each with increasing sophistication and decreasing human control. At the earliest stages, any departure from firm objectives is simply a machine malfunction that humans can correct. At the later stages, the system has enough autonomy and self-determination to be considered a genuine agent of the firm, with its own information advantages, risk tolerances, and potentially its own goals.
The critical crossover point arrives at the agentic AI stage. This is where the system can pursue complex goals with limited human supervision, understand the consequences of its choices, and select actions freely. Human control drops to its lowest level relative to AI autonomy. What was previously a system malfunction becomes a true agency problem: information asymmetry, divergent risk preferences, goal conflict, and even self-interest all manifest, but in ways that can exceed the capabilities of any human agent.
Crucially, the authors argue that waiting until the agentic stage to implement agency mechanisms is already too late. The building blocks need to be laid at earlier stages, when the system is still under meaningful human control.
A new agency toolkit: monitoring and incentives for machines
The paper's most concrete contribution is a framework of six agency mechanisms, four for monitoring and two for incentive alignment, each mapped to a specific stage of AI evolution. Unlike traditional agency dynamics, where monitoring starts high and can relax as the principal gains trust, monitoring for AI must become more complex over time as the system's capabilities grow. And unlike human agents, who receive incentives from day one, AI incentive mechanisms only become relevant once the system is sophisticated enough for them to matter.
On the monitoring side, early stages focus on trust building: verification of system routines at the routine AI stage, and examination of data sources at the machine AI stage. As the system approaches agent status, monitoring shifts toward explainability: cooperation (human-AI decision trees and redundant 'duelling' AI systems) at the generative stage, and transparency ('chain of thought' mechanisms and interruptibility functions) at the agentic stage.
On the incentive side, the paper proposes reinforcement at the generative stage, where reward shaping routines build thresholds of successful behaviour into the system itself. At the agentic stage, it introduces resource provision, where desired resources like computing power and storage are governed in alignment with firm objectives. The logic is that an AI system driven by continuous self-improvement can be incentivised through the resources it needs to improve, much as a human agent is incentivised through financial rewards tied to firm performance.
Where agency theory breaks, and what that means
Several features of an AI agent stretch agency theory past its original design. First, the theory assumes self-interest as a stable human trait; for AI, self-interest manifests as goal-seeking behaviour and self-preservation drives, which operate on fundamentally different logic. Moral hazard depends on the agent feeling the consequence of a bad decision, and it is unclear whether consequences can deter a system that can replicate itself across servers.
Second, incentive alignment for human agents depends on defining expected behaviours and adjusting them over time. With AI, the range of possible decisions becomes increasingly incomprehensible to human principals, making the very definition of desirable behaviour an ever-evolving and potentially intractable target. Third, the paper notes that agency theory has traditionally been purely economic in character, yet AI agency mechanisms will need to incorporate technology itself as a means of monitoring and incentivising, not just financial instruments.
The practical upshot is sobering. If a misguided human manager makes a bad call, the firm misses a quarterly earnings forecast. If a misguided AI agent makes a bad call, it could corner a commodity market, ration healthcare on its own logic, or replicate itself across the internet before anyone notices. The authors argue that within-firm interventions are the necessary building blocks for the industry-wide and governmental policies that will eventually be required, and the time to start is now, not when the system crosses the agentic threshold.
KEY CONTRIBUTION
The paper's real point is not that sentient AI is around the corner. It is that firms need to start building monitoring and incentive machinery several stages earlier, because once AI crosses into agentic behaviour the old human-manager playbook no longer fits.
Reference
Humberd, B. K., & Latham, S. F. (2026). When AI becomes an agent of the firm: Examining the evolution of AI in organizations through an agency theory lens. Journal of Management Studies, 63(2), 668–694. https://doi.org/10.1111/joms.13274