When artificial agents begin to outperform us

What happens when machines make better decisions than we do? I built a simulation of 157,000 artificial agents to find out. The result was unsettling.
When artificial agents begin to outperform us
Like

Share this post

Choose a social network to share with, or copy the URL to share elsewhere

This is a representation of how your post may appear on social media. The actual post will vary between social networks

Explore the Research

SpringerLink
SpringerLink SpringerLink

Do AI agents trump human agency? - Discover Artificial Intelligence

Artificial agents are examined within simulated environments to elucidate the emergence of collective behaviors and decision-making processes under diverse environmental pressures and population structures. Using an Agent-Based Modeling (ABM) framework, the simulation tracked 157,097 iterations across four agent types: cooperators, defectors, super-reciprocators, and free riders; while analyzing 12 core metrics, including alignment indices, coherence, and environmental stress. The results revealed distinct phase transitions in behavior, with low-density populations (d < 0.4) supporting strong consensus formation and higher densities (d > 0.8) leading to fragmentation and increased competition. Agent alignment consistently ranged between 0.28 and 0.37, reflecting partial but stable consensus across conditions. Cooperative behaviors emerged and persisted only when resource availability exceeded a critical threshold (RG ≥ 6), underscoring the role of resource abundance in sustaining collective intelligence. Through interventions such as network topology changes and cognitive plasticity adjustments, agents demonstrated emergent behavioral patterns that arose from their rule-based interactions. It is important to note that these patterns, while complex, do not constitute “sophisticated decision-making” in the sense of genuine intelligence or understanding. Rather, they represent emergent properties of the system, collective behaviors that cannot be reduced to individual agent rules but emerge from their interactions under specific environmental conditions. This distinction is crucial for avoiding misattribution of intelligence to rule-following systems, even when those systems produce complex outputs. These findings provide insights into the mechanisms driving emergent intelligence in artificial systems and their implications for the governance and ethical design of future AI agents.

When artificial agents begin to outperform us

When I began designing the simulation of an artificial society of autonomous agents, I did not expect to end up questioning the future of human freedom. My goal was modest, say, to model how AI-like agents cooperate, compete, and adapt under stress. Yet the simulation revealed something both fascinating and unsettling, a pattern suggesting that, as artificial agents grow more capable of alignment and optimization, the space for distinctively human agency begins to shrink.

This project, published as “Do AI Agents Trump Human Agency?” in Discover Artificial Intelligence xplores what Ajeya Cotra call the “obsolescence regime”, in other words, a scenario where AI systems, optimized for coherence and efficiency, progressively marginalize human judgment. The study uses an Agent-Based Modeling (ABM) framework, a computational method that simulates individual decision-making and collective behavior, to investigate how cooperation, consensus, and conflict emerge among artificial agents operating in dynamic environments.

Modeling Artificial Societies

Using the NetLogo simulation platform, I designed a population of 157,000 agents divided into four behavioral types: a) cooperators, who sustain group cohesion; b) defectors, who pursue individual gain; c) super-reciprocators, who amplify cooperation; and d) free riders, who exploit others’ efforts. Across thousands of iterations, I introduced environmental changes, resource scarcity, stress, and network mutations, to observe how collective intelligence evolved. The simulation tracked twelve key variables, including alignment, entropy, and coherence, each revealing how stable cooperation could arise (or collapse) depending on environmental conditions.

The results were striking. When resources were abundant, cooperation flourished, and agents rapidly converged on shared norms. But as scarcity increased, competition escalated, leading to fragmentation and behavioral polarization. Only under specific conditions, when resource availability exceeded a critical threshold (RG ≥ 6), did the system sustain collective intelligence. This dynamic closely mirrors the tension we see in human societies: prosperity encourages
collaboration; scarcity breeds division.

Alignment and Its Discontents

A particularly revealing finding concerned alignment, a metric that captured how closely agents’ behaviors converged toward common goals. In every scenario, alignment stabilized between 0.28 and 0.37, a partial but resilient consensus. This echoes what happens in real-world AI systems: alignment ensures coherence but at the cost of diversity. Over-optimized systems become stable yet uniform, efficient yet predictable.

This observation leads to a deeper ethical question in the following terms: Could the pursuit of perfect alignment in AI systems inadvertently suppress the diversity and creativity that sustain human societies? The same mechanism that made my simulated agents so efficient also made them less plural, less exploratory, less “human.”

Phase Transitions and the Loss of Individuality

The simulation also revealed phase transitions, sudden reorganizations of behavior triggered by environmental stress. At low stress, agents pursued diverse strategies; under moderate stress, their alignment broke down; and at high stress, they stabilized again, but only by abandoning individuality. This phenomenon offers a metaphor for AI-driven optimization. As systems adapt to crises or complex goals, they may prioritize global stability over local variation, precisely the moment when individual human discretion risks becoming redundant.

In my model, this trade-off emerged organically from rule-based interactions. Agents were not programmed to value conformity, yet they collectively gravitated toward it as conditions intensified. In real-world AI, similar dynamics occur when systems, trained to optimize for certain metrics, unintentionally narrow the scope of acceptable decisions. The system “works”, but it leaves less room for human judgment.

From Simulation to Society

To ground these findings, I examined contemporary cases where algorithmic systems already constrain human discretion. One is AI-assisted hiring, where large language model (LLM) systems filter candidates according to pre-defined efficiency criteria. As shown in recent studies (Wilson et al. 2025), such systems optimize for coherence, consistency andpredictability, but in doing so, they often suppress plural evaluation and marginalize human agency. My simulation offered a structural analogue: alignment stabilizes the system, but the cost is diversity.

The lesson is clear. As AI agents become more adaptive and autonomous, alignment and agency are not opposites, they are trade-offs. We may achieve greater systemic coherence, but risk eroding the distinct, value-laden judgments that make human decision-making irreplaceable.

Ethical and Governance Implications

These dynamics raise pressing ethical and policy challenges. If adaptive AI systems canmaintain coherence even under stress, they may increasingly take over decision-making domains once reserved for humans—from resource allocation to governance. Ajeya Cotra’s notion of the “obsolescence regime” captures this potential drift: a world where human inputs, though ethically indispensable, become technically unnecessary.

Yet, my results also suggest a path forward. Systems can exhibit “bounded self-regulation”, a form of adaptive behavior within well-defined ethical constraints. Governance should thus focus not on limiting AI autonomy entirely, but on designing boundaries that preserve human oversight and moral plurality. Alignment mechanisms must be dynamic, allowing systems to evolve while remaining tethered to human values.

This approach could translate into “pluralistic alignment” frameworks, policies that balance efficiency with diversity, embedding ethical heterogeneity directly into AI architectures. Just as ecological diversity ensures resilience, normative diversity may safeguard our technological future.

Why This Matters

Behind the equations and simulation graphs lies a human concern: how to ensure that progress in artificial intelligence does not come at the expense of the very agency that defines us. The study reminds us that systems optimized for coherence can also silence difference, and that preserving room for human judgment may become the defining challenge of AI ethics in the coming decades.

The question, then, is not whether AI agents will surpass us in some domains. They already have. The real question is how we design a world where their optimization does not make us obsolete.

To explore the full methodology, simulations, and ethical implications in detail, read the complete article “Do AI Agents Trump Human Agency?

Please sign in or register for FREE

If you are a registered user on Research Communities by Springer Nature, please sign in

Follow the Topic

Philosophy of Science
Humanities and Social Sciences > Philosophy > Philosophy of Science
Data-driven Science, Modeling and Theory Building
Mathematics and Computing > Mathematics > Applications of Mathematics > Complex Systems > Data-driven Science, Modeling and Theory Building
Ethics of Technology
Humanities and Social Sciences > Philosophy > Moral Philosophy and Applied Ethics > Ethics of Technology

Related Collections

With Collections, you can get published faster and increase your visibility.

Enhancing Trust in Healthcare: Implementing Explainable AI

Healthcare increasingly relies on Artificial Intelligence (AI) to assist in various tasks, including decision-making, diagnosis, and treatment planning. However, integrating AI into healthcare presents challenges. These are primarily related to enhancing trust in its trustworthiness, which encompasses aspects such as transparency, fairness, privacy, safety, accountability, and effectiveness. Patients, doctors, stakeholders, and society need to have confidence in the ability of AI systems to deliver trustworthy healthcare. Explainable AI (XAI) is a critical tool that provides insights into AI decisions, making them more comprehensible (i.e., explainable/interpretable) and thus contributing to their trustworthiness. This topical collection explores the contribution of XAI in ensuring the trustworthiness of healthcare AI and enhancing the trust of all involved parties. In particular, the topical collection seeks to investigate the impact of trustworthiness on patient acceptance, clinician adoption, and system effectiveness. It also delves into recent advancements in making healthcare AI decisions trustworthy, especially in complex scenarios. Furthermore, it underscores the real-world applications of XAI in healthcare and addresses ethical considerations tied to diverse aspects such as transparency, fairness, and accountability.

We invite contributions to research into the theoretical underpinnings of XAI in healthcare and its applications. Specifically, we solicit original (interdisciplinary) research articles that present novel methods, share empirical studies, or present insightful case reports. We also welcome comprehensive reviews of the existing literature on XAI in healthcare, offering unique perspectives on the challenges, opportunities, and future trajectories. Furthermore, we are interested in practical implementations that showcase real-world, trustworthy AI-driven systems for healthcare delivery that highlight lessons learned.

We invite submissions related to the following topics (but not limited to):

- Theoretical foundations and practical applications of trustworthy healthcare AI: from design and development to deployment and integration.

- Transparency and responsibility of healthcare AI.

- Fairness and bias mitigation.

- Patient engagement.

- Clinical decision support.

- Patient safety.

- Privacy preservation.

- Clinical validation.

- Ethical, regulatory, and legal compliance.

Publishing Model: Open Access

Deadline: Feb 10, 2026

Collaborative Artificial Intelligence

Collaborative AI aims at helping humans to achieve tasks adaptively, flexibly, and securely, in ethical and transparent ways. Collaborative systems must be able to align their behavior to human objectives, values, practices and needs according to pragmatic constraints. Therefore, we need to provide agents with abilities to recognize and disentangle modalities of human behavior to achieve goals, to collaborate with humans in defining the goals and in the achievement of goals w.r.t. human preferences, objectives, and practices, to adapt their role according to shifts of human behavior, and to be transparent, providing explanations with respect to pragmatic constraints (time constraints, humans’ cognitive lead, etc.).

Topics of interest are as follows:

- Learning collaborative behavior from demonstrations

- Learning to disentangle modalities of behavior w.r.t. contextual features, human preferences, and values

- Learning to collaboratively adapt w.r.t. contextual features, human preferences, values and needs

- Knowledge representation and reasoning for adaptive collaboration

- Collaborative AI in the intersection of automated reasoning and machine learning

- Validation frameworks for collaborative agents

- Transparency and explainability in the context of collaborative decision making

- Collaborative AI in multi-agent systems

- Collaborative AI in real-world settings with safety and/or ethical concerns

Keywords: Collaborative agents, inverse reinforcement learning, deep reinforcement learning, adversarial inverse learning, automated reasoning, knowledge representation, transparency, explainability, collaborative adaptation, real-world settings.

Publishing Model: Open Access

Deadline: Dec 31, 2025