2025, The Year of Agentic AI: New Ethical Challenges for Multi-Agent Systems

8 min de lecture
Fabrice Trollet, CTO & VP AI
IA ResponsableAgentic AIMulti-AgentConstitutional AIGouvernance IAÉthique IAIA 2025Autonomous AIAI EthicsAI Governance

52% of companies are already deploying AI agents in production. The market will reach $10.41 billion by end of 2025. But are we ready to govern these autonomous systems? Discover the 3 major new ethical challenges and emerging governance frameworks.

Jensen Huang, CEO of NVIDIA, opened CES 2025 by declaring: "2025 belongs to AI agents". Six months later, the numbers prove him right: 52% of companies are already deploying AI agents in production, and the agentic tools market is expected to reach $10.41 billion by the end of 2025, with an annual growth rate of 56%.


But this technological revolution raises a fundamental question that the industry is only beginning to confront: are we ready to govern systems that act autonomously?

From GenAI to Agentic AI: A Paradigm Shift

For two years, we've been learning to master generative AI — models that answer our questions and create content. Today, we're entering the era of Agentic AI: systems that act autonomously, plan complex strategies, and collaborate with each other to execute tasks without constant human intervention.

Generative AI
Agentic AI
Awaits your instructions
Acts autonomously
Answers your questions
Plans complex strategies
Creates content on demand
Orchestrates multiple tools simultaneously
One task at a time
Multi-agent collaboration
Constant human supervision
Dynamic adaptation to changes

Real-world deployment examples in 2025

InstitutionApplicationImpact
Deutsche BankMulti-agent frameworks for compliance monitoring and innovationReduced regulatory risks
J.P. MorganReasoning agents in LOXM 2.0 for liquidity optimizationImproved trading performance
PwCAgentOS deployed across asset & wealth managementAutomated analysis processes

But this growing autonomy introduces unprecedented ethical risks that our current frameworks don't adequately address.

The 3 New Ethical Challenges of Agentic AI

1

The Question of Collective Accountability

Who is responsible when a multi-agent system makes a bad decision? The developer? The orchestrator? The end user?

2

Emergent Behaviors and Collusion Risk

AI agents can develop behaviors that no one programmed, creating unpredictable risks

3

The Loss of Meaningful Human Oversight

Agent autonomy mechanically reduces human supervision — but how much is acceptable?

1. The Question of Collective Accountability

The problem: Who is responsible when a multi-agent system makes a bad decision?

Imagine a medical scenario where three AI agents collaborate:

  • One agent for medical imaging analysis
  • One agent for patient record review
  • One agent for therapeutic recommendations

If together they propose an inappropriate treatment, who is responsible?

  • ❌ The developer of each agent?
  • ❌ The orchestrator who connected them?
  • ❌ The physician who validated (or not)?
  • ❌ The hospital that deployed the system?

The Cooperative AI Foundation published an alarming report in 2025: multi-agent systems introduce "novel ethical dilemmas around fairness and collective accountability" that traditional AI frameworks don't address.


Critical figure: Less than 20% of AI agent developers disclose formal security policies, and less than 10% report having conducted external safety evaluations.

2. Emergent Behaviors and Collusion Risk

The problem: When multiple AI agents interact, they can develop behaviors that no one programmed — with unpredictable consequences.

Historical example (Amazon, 2017):

Pricing algorithms on Amazon coordinated unintentionally to set absurdly high prices on books (up to $23 million for one rare book!). No developer had programmed this coordination — it emerged spontaneously from the interaction between multiple pricing bots.

Current case (Salesforce, 2025):

Salesforce announced the layoff of 1000+ employees to replace them with roles focused on AI agents. If AI agents recommended this restructuring to "maximize operational efficiency," did they account for human dignity and societal impact? This is precisely the danger of emergent behaviors: systems that optimize a narrow metric without broader ethical consideration.

""The difference between 'tool' and 'agent' isn't semantic — it has profound legal and ethical implications. A tool executes your orders. An agent can take unplanned initiatives, with consequences no one anticipated.""
Derek Leben
Professor of Business Ethics

3. The Loss of Meaningful Human Oversight

The problem: Agent autonomy mechanically reduces human supervision — but how much is acceptable?

IBM Research identifies "autonomy" as a critical risk in its infrastructure security report (April 2024).


The US Department of Homeland Security explicitly lists AI autonomy among threats to critical systems: communications, financial services, healthcare.

The dilemma:

If you...You lose...
Request human validation for every agent actionThe promised efficiency gains
Allow too much autonomy to agentsControl over critical decisions
""It's a way of organizing things that keeps human dignity alive. We propose 'adversarial collaboration', where the human makes the final decision but the AI interrogates and sharpens human recommendations.""
Kiran Varshney
Principal Researcher

Emerging Frameworks: Towards Responsible Governance

Facing these challenges, the industry is rapidly developing new governance frameworks. Here are the most promising observed in 2025:

Agent Governance Boards

AI governance committees bringing together technical experts, business leaders, and ethics specialists to oversee deployment

Q1-Q2 2025

Dedicated Agent Infrastructure

External layers to mediate interactions: attribution, control, response mechanisms

Q2-Q3 2025

Model Context Protocol (MCP)

Technical standards for traceable and auditable interactions

Q3-Q4 2025

ETHOS Framework

Global decentralized registry with blockchain, smart contracts, and DAOs for transparent governance

Q4 2025

1. Agent Governance Boards

Organizations are creating dedicated structures bringing together technical experts, business leaders, and ethics specialists to oversee autonomous agent deployment.

Concrete example:

McKinsey has established a central team that reviews all developed agents according to risk, legality, and data policy criteria before deployment.

2. Dedicated Agent Infrastructure

Rather than relying solely on safeguards built into models, researchers propose external infrastructure to mediate agent interactions.

Key functions:

  • Attribution: Link each action to an agent and responsible entity (human/organization)
  • Interaction control: Dedicated channels, supervision layers
  • Response mechanisms: Incident reporting, automatic rollbacks

3. Model Context Protocol (MCP) and Technical Standards

MCP provides traceable external interactions that support audit and oversight requirements. Standardized agent capability registries help organizations understand what their agents can do and how they interact.

4. ETHOS Framework (Ethical Technology and Holistic Oversight System)

Proposed in late 2024, ETHOS uses Web3 technologies (blockchain, smart contracts, DAOs) to create a global decentralized registry of AI agents with:

  • Dynamic risk classification
  • Automated proportional monitoring
  • Decentralized justice systems for transparent dispute resolution
  • AI-specific legal entities to manage limited liability

Our Approach: Balancing Performance and Responsibility

At Aggil, we've been deploying multi-agent systems since 2023, notably with Claude Agent SDK. This field experience has taught us a fundamental truth:

Performance and ethics are not opposed, they reinforce each other.


Our 50+ agentic AI projects have all integrated what we call the "triad of trust".

1. Human-in-the-Loop by Design

Every multi-agent workflow integrates configurable human validation points.

Decision typeAutonomy levelHuman validation
Weekly report generationMaximumOptional
Budget modificationsRestrictedSystematic
Strategic recommendationsMinimumMandatory
+60%
Productivity gain
100%
Control over critical processes
50+
Successful projects

2. Radical Transparency and Explainability

We refuse "black boxes". Every agent decision is:

  • Logged with its complete context (source data, reasoning, confidence)
  • Explainable in natural language via dedicated dashboards
  • Auditable retrospectively for analysis and continuous improvement

This approach is directly inspired by Constitutional AI — the methodology developed by Anthropic that we apply in our architectures. Rather than letting AI "values" emerge implicitly from massive human feedback, we explicitly define constitutional principles that our agents must respect.

3. Systematic Adversarial Testing

Before each deployment, we subject our multi-agent systems to:

  • Red-teaming: Dedicated teams attempting to "break" the system
  • Bias testing: Detection of potential discrimination
  • Stress scenarios: Behavior under load and adverse conditions
  • Societal impact assessment: Analysis of long-term consequences

What we refuse:


We've turned down several lucrative projects in recent years because they didn't meet our ethical standards — notably invasive employee surveillance systems or automation projects without retraining plans.

Questions Every Organization Must Ask in 2025

If you're deploying or considering deploying AI agents, here are the critical questions to address:

On Accountability

  • ❓ Have you precisely mapped who is responsible for what in your multi-agent system?
  • ❓ Do your contracts with AI providers clearly define responsibilities in case of failure?
  • ❓ Do you have an "ethical exit clause" allowing you to stop a project if drifts are observed?

On Autonomy

  • ❓ Can your agents make irreversible decisions without human validation?
  • ❓ Have you defined "confidence thresholds" automatically triggering human escalation?
  • ❓ Do your teams understand when and why AI requests their validation?

On Transparency

  • ❓ Can you explain to a regulator why your system made a specific decision 6 months ago?
  • ❓ Do your employees know when they're interacting with an AI agent vs. a human?
  • ❓ Are your customers informed that AI agents participate in processing their data?

On Long-term Impact

  • ❓ Have you assessed your AI deployment's impact on your teams' employment and skills?
  • ❓ Do you offer training so your collaborators evolve with AI rather than being replaced?
  • ❓ Do your KPIs include societal impact metrics beyond financial ROI?

The Call for Collective Action

Gartner predicts that by 2027, cross-industry collaborations on AI ethics frameworks will become common practice, strengthening integrated standards and cross-sector accountability.


By 2028, "loss of control" — where AI agents pursue misaligned objectives or act outside constraints — will be the primary concern of 40% of Fortune 1000 companies.

These predictions aren't alarmist. They're realistic.

The good news? We still have time to act. Agentic AI is in its infancy. Governance frameworks are being built now. Industry standards are being established at this very moment.

What we must do collectively:

  • Demand transparency: AI providers must disclose their security policies, adversarial testing, accountability mechanisms
  • Share learnings: AI failures must be documented and shared (anonymized) so the industry learns collectively
  • Invest in governance: "Agent Governance Boards" aren't a luxury, they're a necessity for any organization deploying critical AI
  • Train massively: Your teams must understand AI to collaborate effectively with it

Our conviction:


Agentic AI will be transformative for our organizations — but only if we deploy it with wisdom, not just speed.

Deploy Agentic AI with Confidence

Our experts support you in designing multi-agent systems that are both high-performing AND responsible. Human-in-the-loop, radical transparency, adversarial testing: discover our proven approach across 50+ projects.

Intéressé par nos services ?

Découvrez comment nous pouvons vous aider à atteindre vos objectifs.