Artificial intelligence is no longer confined to isolated models answering questions or classifying data. Increasingly, organisations are deploying agentic AI systems—autonomous software agents that can plan tasks, make decisions, use digital tools, and interact with other systems or agents in order to achieve defined objectives.
Unlike AI models that simply respond to prompts, AI agents can act. For example, an AI agent may retrieve financial data, draft a report, request additional analysis from another agent, validate outputs against policies, and trigger downstream workflows. When multiple agents are connected in a coordinated chain, they form what can be described as a ‘digital workforce’.
At the same time, most organisations already operate within structured control environments built around IT General Controls (ITGCs). ITGCs are the foundational controls that ensure the reliability, security, and integrity of IT systems and data. They typically cover three core areas: access management (ensuring only authorised users or systems can access data and functionality), change management (ensuring major and minor system changes are tested, approved, and documented), and IT operations controls (ensuring systems are stable, monitored, logged, and resilient).
These controls are the backbone of e.g. management reporting reliability, cybersecurity and regulatory compliance. However, as organisations move from traditional IT systems to interconnected AI agents that reason and act (semi-)autonomously, an important question arises: When AI agents start interacting with each other, are traditional IT General Controls still enough?
Traditional IT systems are largely deterministic. Risks typically arise from unauthorised access, uncontrolled changes, or operational failure. Multi-agent AI systems introduce dynamic autonomy and emergent behaviour. For example, in an AI-driven procurement process, one agent analyses demand, another researches suppliers, and a third negotiates contract clauses. Individually, each agent may function correctly. However, interaction effects may cause e.g. compliance requirements or ESG clauses to be unintentionally deprioritised. The resulting contract may not reflect the organisation’s intended risk posture—even though no access violation occurred.
Agents rely on Application Programming Interfaces (APIs), databases and enterprise systems. Traditional controls such as role-based access and least privilege (access permissions granted to the minimum required) remain critical. For example, if a reporting agent preparing quarterly management reports has unrestricted write access to financial ledgers, an erroneous interpretation could overwrite sensitive data. Strong ITGC environments restrict permissions and enforce separation between preparation and approval system roles.
Prompts, orchestration logic (the system-level coordination of multiple autonomous AI agents to solve complex, multi-step tasks that single models cannot handle alone), and model updates directly influence behaviour. For example, imagine an AI agent assisting with regulatory reporting. If its system instructions are simplified from “include all potentially material disclosures” to “prioritise concise reporting,” the agent may begin omitting borderline disclosures that compliance teams would normally include. Prompts and AI policies must therefore be formally version-controlled and subject to structured approval processes.
Infrastructure resilience, monitoring and logging remain essential. However, while logs may show that an agent called a particular tool, they may not explain why a decision was made. This limitation becomes more pronounced as multiple agents interact.
Traditional ITGCs focus on who accessed a system and whether changes were authorised. They do not typically assess whether autonomous behaviour remains aligned with organisational intent. For example: In a customer complaint workflow, one agent categorises complaints, another drafts responses, and a third approves communications. A subtle misclassification at the first stage could cascade into inappropriate responses at scale. All controls may appear compliant—yet the outcome may still create reputational risk.
Additionally, system-level risks may emerge. For example, two financial modelling agents may iteratively refine forecasts, gradually amplifying optimistic assumptions. Individually logical decisions may collectively distort risk assessments.
Periodic reviews are often insufficient for systems that adapt in real time. A treasury agent adjusting strategies based on live market data requires runtime monitoring rather than quarterly review alone.
In practice, organisations do not need to replace traditional ITGCs. They need to extend them in targeted and pragmatic ways. The objective is not to over-engineer controls, but to ensure that autonomous systems remain aligned with business intent, risk appetite and regulatory obligations.
Below are four concrete areas for action:
Organisations beginning to deploy multi-agent AI do not need to redesign their entire control framework at once. Practical first steps can significantly increase governance maturity:
Governance maturity can evolve incrementally; it does not require a full redesign on day one. Understanding the current maturity stage helps prioritise control investments and avoid over- or under-engineering governance measures. The following maturity stages can be defined and applied in doing so:
In most organisations, this evolution does not require entirely new governance structures. Existing bodies—such as change advisory boards, risk committees, data governance councils, and internal audit—can extend their mandate to include oversight of agentic AI systems.
A practical governance test is simple: if an AI agent were to make a materially wrong decision tomorrow, is it immediately clear who in the organisation is accountable? If ownership is ambiguous, the control framework is yet incomplete.
Robust governance should not be seen purely as a defensive measure. Clear guardrails, behavioural monitoring and defined oversight structures reduce incident risk, increase control/confidence and accelerate regulatory acceptance. In practice, mature control environments often enable faster scaling of AI initiatives rather than slowing them down.
Multi-agent AI does not make traditional ITGCs obsolete. Access management, change management and operational controls remain essential foundations. However, when AI agents interact and influence each other within critical business processes, risk increasingly arises from dynamic behaviour and system-level effects rather than isolated control failures.
Organisations must therefore augment—not replace—their ITGC frameworks. By formalising agent governance, introducing behavioural monitoring, strengthening system-level testing and applying proportionate oversight, they can align autonomy with accountability. The challenge is not whether AI agents will collaborate, but whether governance will evolve fast enough to ensure that collaboration remains lawful, ethical, robust, and value-generating.