Oxford Witt Lab
for Trust in AI

The Oxford Witt Lab (OWL) pioneers foundational assurance for AI-human-institution ecosystems. From our work on undetectable threats and multi-agent security to principled evaluations and governance interventions, OWL builds the theoretical and empirical backbone for trust in advanced AI. We work independently, openly, and rigorously to advance anticipatory security in complex sociotechnical systems.

We offer an assurance clinic that advises on mitigating risks in AI deployments of any kind - whether in industry or academia. Email [email protected] to book a free call.

Oxford Witt Lab is based in the Department of Engineering Science, Information Engineering Parks Road, Oxford OX1 3QD

Latest News

Multi-Agent Security Workshop @ DAI'25

OWL has successfully [co-hosted a multi-agent security workshop at DAI'25][https://wittlab.ai/event/masec_dai25]. The workshop took place in King’s College, London and surfaced new threat models and practitioner insights.

Decentralised AI is shifting from isolated agents to networks of interacting agents operating across shared platforms and protocols. This creates security challenges beyond traditional cybersecurity and single-agent safety, where free-form communication and tool use are essential for task generalisation yet open new system-level failure modes. These security vulnerabilities complicate attribution and oversight, and network effects can turn local issues into persistent, systemic risks (e.g., privacy leaks, jailbreak propagation, distributed attacks, or secret collusion). The workshop will address open challenges in multi-agent security as a discipline dedicated to securing interactions among agents, human–AI teams, and institutions—emphasising security–performance–coordination trade-offs, secure interaction protocols and environments, and monitoring/containment that remain effective under emergent behaviour. The main focus will lie on threat model discovery through community interaction.

Witt Lab PI wins Schmidt Sciences AI2050 Early Career Fellowship

The Oxford Witt Lab (OWL) PI Dr. Christian Schroeder de Witt has won a Schmidt AI2050 Early Career Fellowship as one of 21 awardees selected globally this year. See announcements by Forbes, the University of Oxford, and on the Schmidt Sciences website.

Schroeder de Witt receives the 3-year fellowship valued at $500,000 for defining the field of multi-agent security and, in particular, his research agenda on undetectable threats. In the coming years, powerful AI systems will work together, creating new security risks that may be practically infeasible—or even, in some regimes, theoretically impossible—to detect, undermining security approaches based on anomaly detection alone. Schroeder de Witt’s project explores how AI agents might exploit concealed capabilities, secretly share hidden messages, or carry out invisible attacks—and how to prevent them through secure-by-design architectures. The program blends theory and experiments to establish formal detectability limits and develop practical mitigations, including hardened interaction protocols, evaluation and red-team playbooks, and design patterns for resilient multi-agent systems. By acting now, the project aims to keep deployment practice ahead of emerging multi-agent threats.