2025 Cohort

August 4-29, 2025 • London, UK • 20 Participants

AI Security Bootcamp 2025 Cohort

The inaugural AI Security Bootcamp brought together 20 researchers and engineers for four intensive weeks of training in security fundamentals, AI infrastructure security, and AI-specific threats. The program culminated in a week of capstone projects where participants explored cutting-edge security research.

Capstone Projects

MCP Protector Proxy

Andrew

A security layer for the Model Context Protocol (MCP) that provides tool inspection, permission controls, sanitization against prompt injection, and rate limiting.

TAR Poison

Chris

Data poisoning attacks on Support Vector Machines used in legal document discovery for Technology Assisted Review.

Chasseur

Diana & Leo

Catching misbehaving coding agents red-handed through local deployment of EDR to detect and mitigate suspicious command execution attempts.

Breaking Taylor Unswift

Ethan

Demonstrates weight extraction from Taylor series-obfuscated neural networks, undermining model exfiltration protections.

PicoCTF Evals

Irakli

Implementation of capture-the-flag evaluations using the inspect evals framework for AI security testing.

Autonomy caps for Agents governance

Jaeho

Policy framework proposing autonomy caps for AI agents in government systems with practical usage guidelines.

ML Supply Chain Attacks

Jakub

Proof-of-concept pickle file payload injection and MD5 hash collision attacks on ML model distribution.

Red-teaming Activation Monitors

Jord

Adversarial attacks on AI monitoring systems via spurious correlations and backdoor trigger coordination.

Watermark Stealing Replication

Juan

Replication of watermark stealing attacks on machine learning models with interactive Streamlit demo.

ML Supply Chain Vulnerabilities

Katie

Paper replication exploring attack scenarios in outsourced training and transfer learning.

Immune

Leo

Malware detection using static binary analysis features to identify malicious software.

Adversarial Attack Detector

Lorenzo

Using linear probes to detect and analyze adversarial attacks on neural networks.

Audio Attacks on Whisper

Rhita

Adversarial attacks on OpenAI's Whisper automatic speech recognition model.

Here Comes the Worm

Sam

AI security research exploring worm-like behavior in AI systems.