AI Security Bootcamp

4-week intensive programLondon, UK (In-person)Fully funded

August 4–29, 2025 • 20 Participants

AI Security Bootcamp 2025 Cohort

The inaugural AI Security Bootcamp brought together 20 researchers and engineers for four intensive weeks of training in security fundamentals, AI infrastructure security, and AI-specific threats. The program culminated in a week of capstone projects where participants explored cutting-edge security research.

The curriculum is available on GitHub for self-study.

Curriculum

Week 1: Security Fundamentals

  • Fundamentals of Cryptography: Stream ciphers (LCG), block ciphers (DES, SPN), hashing (MD5), message authentication (HMAC), public-key cryptography (RSA), cryptanalysis (crib dragging, meet-in-the-middle, padding oracle attacks)
  • Network Security: Traffic analysis with Wireshark, HTTP/HTTPS man-in-the-middle interception, certificate pinning, NFQUEUE, covert channels (DNS, ICMP)
  • Threat Modeling: STRIDE methodology, attack trees, adversary capability modeling
  • Penetration Testing: Network reconnaissance, enumeration, password cracking, Metasploit exploitation, privilege escalation, persistence

Week 2: Infrastructure Security

  • Containerization: Container fundamentals (chroot, cgroups, namespaces), network isolation, container escapes, syscall monitoring
  • Supply Chain Security: Pickle deserialization attacks, dependency confusion, model provenance
  • Reverse Engineering: Ghidra, buffer overflow, crafting shellcode exploits, bypassing stack canaries
  • Application & Cloud Security: XSS, CSRF, SSRF, SQL injection, command injection (OWASP Top 10), cloud Identity and Access Management

Week 3: AI-Specific Security

  • Adversarial ML: Crafting adversarial examples, attacks on image classifier, watermarking, trojans
  • LLM Security: Tokenization, prompt injection, model weight extraction attacks, model editing, abliteration
  • GPU & Datacenter Security: Nvidia container toolkit exploits, GPU isolation, confidential computing
  • AI Application Security: MCP (Model Context Protocol) security, RAG injection, hardware supply chain

Week 4: Capstone Project

  • Implement novel security solutions, replicate sophisticated attacks, or conduct security research
  • Work with expert mentors on cutting-edge AI security challenges
  • Present findings to cohort and industry professionals

Who was this for?

AISB was designed for researchers and engineers who care about securing the development of AI systems. Ideal participants had prior experience with deep learning (training/evals) and were comfortable with Python.

The program ran full-time, in-person in London. All expenses including flights, accommodation, and meals were covered.

Capstone Projects

MCP Protector Proxy

Andrew

A security layer for the Model Context Protocol (MCP) that provides tool inspection, permission controls, sanitization against prompt injection, and rate limiting.

TAR Poison

Chris

Data poisoning attacks on Support Vector Machines used in legal document discovery for Technology Assisted Review.

Chasseur

Diana & Leo

Catching misbehaving coding agents red-handed through local deployment of EDR to detect and mitigate suspicious command execution attempts.

Breaking Taylor Unswift

Ethan

Demonstrates weight extraction from Taylor series-obfuscated neural networks, undermining model exfiltration protections.

PicoCTF Evals

Irakli

Implementation of capture-the-flag evaluations using the inspect evals framework for AI security testing.

Autonomy caps for Agents governance

Jaeho

Policy framework proposing autonomy caps for AI agents in government systems with practical usage guidelines.

ML Supply Chain Attacks

Jakub

Proof-of-concept pickle file payload injection and MD5 hash collision attacks on ML model distribution.

Red-teaming Activation Monitors

Jord

Adversarial attacks on AI monitoring systems via spurious correlations and backdoor trigger coordination.

Watermark Stealing Replication

Juan

Replication of watermark stealing attacks on machine learning models with interactive Streamlit demo.

ML Supply Chain Vulnerabilities

Katie

Paper replication exploring attack scenarios in outsourced training and transfer learning.

Immune

Leo

Malware detection using static binary analysis features to identify malicious software.

Adversarial Attack Detector

Lorenzo

Using linear probes to detect and analyze adversarial attacks on neural networks.

Audio Attacks on Whisper

Rhita

Adversarial attacks on OpenAI's Whisper automatic speech recognition model.

Here Comes the Worm

Sam

AI security research exploring worm-like behavior in AI systems.

Team

Pranav Gade

Pranav Gade

Research engineer at Conjecture. Created AISB to bridge AI safety and security, and leads curriculum design and program direction.

Jan Michelfeit

Jan Michelfeit

Security lead at Conjecture. Designs AISB's hands-on labs and capstone projects, drawing on 10+ years securing complex systems and ML infrastructure.

Nitzan Shulman

Nitzan Shulman

Head of Cyber Security at Heron AI Security Initiative. 6+ years doing security research specializing in IOT, Robotics, Malware and AI security.

Jinglin Li

Jinglin Li

Software engineer and educator. Keeps AISB running smoothly.


Acknowledgments

This program was supported by Open Philanthropy

Open Philanthropy Logo