Hacking AI Guardrails Workshop (LISA) by Thomas Vissers and Tim Van Hamme
09:00 - 12:30
Abstract:
In this workshop, participants will get hands-on experience both building and breaking AI guardrails. Working in teams, you’ll design guardrail policies to defend sensitive data within a vulnerable AI application, while attempting to bypass those of others. Who will win? The scoreboard will tell, but in the end everyone walks away with a deeper grasp of how LLMs can be manipulated and the challenges of building robust guardrails.
This workshop is delivered as part of the VLAIO research project LISA.
Bio:
Thomas Vissers specializes in the dynamic intersection of cybersecurity and AI. He packs 10 years of experience in developing and operating production-grade AI security systems. Previously at Cloudflare, he served as an engineering leader for data-driven security products, scaling innovations to protect millions of internet users worldwide. Today at KU Leuven, Thomas focuses on empowering organizations to securely and reliably leverage LLM technology.
Tim Van hamme is a computer scientist specialized in AI security and behavioral analysis, leading research at KU Leuven's DistriNet on secure and trustworthy AI adoption. As co-lead of Blue41, he bridges rigorous academic research with industry applications, helping organizations safely deploy AI systems. With experience spanning from CERT.be's threat intelligence infrastructure to behavioral biometrics and AI agent security, Tim combines theoretical expertise with practical solutions for emergent AI security challenges.
Requirements:
Requirements: laptop that can SSH into external machine.
Optional: running Python locally and an IDE.