Red Teaming Latent Spaces & Protecting LLM apps

Experience Level: intermediate
Language: english

This presentation delves into the security challenges associated with Large Language Models (LLMs) implementations, and AI Engineering in general, drawing insights from the HackAPrompt paper “Ignore This Title and HackAPrompt: Exposing Systemic Vulnerabilities of LLMs through a Global Scale Prompt Hacking Competition” (Sander Schulhoff et al.). Exploring a set of attack vectors, and demonstrating how these vulnerabilities can be exploited. Subsequently, we will discuss security measures, frameworks, and services within the Python ecosystem designed to mitigate such threats. The session will feature practical examples illustrating both the execution of these attacks and the implementation of corresponding security measures, equipping attendees with a comprehensive understanding of safeguarding LLM applications.


Raul Pino

Computer Engineer with +10 years in software development, working on e-commerce, biotech, fintech, and AI. Nowadays, Product Labs Dev at Halborn, and previously Staff Engineer at Distro (YC 2024). My last outstanding project was a capstone report for the “Machine Learning Engineer Nanodegree” program at Udacity: “Ensemble of Generative Adversarial Networks as a Data Augmentation Technique for Alzheimer research” which I presented at PyCon Bolivia 2022, and PyCon Italia 2023.

raul-pino