Listen "AI Security - Training Data Attacks"
Episode Synopsis
analysis of training data poisoning, a critical integrity attack against AI and ML systems. It explains how adversaries corrupt the foundational learning phase by manipulating datasets, leading to altered model behavior, ranging from performance degradation to hidden backdoor attacks. The text highlights that large language models (LLMs) and generative AI are particularly vulnerable due to their reliance on vast, often unvetted internet data, and critically notes that larger models can paradoxically be more susceptible to learning malicious behaviors from minimal poisoned data. Finally, it outlines a multi-layered defense strategy, emphasizing data validation, robust model training, and strong operational security controls throughout the MLOps lifecycle, aligned with industry frameworks like NIST and OWASP.
More episodes of the podcast AI Intuition
Agent Builder by Docker
06/09/2025
AI Startup Failure Analysis
03/09/2025
AI Security - Model Denial of Service
02/09/2025
AI Security - Insecure Output Handling
02/09/2025
AI Security - Prompt Injection
02/09/2025
Supervised Fine-Tuning on OpenAI Models
31/08/2025
ZARZA We are Zarza, the prestigious firm behind major projects in information technology.