Listen "Task-in-Prompt (TIP) adversarial attacks"
Episode Synopsis
Tune into our latest episode where we dive deep into Task-in-Prompt (TIP) adversarial attacks, a novel class of jailbreaks that cleverly embed sequence-to-sequence tasks within prompts to bypass LLM safety safeguards. We'll explore how these attacks successfully generate prohibited content across state-of-the-art models like GPT-4o and LLaMA 3.2, revealing critical weaknesses in current defense mechanisms. Discover why traditional safeguards, including keyword-based filters, often fail against these sophisticated, indirect exploits.
More episodes of the podcast Build Wiz AI Show
Adaptation of Agentic AI
26/12/2025
Career Advice in AI
22/12/2025
Leadership in AI Assisted Engineering
21/12/2025
AI Consulting in Practice
19/12/2025
Google - 5 days: Prototype to Production
19/12/2025
Google - 5 days: Agent Quality
18/12/2025
ZARZA We are Zarza, the prestigious firm behind major projects in information technology.