Discover how fine-tuning poisoning can strip LLM safety measures without compromising performance. Dive into the BadGPT attack, a novel approach that bypasses guardrails, avoids token overhead, and retains model efficiency. Learn why securing LLMs is an ongoing challenge in AI alignment.
Learn for free, join the best tech learning community for a price of a pumpkin latte.
Event notifications, weekly newsletter
Delayed access to all content
Immediate access to Keynotes & Panels
Access to Circle community platform
Immediate access to all content
Courses, quizes & certificates
Community chats