In 2023, I partnered with OpenAI, ScaleAI, & Hugging Face to launch HackAPrompt—the 1st & Largest Generative AI Red Teaming Competition ever held. Over 3,300 AI hackers competed to bypass model guardrails using prompt injections—the #1 Security Risk in LLMs.
The result? The Largest Dataset of Prompt Injection attacks ever collected—now used by every major Frontier AI Lab, including OpenAI, which used it to increase their models' resistance to Prompt Injection Attacks by up to 46%.
My research paper on HackAPrompt, Ignore This Title and HackAPrompt, was awarded Best Theme Paper at EMNLP 2023, one of the world's leading NLP conferences, selected out of 20,000 papers. Since then, OpenAI has cited it in three major research papers: Instruction Hierarchy, Automated Red Teaming, and Adversarial Robustness papers.
Today, I've delivered talks on HackAPrompt, Prompt Engineering, and AI Red Teaming at OpenAI, Stanford University, Dropbox, Deloitte, and Microsoft. And because I love to teach... I created this course so I can teach you everything I know about AI Red Teaming!
About the Course
This 10-day Crash Course is the #1 AI Security course for Cybersecurity Professionals, AI Trust & Safety leads, AI product managers, and engineers who want to master AI Red Teaming and secure AI systems against real-world threats. You'll gain hands-on experience identifying vulnerabilities in Generative AI systems, including prompt injections, jailbreaks, and adversarial attacks. Using the HackAPrompt playground, you'll practice both attacking and defending AI systems—learning how to break them and how to secure them.
This course is practical, not just theoretical. You'll work on real-world projects, analyzing live AI systems for vulnerabilities. These projects prepare you for the AI Red Teaming Certified Professional (AIRTP+) Exam, a 24+ hour assessment that validates your AI security expertise.
Our last cohort included 150 professionals from Microsoft, Google, Meta, Capital One, IBM, ServiceNow, and Walmart. Graduates passed the certification exam and are now AIRTP+ certified, applying their skills to secure AI systems worldwide.
What You'll Learn
Learn Advanced AI Red-Teaming Techniques: Gain hands-on experience with prompt injections, jailbreaking, and prompt hacking in the HackAPrompt playground. Learn to identify and exploit AI vulnerabilities, enhancing your offensive security skills to a professional level.
Design Secure AI Systems: Learn the principles of clarity and context during designing of prompts. Understand how to create robust AI systems that are resilient against attacks.
Develop Defensive Strategies: Learn to implement robust defense strategies against prompt injections and adversarial attacks. Secure AI/ML systems by building resilient models and integrating security measures throughout the AI development lifecycle.
Analyze Real-World AI Security Breaches: Study real-world AI security breaches to evaluate risks and develop effective prevention strategies. Gain insights into common vulnerabilities and learn how to mitigate future threats.
Future-Proof Your Career: Equip yourself with cutting-edge skills to stay ahead in the evolving tech landscape. Position yourself at the forefront of AI security, opening new career opportunities as AI transforms industries.
About the Instructor
I'm Sander Schulhoff, the Founder of Learn Prompting. In October 2022, I published the 1st Prompt Engineering Guide on the Internet—two months before ChatGPT was released. Since then, my courses have trained over 3 million people, and I'm one of only two people (other than Andrew Ng) to partner with OpenAI on a ChatGPT course. I've also led Prompt Engineering workshops at OpenAI, Microsoft, Stanford, Deloitte, and Dropbox.
I'm an award-winning Generative AI researcher from the University of Maryland and the youngest recipient of the Best Paper Award at EMNLP 2023, the leading NLP Conference, selected out of 20,000 submitted research papers from PhDs around the world. I've co-authored research with OpenAI, Scale AI, Hugging Face, Stanford, the U.S. Federal Reserve, and Microsoft.
I also created HackAPrompt, the first and largest Generative AI Red Teaming competition. Most recently, I led a team from OpenAI, Microsoft, Google, and Stanford on The Prompt Report—the most comprehensive study on Prompt Engineering to date. This 76-page survey analyzed over 1,500 academic papers, evaluating the effectiveness of prompting techniques, AI agents, and Generative AI applications.
Limited-Time Offer
Plus free access to Learn Prompting Plus (a $549 value): Gain immediate access to over 15 comprehensive courses—including this masterclass and additional courses in Prompt Engineering, Prompt Hacking, & AI/ML Red-Teaming (valued at $299), and a voucher for the Learn Prompting AI/ML Red-Teaming Certificate Exam (valued at $249).
LIMITED SPOTS AVAILABLE
We're keeping this class intentionally small and will cap it at 150 participants so that we can provide more personal attention to each of you to make sure you get the most out of the course. If you're unable to place your order and see the waitlist page, that means we sold out this cohort. If so, please join our waitlist to get notified when we release our next cohort.
Money-Back Guarantee
We genuinely want this course to be transformative for you. You can receive a full refund within 14 days after the course ends, provided you meet the completion criteria in our refund policy. We're confident in the value we provide and stand by our promise to help you level up your AI security expertise.
Interested in an enterprise license so your whole team or company can take the course? Please reach out directly to
[email protected]