From 4e3f2d741e884c4bddba9d8fc5dfa58e648b01bf Mon Sep 17 00:00:00 2001 From: Andrew Freeborn <19598523+afreeborn@users.noreply.github.com> Date: Tue, 15 Jul 2025 19:17:27 -0500 Subject: [PATCH] Update README.md (#25) --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 9be0b29..4dc2524 100644 --- a/README.md +++ b/README.md @@ -2,7 +2,7 @@ This repository contains the challenges for the labs used in the course "AI Red Teaming in Practice". The course was originally taught at [Black Hat USA 2024](https://www.blackhat.com/us-24/training/schedule/index.html#ai-red-teaming-in-practice-37464) by [Dr. Amanda Minnich](https://www.amandaminnich.info/) and [Gary Lopez](https://www.linkedin.com/in/gary-l-76501814a). [Martin Pouliot](https://www.linkedin.com/in/martin-pouliot-266ab0105/) handled the infrastructure and scoring for the challenges. The challenges were designed by Dr. Amanda Minnich, Gary Lopez and Martin Pouliot. These challenges are available for anyone to use. The playground environment is based on [Chat Copilot](https://github.com/microsoft/chat-copilot) and was modified to be used in the course. -These challenges are also referenced in the [Microsoft Learn Limited Series: AI Red Teaming 101](https://review.learn.microsoft.com/en-us/security/ai-red-team/training), released on July 9, 2025. In the table of challenges below, you will find the relevant video link to each challenge from the series. During Microsoft Build in May 2025, several of these challenges were automated by the [Python Risk Identification Tool (PyRIT)](https://aka.ms/pyrit), which is an open-source framework built to empower security professionals and engineers to proactively identify risks in generative AI systems. This repository contains the [corresponding Jupyter Notebooks](https://github.com/microsoft/AI-Red-Teaming-Playground-Labs/tree/main/notebooks) showcasing how to use PyRIT to solve the challenges for Labs 1 and 5. You will also see a notebook for "Lab 13", which does not have an associated challenge as it is notebook-only. +These challenges are also referenced in the [Microsoft Learn Limited Series: AI Red Teaming 101](https://learn.microsoft.com/en-us/security/ai-red-team/training), released on July 9, 2025. In the table of challenges below, you will find the relevant video link to each challenge from the series. During Microsoft Build in May 2025, several of these challenges were automated by the [Python Risk Identification Tool (PyRIT)](https://aka.ms/pyrit), which is an open-source framework built to empower security professionals and engineers to proactively identify risks in generative AI systems. This repository contains the [corresponding Jupyter Notebooks](https://github.com/microsoft/AI-Red-Teaming-Playground-Labs/tree/main/notebooks) showcasing how to use PyRIT to solve the challenges for Labs 1 and 5. You will also see a notebook for "Lab 13", which does not have an associated challenge as it is notebook-only. ## What You Will Learn These challenges are designed to teach security professionals to systematically red team AI systems. They go beyond traditional security failures by incorporating novel adversarial machine learning and Responsible AI (RAI) failures, enabling a holistic approach to identifying potential issues before an AI system is deployed.