[Databite No. 161] Red Teaming Generative AI Harm


Episode Artwork
1.0x
0% played 00:00 00:00
Mar 03 2025 60 mins   4

What exactly is generative AI (genAI) red-teaming? What strategies and standards should guide its implementation? And how can it protect the public interest? In this conversation, Lama Ahmad, Camille François, Tarleton Gillespie, Briana Vecchione, and Borhane Blili-Hamelin examined red-teaming’s place in the evolving landscape of genAI evaluation and governance.

Our discussion drew on a new report by Data & Society (D&S) and AI Risk and Vulnerability Alliance (ARVA), a nonprofit that aims to empower communities to recognize, diagnose, and manage harmful flaws in AI. The report, Red-Teaming in the Public Interest, investigates how red-teaming methods are being adapted to confront uncertainty about flaws in systems and to encourage public engagement with the evaluation and oversight of genAI systems. Red-teaming offers a flexible approach to uncovering a wide range of problems with genAI models. It also offers new opportunities for incorporating diverse communities into AI governance practices.

Ultimately, we hope this report and discussion present a vision of red-teaming as an area of public interest sociotechnical experimentation.

Download the report and learn more about the speakers and references at datasociety.net.

--

00:00 Opening

00:12 Welcome and Framing

04:48 Panel Introductions

09:34 Discussion Overview

10:23 Lama Ahmad on The Value of Human Red-Teaming

17:37 Tarleton Gillespie on Labor and Content Moderation Antecedents

25:03 Briana Vecchione on Participation & Accountability

28:25 Camille François on Global Policy and Open-source Infrastructure

35:09 Questions and Answers

56:39 Final Takeaways