Red-Teaming in the Public Interest

Illustration
CLIENT
Data & Society Research Institute
PROJECT TYPE
Illustration
PROJECT YEAR
2025

Red-teaming genAI raises not only methodological questions — how and when to red-team, who should participate, how results should be used — but also thorny conceptual questions: whose interests are being protected? What counts as problematic model behavior, and who gets to define it? Is the public an object being secured, or a resource being used? In this report, Ranjit Singh, Borhane Blili-Hamelin, Carol Anderson, Emnet Tafesse, Briana Vecchione, Beth Duckles, and Jacob Metcalf offer a vision for red-teaming in the public interest: a process that goes beyond system-centric testing of already built systems to consider the full range of ways the public can be involved in evaluating genAI harms.

The premise of this illustration was to show hands from different backgrounds coming together to examine/evaluate a never ending cubic hole but using traditional tools to measure something that is not traditional and yet to be fully defined. This message is less about collaboration even when it is suggested and focuses on the challenges of red teaming.

Arrow pointing outside
No items found.

Other projects

Let's work together!

Thank you for visiting my portfolio. I'm excited to hear from you! Whether you have a question, a project in mind, or just want to say hello, please feel free to reach out. Fill out the form below and I'll get back to you as soon as possible. Looking forward to connecting with you!

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.