
External
Generating Hate: Bias in Leading AI Models
ADL Center for Technology & Society
This report investigates how major large language models (LLMs) produce antisemitic and anti-Israel content, even in response to neutral prompts. The findings highlight systemic issues in training data, moderation gaps, and model deployment—raising concerns about the role of AI in amplifying bias at scale.
Program-led
Program-led
High School Education Toolkit
By: Sasha Litwin, Hunter Dunn, Andrew Chang, & Nicole Aaberg
A curated collection of educational materials designed to help high school students critically engage with media. This toolkit provides teachers with classroom-ready resources that emphasize media literacy, civic understanding, and responsible digital engagement.
Program-led
Political Coverage Aggregator
Rattlesnake Team: Andrew Quach, Kadon Chia, Jordan Rosen, Liam Riley, Max Young
A real-time election coverage tool inspired by NFL RedZone. This Chrome extension aggregates news, social media, and prediction markets, helping users track political updates while minimizing misinformation and cognitive overload.
Program-led
AI-Driven Bias Detection
Chiara Izzo | PAX Technologies
An AI-powered platform designed to help users identify misinformation, uncover media bias, and evaluate source credibility in real time. Built on large language models and multimodal input, this project focuses on increasing transparency and empowering informed media consumption.






