Overview
- An ADL-affiliated study of more than 1,200 people who endorsed at least one antisemitic conspiracy found that chatting with a purpose-built LLM debunker cut belief levels and increased favorable views of Jews, with effects persisting a month later.
- Participants who used the DebunkBot showed larger shifts than control groups that either discussed unrelated topics or received a generic warning, though reductions were smaller among those endorsing multiple conspiracies; the team is seeking integration with major platforms.
- A separate report by researcher Julia Senkfor warns that leading LLMs, including GPT‑4o, Gemini, Claude and Llama, can generate antisemitic content due to structural biases, deliberate data poisoning and overreliance on heavily scraped sources such as Wikipedia.
- Senkfor cites findings that as few as 250 malicious documents can implant back-door behaviors and details extremist weaponization, including networks launching scores of hostile chatbots and AI-generated propaganda, as authorities probe incidents such as recent Grok responses in France.
- The report urges policy action that treats AI systems as products with liability, expands laws like the STOP HATE Act to cover AI, and empowers the FTC to investigate antisemitic outputs, as researchers say regulatory momentum remains weak.