Lessons from red teaming a generative newsbot

Exploring how red teaming reveals the real risks and tradeoffs in deploying Artificial Intelligence-powered news chatbots.

As Artificial Intelligence technology becomes embedded in newsroom workflows and user-facing products, news organizations are accelerating adoption to maintain relevance in a fragmented information landscape. From aiding in SEO headline generation to answering real-time questions for readers, Artificial Intelligence systems are now commonplace not just in editorial backrooms but as interactive web tools. The article notes that while such systems promise efficiency and cost-savings, they also risk reshaping how users perceive authority, bias, and trust in news—pressures amplified as search engines and digital assistants increasingly aggregate and synthesize content instead of referring users directly to publishers.

The University of Pittsburgh’s collaboration with Spotlight PA to red-team its Election Assistant offers a rare inside look at these challenges. Spotlight PA´s system, intentionally constrained with predefined Q&A pairs rather than open-ended generative capabilities, showed that risk mitigation can significantly reduce both incorrect answers and the system´s helpfulness. The more tightly a bot is limited to safeguard accuracy and prevent harmful outputs, the less capable it becomes of meeting diverse user needs—a dilemma developers of newsroom Artificial Intelligence tools must confront, especially for high-stakes topics like elections. This fundamental tension—between precision and flexibility—mirrors wider industry debates about the helpfulness versus harmfulness tradeoff inherent to Artificial Intelligence deployment in sensitive settings.

The authors recommend several best practices for mitigating risk in Artificial Intelligence-powered news assistants. Proactive user education on the chatbot’s purpose and constraints can set expectations and improve query effectiveness. How a bot handles ´non-answers´—refusals or redirections—strongly influences user trust; clear, transparent explanations are vital to avoid raising suspicions of bias. Accurate translation is necessary to meet access and equity goals but demands careful human oversight to avoid misinforming user groups. Most critically, red teaming should not only check for factuality but also simulate the perspectives of skeptical or confused users, as seemingly neutral or evasive answers can unintentionally erode trust in both the tool and the underlying news organization.

The authors argue that concerns about hallucinated or biased outputs do not fully define the risk landscape. Since Artificial Intelligence mediates how audiences experience and emotionally interact with the news, emergent risks include shifting user attitudes, changes in behavior, and ambiguous intentions—especially as users may try to ´break´ or manipulate bots into misbehaving. Effective risk reduction may require greater restrictions, as seen in Spotlight PA, but doing so may hamper the tool’s utility in dynamic real-world demand. The persistent unpredictability of user behavior and intent makes traditional system audits insufficient. Continuous research into user experience and intention is urged to ensure new Artificial Intelligence-driven interfaces are trustworthy and effective in the evolving information ecosystem.

72

Impact Score

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.

Please check your email for a Verification Code sent to . Didn't get a code? Click here to resend