Artificial intelligence testing draws key lessons from cybersecurity practices

Experts reveal how cybersecurity´s differentiated standards, partnerships, and red teaming shape responsible Artificial Intelligence testing and evaluation.

Generative Artificial Intelligence is accelerating the need for robust governance and effective evaluation strategies, prompting industry leaders and researchers to mine lessons from established domains like cybersecurity. In a recent Microsoft Research podcast, host Kathleen Sullivan convenes Professor Ciaran Martin, former head of the UK’s National Cyber Security Centre, and Tori Westerhoff, principal director of Microsoft’s Artificial Intelligence Red Team, to unpack how risk assessment, differentiated standards, and collaborative frameworks in cybersecurity can guide emerging Artificial Intelligence oversight.

Martin provides an in-depth look at cybersecurity´s structured approach to risk assessment, highlighting that organizations need tailored solutions based on their size and threat profile. He emphasizes the folly of imposing nation-state security demands on small businesses, instead advocating for differentiated, practical standards that avoid stifling innovation. Martin underscores that standards like those from NIST and ISO, though technically complex, boil down to clear risk identification, perimeter protection, intrusion detection, mitigation strategies, and recovery protocols. Public-private partnerships, according to Martin, are most effective when pragmatic and purpose-driven rather than merely bureaucratic; real progress arises when both sectors co-develop credible, manageable regulation.

Transitioning to Artificial Intelligence, Westerhoff shares insights from red teaming hundreds of generative Artificial Intelligence products, stressing that Artificial Intelligence security—like cybersecurity—is fundamentally a team sport. She draws parallels between the use of established frameworks in cybersecurity and efforts within Artificial Intelligence to develop a shared risk language and actionable methodologies. The Microsoft Artificial Intelligence Red Team acts as both an early warning system and an incubator for new evaluation strategies, embedding security thinking throughout product development. Westerhoff highlights the necessity for hyper-customized testing methods that account for the unique impacts and configurations of various Artificial Intelligence systems, and she points to the growing convergence of Artificial Intelligence with traditional cybersecurity standards as a way to anchor innovation in reliable foundations.

Both experts underscore that responsible Artificial Intelligence governance must balance innovation with public trust, ensuring that standards and red-teaming practices are applied thoughtfully across sectors and company sizes. As policymakers, technologists, and industry leaders grapple with the speed and complexity of generative Artificial Intelligence, practical lessons from cybersecurity—especially those focused on transparent collaboration, differentiated standards, and adaptive testing—provide an evolving blueprint for safer and more trustworthy Artificial Intelligence deployment.

74

Impact Score

ChatGPT Images adds thinking capability

OpenAI has upgraded ChatGPT Images with a new thinking mode that can search the internet, generate multiple images, and verify outputs before finalizing results. The update also improves text rendering, dense compositions, multilingual support, and style flexibility.

YouTube expands deepfake detection to Hollywood talent

YouTube is opening its likeness protection system to actors, athletes, musicians and creators beyond its own platform. The move gives public figures a way to flag and request removal of damaging Artificial Intelligence-generated replicas while YouTube weighs broader rules and possible future monetization.

Adobe plans outcome-based pricing for Artificial Intelligence agents

Adobe is positioning its Artificial Intelligence agents around performance-based pricing, charging only when the software completes useful work. The approach points to a more results-oriented model for selling generative Artificial Intelligence tools to business customers.

Tech firms commit billions to Artificial Intelligence infrastructure

Amazon, OpenAI, Nvidia, Meta, Google and others are signing increasingly large cloud, chip and data center agreements as demand for Artificial Intelligence infrastructure accelerates. The latest wave of deals spans investments, compute purchases, chip supply agreements and data center buildouts.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.