Artificial intelligence testing draws key lessons from cybersecurity practices

Experts reveal how cybersecurity´s differentiated standards, partnerships, and red teaming shape responsible Artificial Intelligence testing and evaluation.

Generative Artificial Intelligence is accelerating the need for robust governance and effective evaluation strategies, prompting industry leaders and researchers to mine lessons from established domains like cybersecurity. In a recent Microsoft Research podcast, host Kathleen Sullivan convenes Professor Ciaran Martin, former head of the UK’s National Cyber Security Centre, and Tori Westerhoff, principal director of Microsoft’s Artificial Intelligence Red Team, to unpack how risk assessment, differentiated standards, and collaborative frameworks in cybersecurity can guide emerging Artificial Intelligence oversight.

Martin provides an in-depth look at cybersecurity´s structured approach to risk assessment, highlighting that organizations need tailored solutions based on their size and threat profile. He emphasizes the folly of imposing nation-state security demands on small businesses, instead advocating for differentiated, practical standards that avoid stifling innovation. Martin underscores that standards like those from NIST and ISO, though technically complex, boil down to clear risk identification, perimeter protection, intrusion detection, mitigation strategies, and recovery protocols. Public-private partnerships, according to Martin, are most effective when pragmatic and purpose-driven rather than merely bureaucratic; real progress arises when both sectors co-develop credible, manageable regulation.

Transitioning to Artificial Intelligence, Westerhoff shares insights from red teaming hundreds of generative Artificial Intelligence products, stressing that Artificial Intelligence security—like cybersecurity—is fundamentally a team sport. She draws parallels between the use of established frameworks in cybersecurity and efforts within Artificial Intelligence to develop a shared risk language and actionable methodologies. The Microsoft Artificial Intelligence Red Team acts as both an early warning system and an incubator for new evaluation strategies, embedding security thinking throughout product development. Westerhoff highlights the necessity for hyper-customized testing methods that account for the unique impacts and configurations of various Artificial Intelligence systems, and she points to the growing convergence of Artificial Intelligence with traditional cybersecurity standards as a way to anchor innovation in reliable foundations.

Both experts underscore that responsible Artificial Intelligence governance must balance innovation with public trust, ensuring that standards and red-teaming practices are applied thoughtfully across sectors and company sizes. As policymakers, technologists, and industry leaders grapple with the speed and complexity of generative Artificial Intelligence, practical lessons from cybersecurity—especially those focused on transparent collaboration, differentiated standards, and adaptive testing—provide an evolving blueprint for safer and more trustworthy Artificial Intelligence deployment.

74

Impact Score

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.

Please check your email for a Verification Code sent to . Didn't get a code? Click here to resend