Artificial intelligence testing draws key lessons from cybersecurity practices

Experts reveal how cybersecurity´s differentiated standards, partnerships, and red teaming shape responsible Artificial Intelligence testing and evaluation.

Generative Artificial Intelligence is accelerating the need for robust governance and effective evaluation strategies, prompting industry leaders and researchers to mine lessons from established domains like cybersecurity. In a recent Microsoft Research podcast, host Kathleen Sullivan convenes Professor Ciaran Martin, former head of the UK’s National Cyber Security Centre, and Tori Westerhoff, principal director of Microsoft’s Artificial Intelligence Red Team, to unpack how risk assessment, differentiated standards, and collaborative frameworks in cybersecurity can guide emerging Artificial Intelligence oversight.

Martin provides an in-depth look at cybersecurity´s structured approach to risk assessment, highlighting that organizations need tailored solutions based on their size and threat profile. He emphasizes the folly of imposing nation-state security demands on small businesses, instead advocating for differentiated, practical standards that avoid stifling innovation. Martin underscores that standards like those from NIST and ISO, though technically complex, boil down to clear risk identification, perimeter protection, intrusion detection, mitigation strategies, and recovery protocols. Public-private partnerships, according to Martin, are most effective when pragmatic and purpose-driven rather than merely bureaucratic; real progress arises when both sectors co-develop credible, manageable regulation.

Transitioning to Artificial Intelligence, Westerhoff shares insights from red teaming hundreds of generative Artificial Intelligence products, stressing that Artificial Intelligence security—like cybersecurity—is fundamentally a team sport. She draws parallels between the use of established frameworks in cybersecurity and efforts within Artificial Intelligence to develop a shared risk language and actionable methodologies. The Microsoft Artificial Intelligence Red Team acts as both an early warning system and an incubator for new evaluation strategies, embedding security thinking throughout product development. Westerhoff highlights the necessity for hyper-customized testing methods that account for the unique impacts and configurations of various Artificial Intelligence systems, and she points to the growing convergence of Artificial Intelligence with traditional cybersecurity standards as a way to anchor innovation in reliable foundations.

Both experts underscore that responsible Artificial Intelligence governance must balance innovation with public trust, ensuring that standards and red-teaming practices are applied thoughtfully across sectors and company sizes. As policymakers, technologists, and industry leaders grapple with the speed and complexity of generative Artificial Intelligence, practical lessons from cybersecurity—especially those focused on transparent collaboration, differentiated standards, and adaptive testing—provide an evolving blueprint for safer and more trustworthy Artificial Intelligence deployment.

74

Impact Score

Saudi Artificial Intelligence startup launches Arabic LLM

Misraj Artificial Intelligence unveiled Kawn, an Arabic large language model, at AWS re:Invent and launched Workforces, a platform for creating and managing Artificial Intelligence agents for enterprises and public institutions.

Introducing Mistral 3: open artificial intelligence models

Mistral 3 is a family of open, multimodal and multilingual Artificial Intelligence models that includes three Ministral edge models and a sparse Mistral Large 3 trained with 41B active and 675B total parameters, released under the Apache 2.0 license.

NVIDIA and Mistral Artificial Intelligence partner to accelerate new family of open models

NVIDIA and Mistral Artificial Intelligence announced a partnership to optimize the Mistral 3 family of open-source multilingual, multimodal models across NVIDIA supercomputing and edge platforms. The collaboration highlights Mistral Large 3, a mixture-of-experts model designed to improve efficiency and accuracy for enterprise artificial intelligence deployments starting Tuesday, Dec. 2.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.