Microsoft Defender for Cloud adds granular alerts for securing Artificial Intelligence services and agents

Microsoft Defender for Cloud now exposes a detailed set of alerts tailored to threats against Azure Artificial Intelligence applications and agent-based systems, mapping each signal to specific attack tactics and severities. The alerts cover credential leaks, jailbreaks, phishing, access anomalies, wallet abuse and reconnaissance on both model deployments and agentic workloads.

Microsoft Defender for Cloud introduces a catalog of security alerts tailored to Azure Artificial Intelligence services, with coverage split between Artificial Intelligence applications and Artificial Intelligence agents. The alerts are designed to notify security operations teams about credential exposure, jailbreak attempts, phishing, anomalous access patterns and reconnaissance targeting models and agents. Each alert is aligned with specific mitre tactics such as Credential Access, Privilege Escalation, Defense Evasion, Execution, Reconnaissance, Impact and Collection, and carries a defined severity level from low to high.

For Artificial Intelligence applications, Defender for Cloud flags credential theft when credentials appear in generative Artificial Intelligence model responses, highlighting risks unique to generative workloads. Multiple jailbreak alerts focus on direct prompt injection attempts to manipulate system prompts and bypass safeguards, differentiating between cases blocked or only detected by Azure Responsible Artificial Intelligence Content Safety Prompt Shields. Phishing-focused alerts distinguish between corrupted applications or models that respond with known malicious URLs, user-originated phishing URLs, and unknown URL origins, mapping to Impact, Collection and data poisoning scenarios. Additional detections include suspicious user agents and Tor or other suspicious IP addresses accessing Artificial Intelligence resources, wallet attack patterns based on duplicate requests or anomalous traffic volumes, and access anomalies tied to shifts in user agents, IP ranges or authentication methods. Defender for Cloud also monitors anomalous tool invocation in applications connected to Azure OpenAI deployments, sensitive data anomalies in model responses, and reconnaissance patterns aimed at extracting system instructions or testing guardrails.

For Artificial Intelligence agents, Microsoft extends similar protections in preview for the Foundry Agent service. Jailbreak attempts against agents are detected or blocked by Prompt Shields, reflecting the same focus on prompt injection and safeguard bypass. Agent-specific alerts also watch for ascii smuggling attacks that hide instructions in invisible text, phishing URLs sent by users or produced in agent responses or tool outputs, and suspicious, anonymized or malicious user agent and IP access. Wallet attacks on agents are tracked through volume anomalies that diverge from historical usage. New detections highlight instruction prompt leakage, where attackers attempt to extract hidden system-level instructions, and agent reconnaissance behavior that can precede more invasive prompt injection or jailbreak attacks. Together these alerts provide a structured signal set that helps organizations monitor, triage and respond to threats across their Artificial Intelligence models, applications and agentic systems within Azure.

55

Impact Score

Google Vids opens free video generation to all Google users

Google has made Google Vids available to anyone with a Google account, adding free access to video generation with its latest models. The move expands Google’s end-to-end video workflow and increases pressure on rivals that charge for similar tools.

Court warns against chatbot legal advice in Heppner case

A federal court found that chats with a publicly available generative Artificial Intelligence tool were not protected by attorney-client privilege or the work-product doctrine. The ruling highlights litigation risks when executives or employees use chatbots for legal guidance without lawyer supervision.

Newsom orders California to weigh Artificial Intelligence harms in contract rules

Gov. Gavin Newsom has signed an executive order directing California agencies to account for potential Artificial Intelligence harms in state contracting while expanding approved use of generative tools across government. The move follows a dispute involving Anthropic and reflects a broader split between California and the Trump administration on Artificial Intelligence oversight.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.