Reducing privacy leaks in artificial intelligence: two approaches to contextual integrity

New research from Microsoft Research presents two methods to reduce privacy leaks in artificial intelligence using the principle of contextual integrity. One method applies lightweight, inference-time checks and the other builds contextual awareness into models through reasoning and reinforcement learning.

Microsoft Research published new work that explores how to strengthen privacy safeguards for artificial intelligence agents by applying the concept of contextual integrity. The research frames privacy leaks as failures of contextual norms and investigates practical ways to align model behavior with those norms. The post highlights two distinct approaches developed or analyzed by the researchers, situating the effort as part of ongoing work to make models more sensitive to when and how private information should be shared.

The first approach described in the research uses lightweight, inference-time checks. These checks operate at the moment a model generates a response and act as an additional layer that evaluates whether a potential output would violate contextual privacy expectations. Because they are applied at inference time and are characterized as lightweight, they are presented as a way to add privacy safeguards without rebuilding underlying model architectures or retraining large systems.

The second approach integrates contextual awareness directly into models through explicit reasoning and reinforcement learning. Instead of relying on post hoc checks, this method aims to teach models to internalize contextual integrity during training or through reward-driven learning so that their outputs reflect privacy-aware behavior by design. The research thus places two different strategies side by side: one that supplements existing models at inference and one that seeks to embed contextual norms within model reasoning and learning dynamics.

55

Impact Score

Samsung starts sampling 3 GB GDDR7 running at 36 Gbps

Samsung has begun sampling its fastest-ever GDDR7 memory at 36 Gbps in 24 Gb dies that translate to 3 GB per chip, and it is also mass producing 28.0 Gbps 3 GB modules reportedly aimed at a mid-cycle NVIDIA refresh.

FLUX.2 image generation models now released, optimized for NVIDIA RTX GPUs

Black Forest Labs, the frontier Artificial Intelligence research lab, released the FLUX.2 family of visual generative models with new multi-reference and pose control tools and direct ComfyUI support. NVIDIA collaboration brings FP8 quantizations that reduce VRAM requirements by 40% and improve performance by 40%.

Aligning VMware migration with business continuity

Business continuity planning long focused on physical disasters, but cyber incidents, particularly ransomware, are now more common and often more damaging. In a survey of more than 500 CISOs, almost three-quarters (72%) said their organization had dealt with ransomware in the previous year.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.