EU Artificial Intelligence Office Issues Guidelines on AI Act and Data Protection

The EU´s Artificial Intelligence Office has released new guidelines clarifying the relationship between the Artificial Intelligence Act and existing European data protection laws.

The European Artificial Intelligence Office has published Guidelines 02 to address the intersection of the newly enacted Artificial Intelligence Act and the European Union´s robust data protection framework. The guidance aims to assist organizations, regulators, and developers in navigating the complexities arising from the coexistence of two major regulatory regimes. With the Artificial Intelligence Act set to become the first comprehensive legal framework for Artificial Intelligence systems globally, concerns regarding compliance with the General Data Protection Regulation (GDPR) and other privacy laws are of paramount importance for stakeholders across the continent.

The guidelines outline best practices for aligning Artificial Intelligence development and deployment with EU data protection standards, emphasizing lawful processing, data minimization, transparency, and accountability. The Office clarifies key obligations—including assessing the compatibility of training data with GDPR, ensuring comprehensive data subject rights, and managing risks related to automated decision-making. These rules are particularly relevant for high-risk Artificial Intelligence systems identified under the new Act, which are subject to strict conformity assessments, record-keeping, and human oversight requirements.

Legal experts indicate that the guidance provides much-needed clarity for transatlantic businesses and European startups who face uncertainty around the overlapping obligations of the Artificial Intelligence Act and GDPR. By addressing potential conflicts, the EU aims to encourage responsible innovation while safeguarding fundamental rights. The guidelines also underscore the role of Data Protection Authorities in monitoring Artificial Intelligence systems and highlight avenues for collaboration between national regulators and the new centralized Artificial Intelligence Office. These efforts are expected to influence similar regulatory developments in other jurisdictions and reinforce the EU´s role as a global standard-setter in technology governance.

73

Impact Score

Adaptive training method boosts reasoning large language model efficiency

Researchers have developed an adaptive training system that uses idle processors to train a smaller helper model on the fly, doubling reasoning large language model training speed without sacrificing accuracy. The method aims to cut costs and energy use for advanced applications such as financial forecasting and power grid risk detection.

How to run MiniMax M2.5 locally with Unsloth GGUF

MiniMax-M2.5 is a new open large language model optimized for coding, tool use, search, and office tasks, and Unsloth provides quantized GGUF builds and usage recipes for running it locally. The guide focuses on memory requirements, recommended decoding parameters, and deployment via llama.cpp and llama-server with an OpenAI-compatible interface.

Y Combinator backs new wave of computer vision startups in 2026

Y Combinator’s 2026 computer vision cohort spans infrastructure, developer tools, and industry-specific applications from retail security to aquaculture and healthcare. Startups are increasingly pairing computer vision with large vision language models and foundation models to tackle real-time video, automation, and domain-specific analysis.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.