California delays its Artificial Intelligence Transparency Act and passes new content laws

California enacted AB 853, pushing the Artificial Intelligence Transparency Act’s start date to August 2, 2026, and adding new disclosure and detection duties for generative Artificial Intelligence providers, large platforms, and device makers. Platforms face standardized source data checks and latent disclosures in 2027, with capture devices offering similar options in 2028.

On October 13, the governor of California signed AB 853 to amend the compliance date of the California Artificial Intelligence Transparency Act and to add new obligations for large online platforms. The amendment moves the act’s effective date from January 1, 2026, to August 2, 2026. It also requires covered providers that create, code or produce a generative Artificial Intelligence system to make available a free Artificial Intelligence detection tool that helps users determine whether media content was created or altered by the provider’s generative Artificial Intelligence system, and that outputs any system source data detected in the content.

Beginning January 1, 2027, large online platforms face additional responsibilities. The law defines these platforms as public-facing social media, file-sharing, mass messaging platforms, or stand-alone search engines that distribute content to users who did not create it and that exceeded 2 million unique monthly users during the preceding 12 months. These platforms must detect whether source data, compliant with widely adopted standards, is embedded in or attached to the content they distribute. They must then provide users with a latent disclosure that indicates whether the content was generated or altered by a generative Artificial Intelligence system.

Starting January 1, 2028, capture device manufacturers must offer users the option to include a latent disclosure in content captured by devices first produced for sale in California on or after that date. The required disclosure must communicate the manufacturer’s name, the device name and version, and the time and date when the content was created or altered. Together, the delayed timeline and new requirements establish a phased approach that begins with generation-layer detection tools from providers, extends to platform-level source data checks and user disclosures, and reaches device-level options for embedding standardized disclosures in newly produced hardware.

68

Impact Score

What businesses need to know about the EU cyber resilience act

The EU cyber resilience act is turning product cybersecurity into a legal requirement for companies that sell digital products into the European Union. A key compliance milestone arrives in September 2026, well before the full regulation takes effect in 2027.

Claude Mythos and cyber insurance’s next inflection point

Claude Mythos is being treated by governments and regulators as a potential systemic cyber risk with implications for financial stability and insurance markets. Its emergence is intensifying pressure on insurers to clarify whether Artificial Intelligence-enabled cyber losses are covered, excluded, or require new stand-alone products.

OpenAI expands ChatGPT ads with self-serve manager

OpenAI is widening its ChatGPT ads pilot with a beta self-serve Ads Manager, new bidding options and broader measurement tools. The push signals a deeper move into advertising as the company expands the program into several international markets.

OpenAI launches Artificial Intelligence deployment consulting unit

OpenAI has created a new consulting and deployment business aimed at helping enterprises build and roll out Artificial Intelligence systems. The move mirrors a similar push by Anthropic and signals a broader effort by model providers to capture more of the enterprise services market.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.