Why opinion on Artificial Intelligence is so divided

Stanford’s 2026 Artificial Intelligence Index highlights a sharp disconnect between rapid technical progress and uneven real-world performance. That split is also visible in how experts and the public judge Artificial Intelligence’s impact on jobs, the economy, and daily life.

Stanford’s 2026 Artificial Intelligence Index captures an industry moving quickly while remaining difficult to assess clearly. The report pairs signs of concentration and momentum with evidence of uneven capability. The US is pushing harder on Artificial Intelligence infrastructure than other countries, and it hosts 5,427 data centers (and counting). That’s more than 10 times as many as any other country. The hardware stack also depends heavily on a narrow manufacturing base, with TSMC fabricating almost every leading Artificial Intelligence chip and making the global Artificial Intelligence hardware supply chain dependent on one foundry in Taiwan.

The strongest theme is contradiction. Systems can post elite results in some narrow domains while failing at tasks that appear much simpler. The Stanford report notes that Google DeepMind’s top reasoning model, Gemini Deep Think, scored a gold medal in the International Math Olympiad but is unable to read analog clocks half the time. That disconnect helps explain why opinion on Artificial Intelligence feels so unstable, with the technology described at once as transformative, overhyped, threatening to jobs, and still prone to obvious mistakes.

A major divide runs between experts and the public. “AI experts and the general public view the technology’s trajectory very differently,” the authors of the AI Index write. “Assessing AI’s impact on jobs, 73% of U.S. experts are positive, compared with only 23% of the public, a 50 percentage point gap. Similar divides emerge with respect to the economy and medical care.” The report defines experts as US-based researchers who took part in Artificial Intelligence conferences in 2023 and 2024.

One reason for that gap may be that different groups encounter very different versions of the technology. People using the newest tools for coding, math, or research are seeing systems at their most capable, especially as model makers focus on technical tasks with clear right or wrong answers. Those products are also becoming commercially valuable, which drives further improvement. Outside those use cases, large language models remain unreliable and inconsistent, a pattern often described as the “jagged frontier.”

The result is that two people can talk about Artificial Intelligence while effectively referring to different experiences. A power user paying for top-tier tools and tracking the latest model releases may see dramatic progress, while someone relying on a free version for a more open-ended task may see little reason for confidence. Artificial Intelligence is advancing faster and performing better than many people realize, but it also remains weak in many areas that matter to everyday users.

51

Impact Score

Stanford report charts the rapid rise of Artificial Intelligence

Stanford’s 2026 index portrays Artificial Intelligence as a fast-improving technology spreading through business, education, and daily life faster than past computing shifts. It also highlights mounting pressure on energy, jobs, benchmarking, and regulation as governments and researchers struggle to keep pace.

ClearBank Europe gains MiCAR crypto-asset status in the Netherlands

ClearBank Europe has become the first Dutch credit institution to complete the MiCAR notification process and receive Crypto Asset Service Provider status. The move allows the bank to offer regulated digital asset services across the European Union, beginning with access to EURC and USDC through Circle’s Mint platform.

OpenAI pauses North Tyneside data centre plan

OpenAI has paused its Stargate UK scheme, including a planned data centre at Cobalt Park in North Tyneside, citing energy costs and regulation. Local and national figures say the setback highlights broader pressures facing the UK’s Artificial Intelligence infrastructure ambitions.

California courts set rules for generative Artificial Intelligence use

California’s superior courts are increasingly permitting generative Artificial Intelligence in court work, while courts and lawyers across the legal sector face rising pressure to prevent fabricated citations and other tool-related errors. Sanctions, new policies and technology adoption are advancing in parallel.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.