Amazon signs inference chip supply deal with Cerebras

Amazon is partnering with Cerebras to supply inference processors, expanding its options beyond established semiconductor vendors and deepening its investment in specialized artificial intelligence hardware.

Amazon has entered into a new supply agreement with Cerebras to use the startup’s inference chips in its infrastructure, signaling a push to diversify beyond long-dominant graphics processors in artificial intelligence workloads. The deal focuses on processors optimized specifically for running trained artificial intelligence models in production, rather than for the initial training phase, and reflects rising demand for specialized silicon that can deliver lower latency and improved efficiency at scale.

Cerebras has built its strategy around very large single-wafer chips that integrate massive numbers of cores, high on-chip memory bandwidth and tightly coupled interconnects. By aligning with Amazon, Cerebras gains access to a major cloud and e-commerce platform that is seeking alternatives to incumbent suppliers for critical artificial intelligence infrastructure. The agreement highlights how hyperscale technology companies are exploring a broader mix of accelerators to balance performance, availability and cost as artificial intelligence applications proliferate across consumer and enterprise services.

The partnership also underscores intensifying competition in the semiconductor market for artificial intelligence inference, an area where power consumption, cost per query and hardware utilization are becoming as important as raw training throughput. For Amazon, adding Cerebras inference chips expands the portfolio of custom and third party silicon it can deploy to support internal products and cloud customers, while for Cerebras the relationship provides a high profile reference customer and potential validation of its architecture in large scale, real world workloads.

55

Impact Score

Nvidia launches nemotron 3 nano omni for enterprise agents

Nvidia has introduced Nemotron 3 Nano Omni, a multimodal open model designed to support enterprise agents that reason across vision, speech and language. The launch extends Nvidia’s push beyond hardware into models and services while targeting more efficient agentic workflows.

Intel 18A-P node improves performance and efficiency

Intel plans to present new results for its 18A-P process at the VLSI 2026 Symposium, highlighting gains in performance, power efficiency, and manufacturing predictability. The updated node is positioned as a stronger option for customers seeking 18A density with better operating characteristics.

EA CEO defends broader Artificial Intelligence use in game development

EA CEO Andrew Wilson defended the company’s internal use of Artificial Intelligence after employee claims that the tools were slowing work rather than helping. He framed the technology as an aid for repetitive quality assurance tasks, even as concerns persist over its broader impact on development.

Generative Artificial Intelligence is reshaping cybercrime less than feared

Research into criminal underground forums suggests generative Artificial Intelligence is being used mainly as a productivity tool rather than a transformative criminal breakthrough. The biggest near-term risks may come from automation, fraud support, and attackers adapting content to influence chatbot outputs.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.