Think SMART: How to optimize artificial intelligence factory inference performance

The Think SMART framework outlines how to optimize artificial intelligence inference at scale by balancing workload complexity, multidimensional performance and ecosystem considerations. The article highlights architecture, software and return-on-investment levers that AI factories can use to maximize tokens per watt and cost efficiency.

Gearing up for the gigawatt artificial intelligence data center age

Artificial intelligence factories are shifting data center design from web-serving to GPU‑dense facilities that require new networking and cooling approaches. NVIDIA presents NVLink, Quantum InfiniBand and Spectrum‑X as a layered strategy to scale GPUs inside racks and across clusters.

Inside Ukraine’s largest Starlink repair shop

A volunteer workshop in Lviv has become possibly the largest unofficial repair hub for Starlink terminals supporting Ukraine’s military, repairing or modifying thousands of units since 2022.

Google releases per-prompt energy data for Gemini artificial intelligence

Google published a technical report estimating the energy, water and carbon footprint of a text prompt to its Gemini artificial intelligence, with a median prompt using 0.24 watt-hours of electricity. The report provides a detailed breakdown of how that figure was calculated and highlights limits to its scope.

Ionstream.ai offers NVIDIA B200 bare metal for Artificial Intelligence workloads

Ionstream.ai is offering NVIDIA HGX B200 bare metal through its GPU-as-a-Service platform for Artificial Intelligence workloads. The article lists an ambiguous hourly rate token (´?.50 per hour´) for one-month contracts and says lower pricing is available for longer commitments; the precise hourly price is not stated.