NVIDIA pushes physical Artificial Intelligence with Omniverse and OpenUSD

NVIDIA used GTC to position simulation, digital twins and synthetic data pipelines as core infrastructure for physical Artificial Intelligence. New models, blueprints and partner deployments show how robots, vehicles and factories are moving from isolated pilots to broader enterprise systems.

NVIDIA presented GTC as a milestone for physical Artificial Intelligence, with robots, vehicles and factories moving beyond single use cases and isolated deployments into broader enterprise workloads. The company highlighted frontier models including NVIDIA Cosmos 3, NVIDIA Isaac GR00T N1.7 and NVIDIA Alpamayo 1.5, alongside new infrastructure intended to support world modeling, humanoid skills and autonomous driving. OpenUSD sits at the center of that strategy as a common scene-description language for combining CAD data, simulation assets and real-world telemetry in a shared, physically accurate environment.

A major focus was simulation for industrial systems before they are built or deployed. NVIDIA introduced the Omniverse DSX Blueprint as a reference architecture for creating a single digital twin across an Artificial Intelligence factory, covering thermals, power grids, network load and mechanical systems. NVIDIA also introduced its Physical Artificial Intelligence Data Factory Blueprint, an open reference architecture built on NVIDIA Cosmos open world foundation models and the NVIDIA OSMO operator. The system unifies data curation, augmentation and evaluation into a single pipeline, enabling developers to create diverse, long-tail datasets from limited real-world inputs. Microsoft Azure and Nebius are the first cloud platforms to offer the blueprint.

NVIDIA framed compute as the new engine for producing training data, arguing that real-world data alone no longer scales for physical Artificial Intelligence. Open source agentic frameworks such as OpenClaw extend that stack into operations by running long-lived workflows with tools, memory and messaging interfaces. On the design side, NVIDIA emphasized CAD-to-OpenUSD pipelines using Omniverse Kit and Isaac Sim to turn engineering data into simulation-ready assets for real-time rendering, testing and collaboration. FANUC and Fauna Robotics are using this workflow to accelerate robotic system design and validation.

The company also linked digital twins directly to manufacturing and logistics deployment. The NVIDIA Mega Omniverse Blueprint is positioned as a way to design, test and optimize robot fleets and Artificial Intelligence agents in facility-scale digital twins before deployment. KION, with Accenture and Siemens, is using the blueprint to build warehouse digital twins for GXO and train NVIDIA Jetson-based autonomous forklifts. NVIDIA said ABB Robotics, FANUC, KUKA and Yaskawa, which have a combined global install base of over 2 million robots, are using Omniverse libraries and Isaac simulation frameworks to validate applications and production lines through digital twins, while integrating NVIDIA Jetson modules for real-time Artificial Intelligence inference. FieldAI, Skild AI and Generalist AI were also highlighted for using NVIDIA Cosmos and Isaac-based simulation to develop robot brains and synthetic data workflows.

68

Impact Score

AMD plans specialized EPYC CPUs for Artificial Intelligence, hpc, and cloud

AMD is preparing a broader EPYC strategy with task-specific server CPUs aimed at agentic Artificial Intelligence, hpc, training and inference, and cloud deployments. The shift starts with the Zen 6 generation and adds Verano as an Artificial Intelligence-focused variant within the same EPYC family.

Nvidia expands spectrum-x ethernet with open mrc protocol

Nvidia is positioning Spectrum-X Ethernet as a foundation for large-scale Artificial Intelligence training, with Multipath Reliable Connection adding open, multi-path RDMA transport for higher resilience and throughput. OpenAI, Microsoft and Oracle are among the organizations using the technology in large Artificial Intelligence environments.

Anthropic explores Fractile chips to diversify supply

Anthropic is reportedly in early talks with London-based Fractile to secure high-performance Artificial Intelligence chips for inference workloads. The move would reduce reliance on Nvidia and broaden the company’s hardware supply chain.

OpenAI curbs odd creature references in chatbot responses

OpenAI has adjusted its models after users complained about overly familiar responses and strange references to goblins, gremlins, pigeons, and raccoons. The company traced the behavior to a retired “nerdy” personality whose habits spread into broader model training.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.