Amazon Devices advances zero-touch manufacturing with NVIDIA digital twins

Amazon Devices & Services is using NVIDIA digital twin technologies and Artificial Intelligence to train robotic arms on synthetic data, enabling zero-touch manufacturing and faster product integration.

Amazon Devices & Services has deployed a simulation-first manufacturing solution this month that pairs Amazon-created software with NVIDIA digital twin technologies to move toward zero-touch production. The system trains robotic arms to inspect a range of products for quality auditing and to integrate new goods onto the line without hardware changes. Amazon´s approach relies heavily on synthetic data and a modular workflow, aiming to replace time-consuming physical prototyping with virtual, repeatable experiments.

The technical pipeline stitches together several NVIDIA stacks and Amazon cloud services. Amazon imports CAD models into NVIDIA Isaac Sim on the Omniverse platform, then generates more than 50,000 synthetic images per device to train object- and defect-detection models. NVIDIA Isaac ROS produces robotic trajectories, while cuMotion and the nvblox library support fast collision-free planning on NVIDIA Jetson AGX Orin modules. FoundationPose, a foundation model trained on 5 million synthetic images, provides pose estimation and object tracking that can generalize to unseen objects. On the cloud side, AWS accelerated model development using Amazon EC2 G6 instances and AWS Batch; Amazon Bedrock and Bedrock AgentCore handle higher-level task planning and ingest multimodal product specifications.

That collection of tools is designed to enable what Amazon describes as ´zero-shot manufacturing´ and a broader move to generalized manufacturing. Lines can switch from auditing one product to another with software updates alone. The modular design already supports defect detection during production and is built to integrate more advanced reasoning components in the future, including NVIDIA Cosmos Reason for deeper analysis and decision-making. Eliminating the need for physical prototypes reduces cost and shortens the time it takes to bring new devices to consumers.

The deployment at an Amazon Devices facility demonstrates a path from simulation to the real world, with simulated stations matching real ones for training and validation. The result is faster on-boarding of new products, more flexible robotic operations, and a clearer route to autonomous, software-driven manufacturing pipelines that scale across products and stations.

76

Impact Score

Chip giants back Ayar Labs to push optical interconnects for Artificial Intelligence

Ayar Labs has attracted investments from Nvidia, AMD, Intel, MediaTek and major funds by promising optical interconnects that tackle bandwidth, latency and power bottlenecks in Artificial Intelligence data centers. Its TeraPHY and SuperNova platform combines silicon photonics with open chiplet standards to link accelerators over distances from millimeters to kilometers.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.