Nvidia Microservices Target Humanoid Robot Development

Aug. 23, 2024
To accelerate humanoid robot development, Nvidia is providing robot manufacturers, AI model developers and software makers with a suite of services, models and computing platforms to develop, train and build the next generation of humanoid robots.
“The next wave of AI is robotics and one of the most exciting developments is humanoid robots,” said Jensen Huang, founder and CEO of Nvidia. “We’re advancing the entire Nvidia robotics stack, opening access for worldwide humanoid developers and companies to use the platforms, acceleration libraries and AI models best suited for their needs.”
 
Among the offerings are new Nvidia NIM microservices and frameworks for robot simulation and learning, the Nvidia Osmo orchestration service for running multi-stage robotics workloads, and an AI- and simulation-enabled teleoperation workflow that allows developers to train robots using small amounts of human demonstration data. 
 
The MimicGen NIM microservice generates synthetic motion data based on recorded teleoperated data from spatial computing devices like Apple Vision Pro. The Robocasa NIM microservice generates robot tasks and simulation-ready environments in OpenUSD, a universal framework for developing and collaborating within 3D worlds.
 
NVIDIA Osmo is a cloud-native managed service that allows users to orchestrate and scale complex robotics development workflows across distributed computing resources, whether on premises or in the cloud. According to Nvidia, Osmo simplifies robot training and simulation workflows, cutting deployment and development cycle times from months to less than a week. Users can visualize and manage a range of tasks, such as generating synthetic data, training models, conducting reinforcement learning and implementing software-in-the-loop testing at scale for humanoids, autonomous mobile robots and industrial manipulators.
 
An NVIDIA AI- and Omniverse-enabled teleoperation reference workflow allows researchers and AI developers to generate massive amounts of synthetic motion and perception data from a minimal amount of remotely captured human demonstrations. With this approach, Nvidia is seeking to minimize the costs and time typically required for teleoperation. This is seen as a key step in humanoid robot development as the building of training foundation models for humanoid robots requires an incredible amount of data. 
To reduce teleoperation time and costs, developers can use Apple Vision Pro to capture a small number of teleoperated demonstrations. Then, they simulate the recordings in Nvidia Isaac Sim and use the MimicGen NIM microservice to generate synthetic datasets from the recordings.
 
The developers train the Project GR00T humanoid foundation model with real and synthetic data, enabling developers to save time and reduce costs. Following this step, they use the Robocasa NIM microservice in Isaac Lab, a framework for robot learning, to generate experiences to retrain the robot model. Throughout the workflow, NVIDIA Osmo reportedly assigns computing jobs to different resources, saving the developers weeks of administrative tasks.

Sponsored Recommendations

Meet our experts - Reduce complexity of a DCS Migration

Sign up for a complementary onsite assessment.

Revolutionizing Germany’s energy landscape: The Wilhelmshaven floating LNG terminal

The German LNG terminal lays the groundwork for future sustainable energy initiatives. Here's how Schneider Electric helped make it happen.

Navigating Distributed Control Systems Migration

Navigating Distributed Control System (DCS) migrations doesn't have to be as complex as it seems. Whether you are planning a migration or seeking to enhance ...

Revolutionize process safety with Tricon CX V12

The most versatile TÜV certified safety instrumented system. One system with a choice of architectures and form factors delivering a lifetime of safe, resili...