NVIDIA debuts open-source physical and digital AI models

NVIDIA has released new open-source AI models and tools designed to advance both digital and physical AI, particularly for autonomous driving and robotics.

The NVIDIA Drive Alpamayo-R1 is dubbed as the world’s first open reasoning Vision Language Action model for autonomous vehicle (AV) research. It integrates chain-of-thought AI reasoning with path planning to give AVs the common sense needed to safely navigate complex road scenarios, such as pedestrian-heavy intersections, taking a crucial step toward Level 4 autonomy.

To support the growing physical AI ecosystem, the open-source Cosmos Cookbook is a comprehensive guide for developers to customise Cosmos-based models.

New tools, such as the LidarGen world model, allow for the generation of realistic lidar data for AV simulation.

In digital AI, new models that bolster the developer toolkit include the MultiTalker Parakeet automatic speech recognition model that can understand multiple speakers in real-time, and new resources for AI safety, such as the Nemotron Content Safety Reasoning model and a synthetic audio dataset.