At this year's NeurIPS conference, NVIDIA unveiled a new AI system aimed at accelerating the adoption of autonomous vehicles.
During the event in San Diego, the company introduced Alpamayo-R1 (AR1), which it describes as the world’s first industrial-scale, open-source reasoning Vision-Language-Action (VLA) model for autonomous driving.
VLA models are capable of processing both text and images simultaneously, enabling vehicle sensors to translate what they “see” into natural language descriptions.
NVIDIA’s software—named after a notoriously challenging peak in Peru’s Andes Mountains—combines chain-of-thought AI reasoning with trajectory planning. This allows the system to handle complex driving scenarios more effectively than previous autonomous software by breaking down scenes, evaluating all possible options, and then deciding on the best course of action—much like a human driver would.
NVIDIA emphasized that this capability is essential for achieving Level 4 autonomy, which the Society of Automotive Engineers defines as full vehicle control under specific conditions without human intervention.
In a blog post released alongside the AR1 announcement, Bryan Catanzaro, Vice President of Applied Deep Learning Research at NVIDIA, illustrated how the system operates in practice.
“By leveraging AR1’s chain-of-thought reasoning,” Catanzaro explained, “an autonomous vehicle navigating near a crowded bike lane can extract data from its path, generate reasoning traces that explain why certain actions are taken, and use that insight to plan future trajectories—such as veering away from the bike lane or stopping for a potential jaywalker.”
NVIDIA also noted that AR1’s human-like reasoning enhances performance in other nuanced situations, including busy intersections, lanes about to close, or double-parked vehicles blocking bike lanes.
By making its decision-making process transparent through explicit reasoning, AR1 gives engineers clearer insight into why specific choices are made—ultimately helping them design safer autonomous systems.
The model builds upon NVIDIA’s earlier Cosmos Reason framework, released earlier this year. Its open-access nature enables researchers to customize AR1 for non-commercial applications, whether for benchmarking or developing their own safety-focused tools.
AR1 is now available on GitHub and Hugging Face. According to Catanzaro, post-training reinforcement learning has proven “particularly effective,” with researchers reporting “significant improvements” in reasoning capabilities.