NVIDIA today announced the world’s first commercially available Level 2+ automated driving system, NVIDIA DRIVE AutoPilot, which integrates multiple breakthrough AI technologies that will enable supervised self-driving vehicles to go into production by next year.
At CES 2019, leading automotive suppliers Continental and ZF announced Level 2+ self-driving solutions based on NVIDIA DRIVE, with production starting in 2020.
As a Level 2+ self-driving solution, NVIDIA DRIVE AutoPilot uniquely provides both world-class autonomous driving perception and a cockpit rich in AI capabilities. Vehicle manufacturers can use it to bring to market sophisticated automated driving features — as well as intelligent cockpit assistance and visualisation capabilities — that far surpass today’s ADAS offerings in performance, functionality and road safety.
DRIVE AutoPilot integrates for the first time high-performance NVIDIA Xavier system-on-a-chip (SoC) processors and the latest NVIDIA DRIVE Software to process many deep neural networks (DNNs) for perception as well as complete surround camera sensor data from outside the vehicle and inside the cabin. This combination enables full self-driving autopilot capabilities, including highway merge, lane change, lane splits and personal mapping. Inside the cabin, features include driver monitoring, AI copilot capabilities and advanced in-cabin visualisation of the vehicle’s computer vision system.
DRIVE AutoPilot is part of the open, flexible NVIDIA DRIVE platform, which is being used by hundreds of companies worldwide to build autonomous vehicle solutions that increase road safety while reducing driver fatigue and stress on long drives or in stop-and-go traffic. The new Level 2+ system complements the NVIDIA DRIVE AGX Pegasus system that provides Level 5 capabilities for robotaxis.
DRIVE AutoPilot addresses the limitations of existing Level 2 ADAS systems, which a recent Insurance Institute for Highway Safety study showed offer inconsistent vehicle detections and poor ability to stay within lanes on curvy or hilly roads, resulting in a high occurrence of system disengagements where the driver abruptly had to take control.
Xavier SoC: Processing at 30 Teraops a Second
Central to NVIDIA DRIVE AutoPilot is the Xavier SoC, which delivers 30 trillion operations per second of processing capability. Architected for safety, Xavier has been designed for redundancy and diversity, with six types of processors and 9 billion transistors that enable it to process vast amounts of data in real time.
Xavier is the world’s first automotive-grade processor for autonomous driving and is in production today. Global safety experts have assessed its architecture and development process as suitable for building a safe product.
AI Inside and Out
The DRIVE AutoPilot software stack integrates DRIVE AV software for handling challenges outside the vehicle, as well as DRIVE IX software for tasks inside the car.
DRIVE AV uses surround sensors for full, 360-degree perception and features highly accurate localisation and path-planning capabilities. These enable supervised self-driving on the highway, from on-ramp to off-ramp. Going beyond basic adaptive cruise control, lane keeping and automatic emergency braking, its surround perception capabilities handle situations where lanes split or merge, and safely perform lane changes.
DRIVE AV also includes a diverse and redundant set of advanced DNN technologies that enable the vehicle to perceive a wide range of objects and driving situations, including DriveNet, SignNet, LaneNet, OpenRoadNet and WaitNet. This sophisticated AI software understands where other vehicles are, reads lane markings, detects pedestrians and cyclists, distinguishes different types of lights and their colours, recognises traffic signs and understands complex scenes.
In addition to providing precise localisation to the world’s HD maps for vehicle positioning on the road, DRIVE AutoPilot offers a new personal mapping feature called “My Route,” which remembers where you have driven and can create a self-driving route even if no HD map is available.
Within the vehicle, DRIVE IX intelligent experience software enables occupant monitoring to detect distracted or drowsy drivers and provide alerts or take corrective action if needed. It is also used to create intelligent user experiences, including the new ability for augmented reality. Displaying a visualisation of the surrounding environment sensed by the vehicle, as well as planned route, instils trust in the system.
For next-generation user experiences in the vehicle, the AI capabilities of DRIVE IX can also be used to accelerate natural language processing, gaze tracking or gesture recognition.
Adopted by Industry Leaders
Continental is developing a scalable and affordable automated driving architecture that will bridge from Premium Assist to future automated functionalities. It uses Continental’s portfolio of radar, lidar, camera and Automated Driving Control Unit technology powered by NVIDIA DRIVE.
ZF ProAI offers a unique modular hardware concept and open software architecture, utilising NVIDIA DRIVE Xavier processors and DRIVE Software....