At CES 2019, NVIDIA announced the first commercially available Level 2+ automated driving system, NVIDIA DRIVE AutoPilot, which integrates multiple breakthrough AI technologies that will enable supervised self-driving vehicles to go into production by next year. At CES 2019, Continental and ZF announced Level 2+ self-driving solutions based on NVIDIA DRIVE, with production starting in 2020.
As a Level 2+ self-driving solution, NVIDIA DRIVE AutoPilot provides both world-class autonomous driving perception and a cockpit rich in AI capabilities. Vehicle manufacturers can use it to bring to market sophisticated automated driving features as well as intelligent cockpit assistance and visualization capabilities that surpass today’s ADAS offerings in performance, functionality and road safety.
A full-featured, Level 2+ system requires significantly more computational horsepower and sophisticated software than what is on the road today. NVIDIA DRIVE AutoPilot provides these, making it possible for carmakers to quickly deploy advanced autonomous solutions by 2020 and to scale this solution to higher levels of autonomy faster.—Rob Csongor, vice president of Autonomous Machines at NVIDIA
DRIVE AutoPilot integrates for the first time high-performance NVIDIA Xavier system-on-a-chip (SoC) processors and the latest NVIDIA DRIVE Software to process many deep neural networks (DNNs) for perception as well as complete surround camera sensor data from outside the vehicle and inside the cabin.
This combination enables full self-driving autopilot capabilities, including highway merge, lane change, lane splits and personal mapping. Inside the cabin, features include driver monitoring, AI copilot capabilities and advanced in-cabin visualization of the vehicle’s computer vision system.
DRIVE AutoPilot is part of the open, flexible NVIDIA DRIVE platform, which is being used by hundreds of companies worldwide to build autonomous vehicle solutions that increase road safety while reducing driver fatigue and stress on long drives or in stop-and-go traffic. The new Level 2+ system complements the NVIDIA DRIVE AGX Pegasus system that provides Level 5 capabilities for robotaxis.
DRIVE AutoPilot addresses the limitations of existing Level 2 ADAS systems, which a recent Insurance Institute for Highway Safety study showed offer inconsistent vehicle detections and poor ability to stay within lanes on curvy or hilly roads, resulting in a high occurrence of system disengagements where the driver abruptly had to take control.
Lane keeping and adaptive cruise control systems on the market today are simply not living up to the expectations of consumers. The high-performance AI solutions from NVIDIA will deliver more effective active safety and more reliable automated driving systems in the near future.—Dominique Bonte, vice president of Automotive Research at ABI Research
Central to NVIDIA DRIVE AutoPilot is the Xavier SoC, which delivers 30 trillion operations per second of processing capability. Architected for safety, Xavier has been designed for redundancy and diversity, with six types of processors and 9 billion transistors that enable it to process vast amounts of data in real time.
Xavier is the first automotive-grade processor for autonomous driving and is in production today. Global safety experts have assessed its architecture and development process as suitable for building a safe product.
The DRIVE AutoPilot software stack integrates DRIVE AV software for handling challenges outside the vehicle, as well as DRIVE IX software for tasks inside the car.
DRIVE AV uses surround sensors for full, 360-degree perception and features highly accurate localization and path-planning capabilities. These enable supervised self-driving on the highway, from on-ramp to off-ramp. Going beyond basic adaptive cruise control, lane keeping and automatic emergency braking, its surround perception capabilities handle situations where lanes split or merge, and safely perform lane changes.
DRIVE AV also includes a diverse and redundant set of advanced DNN technologies that enable the vehicle to perceive a wide range of objects and driving situations, including DriveNet, SignNet, LaneNet, OpenRoadNet and WaitNet. This sophisticated AI software understands where other vehicles are, reads lane markings, detects pedestrians and cyclists, distinguishes different types of lights and their colors, recognizes traffic signs and understands complex scenes.
In addition to providing precise localization to the world’s HD maps for vehicle positioning on the road, DRIVE AutoPilot offers a new personal mapping feature called “My Route,” which remembers where you have driven and can create a self-driving route even if no HD map is available.
Within the vehicle, DRIVE IX intelligent experience software enables occupant monitoring to detect distracted or drowsy drivers and provide alerts, or take corrective action if needed. It is also used to create intelligent user experiences, including the new ability for augmented reality. Displaying a visualization of the surrounding environment sensed by the vehicle, as well as planned route, instills trust in the system.
For next-generation user experiences in the vehicle, the AI capabilities of DRIVE IX can also be used to accelerate natural language processing, gaze tracking or gesture recognition.