• Skip to primary navigation
  • Skip to main content
  • Skip to primary sidebar
  • Skip to secondary sidebar
  • About
    • Contact
    • Privacy
    • Terms of use
  • Advertise
    • Advertising
    • Case studies
    • Design
    • Email marketing
    • Features list
    • Lead generation
    • Magazine
    • Press releases
    • Publishing
    • Sponsor an article
    • Webcasting
    • Webinars
    • White papers
    • Writing
  • Subscribe to Newsletter

Robotics & Automation News

Where Innovation Meets Imagination

  • Home
  • News
  • Features
  • Editorial Sections A-Z
    • Agriculture
    • Aircraft
    • Artificial Intelligence
    • Automation
    • Autonomous Vehicles
    • Business
    • Computing
    • Construction
    • Culture
    • Design
    • Drones
    • Economy
    • Energy
    • Engineering
    • Environment
    • Health
    • Humanoids
    • Industrial robots
    • Industry
    • Infrastructure
    • Investments
    • Logistics
    • Manufacturing
    • Marine
    • Material handling
    • Materials
    • Mining
    • Promoted
    • Research
    • Robotics
    • Science
    • Sensors
    • Service robots
    • Software
    • Space
    • Technology
    • Transportation
    • Warehouse robots
    • Wearables
  • Press releases
  • Events
nvidia drive ai

How deep learning enables autonomous vehicles to understand their environment

August 16, 2018 by David Edwards

Humans are constantly taking in data from the world around them using five primary senses. You hear your phone ring, see a notification on your computer screen or touch something hot.

However, without perception, there’s no way to decipher those inputs and determine what’s relevant. That you should answer the call, know there’s an email to respond to or pull away your hand before it’s burned.

Now imagine driving on a highway, where a constant stream of information surrounds you. From lane markings and street signs to lane-splitting motorcyclists, merging trucks and traffic jams – the ability to make instant, informed decisions is not just a skill, it’s an imperative. 

Just as perception enables humans to make instant associations and act on them, the ability to extract relevant knowledge from immediate surroundings is a fundamental pillar for the safe operation of an autonomous vehicle.

With the power of perception, a car can detect vehicles ahead using cameras and other sensors, identify if they become potential hazards and know to continuously track their movements.

This capability extends to the 360-degree field around the vehicle, enabling the car to detect and track all moving and static objects as it travels.

Perception is the first stage in the computational pipeline for the safe functioning of a self-driving car. Once the vehicle is able to extract relevant data from the surrounding environment, it can plan the path ahead and actuate, all without human intervention.

Finding the signal through the noise

Autonomous vehicle sensors generate massive amounts of data every second. From other cars, to pedestrians, to street signs, to traffic lights, every mile contains indicators for where the self-driving car should and shouldn’t go.

Identifying these indicators and determining those needed to safely move is incredibly complex, requiring a diversity of deep neural networks working in parallel.

The Nvidia Drive software stack – a primary component of the Nvidia Drive platform – contains libraries, frameworks and source packages that allow the necessary deep neural networks to work together for comprehensive perception.

These networks include DriveNet, which detects obstacles, and OpenRoadNet, which detects drivable space. To plan a path forward, LaneNet detects lane edges and PilotNet detects drivable paths.

Nvidia Drive software enables this integration by building on top of highly optimized and flexible libraries. These diverse networks run simultaneously and can overlap, providing redundancy, a key element to safety.

‘Inherently safe’

In addition to the redundancy within the perception layer, these networks back up the overall function of the vehicle, enhancing safety at every level, according to Nvidia, which claims the system is “inherently safe”.

For example, the car’s high-definition map can indicate a four-way intersection, and when paired with real-time sensor data, the perception layer shows the car precisely where to stop, enabling a more powerful way to pinpoint the car’s location.

Perception also contributes to the diversity of autonomous vehicle capabilities, enabling the car to see the world with the same sophistication as humans.

Rather than just identify obstacles, it can discern stationary objects as well as moving ones, and determine their path.

With added software capabilities, like those offered by Nvidia partner Perceptive Automata, the car can even predict human behavior by reading body language and other markers.

This added human behavior perception capability can run simultaneously with other algorithms operating an autonomous vehicle thanks to the computing horsepower from the Nvidia Drive platform.

With this combined hardware and software solution, developers are continuously adding new perception abilities to the car’s self-driving brain.

Print Friendly, PDF & Email

Share this:

  • Click to print (Opens in new window) Print
  • Click to share on Facebook (Opens in new window) Facebook
  • Click to share on LinkedIn (Opens in new window) LinkedIn
  • Click to share on Reddit (Opens in new window) Reddit
  • Click to share on X (Opens in new window) X
  • Click to share on Tumblr (Opens in new window) Tumblr
  • Click to share on Pinterest (Opens in new window) Pinterest
  • Click to share on WhatsApp (Opens in new window) WhatsApp
  • Click to share on Telegram (Opens in new window) Telegram
  • Click to share on Pocket (Opens in new window) Pocket

Related stories you might also like…

Filed Under: Features, Transportation Tagged With: autonomous car platform, autonomous cars, deep learning, nvidia, nvidia drive

Primary Sidebar

Search this website

Latest articles

  • Sonair raises $6 million to accelerate launch of ‘world’s first safe 3D ultrasonic sensor for robots’
  • ASG Power advances sustainability and efficiency through new training initiative
  • GMI and AINEXXO form strategic alliance to launch ‘self-aware and self-protecting factory’
  • SoftBank develops ‘robot-friendly’ server rack to enable automation at data centers
  • Georgia Tech team designs robot guide dog to assist the visually impaired
  • Figure raises over $1 billion in Series C funding at $39 billion valuation
  • Tennant Company and Brain Corp to debut next-generation autonomous cleaning at CMS 2025
  • Swisslog Healthcare partners with Diligent Robotics to enhance hospital logistics
  • Niryo launches robotic platform to make industrial automation ‘more accessible’
  • ABB and LandingAI unleash the power of generative AI for robotic vision

Secondary Sidebar

Copyright © 2025 · News Pro on Genesis Framework · WordPress · Log in

We use cookies on our website to give you the most relevant experience by remembering your preferences and repeat visits. By clicking “Accept”, you consent to the use of ALL the cookies.
Do not sell my personal information.
Cookie SettingsAccept
Manage consent

Privacy Overview

This website uses cookies to improve your experience while you navigate through the website. Out of these, the cookies that are categorized as necessary are stored on your browser as they are essential for the working of basic functionalities of the website. We also use third-party cookies that help us analyze and understand how you use this website. These cookies will be stored in your browser only with your consent. You also have the option to opt-out of these cookies. But opting out of some of these cookies may affect your browsing experience.
Necessary
Always Enabled
Necessary cookies are absolutely essential for the website to function properly. These cookies ensure basic functionalities and security features of the website, anonymously.
CookieDurationDescription
cookielawinfo-checkbox-analytics11 monthsThis cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Analytics".
cookielawinfo-checkbox-functional11 monthsThe cookie is set by GDPR cookie consent to record the user consent for the cookies in the category "Functional".
cookielawinfo-checkbox-necessary11 monthsThis cookie is set by GDPR Cookie Consent plugin. The cookies is used to store the user consent for the cookies in the category "Necessary".
cookielawinfo-checkbox-others11 monthsThis cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Other.
cookielawinfo-checkbox-performance11 monthsThis cookie is set by GDPR Cookie Consent plugin. The cookie is used to store the user consent for the cookies in the category "Performance".
viewed_cookie_policy11 monthsThe cookie is set by the GDPR Cookie Consent plugin and is used to store whether or not user has consented to the use of cookies. It does not store any personal data.
Functional
Functional cookies help to perform certain functionalities like sharing the content of the website on social media platforms, collect feedbacks, and other third-party features.
Performance
Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.
Analytics
Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics the number of visitors, bounce rate, traffic source, etc.
Advertisement
Advertisement cookies are used to provide visitors with relevant ads and marketing campaigns. These cookies track visitors across websites and collect information to provide customized ads.
Others
Other uncategorized cookies are those that are being analyzed and have not been classified into a category as yet.
SAVE & ACCEPT