idw – Informationsdienst Wissenschaft

Nachrichten, Termine, Experten

Grafik: idw-Logo
Grafik: idw-Logo

idw - Informationsdienst
Wissenschaft

Science Video Project
idw-Abo

idw-News App:

AppStore

Google Play Store



Instance:
Share on: 
09/13/2022 13:28

Unlocking human-like perception in self-driving vehicles

Bastian Strauch Hochschul- und Wissenschaftskommunikation
Albert-Ludwigs-Universität Freiburg im Breisgau

    • Researchers at the University of Freiburg have introduced the amodal panoptic segmentation task whose solution may soon enable more human-like environment perception for self-driving vehicles through AI algorithms
    • Unveiled at the AutoSens conference in the Autoworld Museum in Brussels
    • “We are confident that novel AI algorithms for this task will enable robots to emulate the visual experience that humans have by perceiving complete physical structures of objects,” says Valada.

    How can mobile robots perceive and understand the environment correctly, even if parts of the environment are occluded by other objects? This is a key question that must be solved for self-driving vehicles to safely navigate in large crowded cities. While humans can imagine complete physical structures of objects even when they are partially occluded, existing artificial intelligence (AI) algorithms that enable robots and self-driving vehicles to perceive their environment do not have this capability. Robots with AI can already find their way around and navigate on their own once they have learned what their environment looks like. However, perceiving the entire structure of objects when they are partially hidden, such as people in crowds or vehicles in traffic jams, has been a significant challenge. A major step towards solving this problem has now been taken by Freiburg robotics researchers Prof. Dr. Abhinav Valada and PhD student Rohit Mohan from the Robot Learning Lab at the University of Freiburg, which they have presented in two joint publications.

    A task whose solution promises more safety

    The two Freiburg scientists have developed the amodal panoptic segmentation task and demonstrated its feasibility using novel AI approaches. Until now, self-driving vehicles have used panoptic segmentation to understand their surroundings. This means that they can so far only predict which pixels of an image belong to which “visible” regions of an object such as a person or car, and identify instances of those objects. What they lack so far is being able to also predict the entire shape of objects even when they are partially occluded by other objects next to them. The new task of perception with amodal panoptic segmentation makes this holistic understanding of the environment possible.

    Amodal refers to the case that any partial occlusion of objects must be abstracted and instead of viewing them as fragments, there should be a general understanding of viewing them as a whole. Thus, this improved ability of visual recognition will lead to enormous progress in improving the safety of self-driving vehicles.

    Potential to revolutionize urban visual scene understanding

    In a new paper published at the IEEE/CVF Computer Vision and Pattern Recognition Conference (CVPR), the researchers have added the new task to established benchmark datasets and made them publicly available. They are now calling on scientists to participate in the benchmarking with their own AI algorithms. The goal of this task is the pixel-wise semantic segmentation of the visible regions of amorphous background classes such as roads, vegetation, sky, and the instance segmentation of both the visible and occluded object regions of countable classes such as cars, trucks, and pedestrians.

    The benchmark and datasets are publicly available on the website, including two proposed novel learning algorithms. “We are confident that novel AI algorithms for this task will enable robots to emulate the visual experience that humans have by perceiving complete physical structures of objects,” Valada explains. “Amodal panoptic segmentation will significantly help downstream automated driving tasks where occlusion is a major challenge such as depth estimation, optical flow, object tracking, pose estimation, motion prediction, etc. With more advanced AI algorithms for this task, visual recognition ability for self-driving cars can be revolutionized. For example, if the entire structure of road users is perceived at all times, regardless of partial occlusions, the risk of accidents can be significantly minimized.”

    In addition, by inferring the relative depth ordering of objects in a scene, automated vehicles can make complex decisions such as in which direction to move toward the object to get a clearer view. In order to make these visions a reality, the task and its benefits were presented to leading automotive industry professionals at AutoSens, which was held at the Autoworld Museum in Brussels.

    Overview of facts:
    • Abhinav Valada is an assistant professor and head of the Robot Learning Lab at the Department of Computer Science at the University of Freiburg as well as a member of the BrainLinks-BrainTools centre.
    • He primarily conducts research on topics at the intersection of robotics, machine learning, and computer vision, such as fundamental problems in robot perception, state estimation, and planning using machine learning approaches to enable robots to reliably operate in more complex domains and diverse environments. His goal is to develop scalable, lifelong learning robotic systems.
    • Robot Learning Lab: Valada's team focuses on developing scalable robot learning algorithms for mobility and manipulation. They combine modern machine learning techniques such as deep learning and reinforcement learning, with robotics to expand the abilities of robots to operate in diverse real-world environments.
    • Original publications: Mohan, Rohit, Abhinav Valada: “Amodal Panoptic Segmentation”, IEEE/ CVF International Conference on Computer Vision and Pattern Recognition (CVPR), pp. 21023-21032, 2022.
    Mohan, Rohit, Valada, Abhinav: “Perceiving the Invisible: Proposal-Free Amodal Panoptic Segmentation”, IEEE Robotics and Automation Letters (RA-L), vol. 7, no. 4, pp. 9302-9309, 2022.
    • For more information: http://amodal-panoptic.cs.uni-freiburg.de
    • Explanatory video: “Amodal Panoptic Segmentation”


    Contact for scientific information:

    Prof. Dr. Abhinav Valada
    Robot Learning Lab
    Faculty of Engineering
    University of Freiburg
    Tel.: 0761 203-8025
    e-mail: valada@cs.uni-freiburg.de


    Original publication:

    Mohan, Rohit, Abhinav Valada: “Amodal Panoptic Segmentation”, IEEE/ CVF International Conference on Computer Vision and Pattern Recognition (CVPR), pp. 21023-21032, 2022.
    Mohan, Rohit, Valada, Abhinav: “Perceiving the Invisible: Proposal-Free Amodal Panoptic Segmentation”, IEEE Robotics and Automation Letters (RA-L), vol. 7, no. 4, pp. 9302-9309, 2022.


    More information:

    https://kommunikation.uni-freiburg.de/pm-en/press-releases-2022/unlocking-human-...


    Images

    Criteria of this press release:
    Journalists
    Electrical engineering, Traffic / transport
    transregional, national
    Research results, Transfer of Science or Research
    English


     

    Help

    Search / advanced search of the idw archives
    Combination of search terms

    You can combine search terms with and, or and/or not, e.g. Philo not logy.

    Brackets

    You can use brackets to separate combinations from each other, e.g. (Philo not logy) or (Psycho and logy).

    Phrases

    Coherent groups of words will be located as complete phrases if you put them into quotation marks, e.g. “Federal Republic of Germany”.

    Selection criteria

    You can also use the advanced search without entering search terms. It will then follow the criteria you have selected (e.g. country or subject area).

    If you have not selected any criteria in a given category, the entire category will be searched (e.g. all subject areas or all countries).