One necessary condition for autonomous driving is an accurate and reliable representation of the environment around the vehicle. Current architectures rely on cameras, radars, and lidars to capture the visual environment and to localize and track other traffic participants. Human drivers can see but also hear and use a lot of auditory information for understanding the environment in addition to visual cues. In this paper, we present a pure sound localization and recognition system to extract an auditory representation of the environment. First, the environmental sound is classified into seven main categories of traffic objects followed by six specific kinds of sirens in an emergency case using a simple neural network layout. Second, each object is localized via a combined time-delay of arrival and amplitude-based localization algorithm. The system is evaluated on real-world data focusing on a robust detection and accurate localization of emergency vehicles.


    Access

    Check access

    Check availability in my library

    Order at Subito €


    Export, share and cite



    Title :

    Auditory Scene Understanding for Autonomous Driving


    Contributors:


    Publication date :

    2021-07-11


    Size :

    1897000 byte




    Type of media :

    Conference paper


    Type of material :

    Electronic Resource


    Language :

    English



    DRIVING SCENE UNDERSTANDING

    DING SHUGUANG / JIN YUEXIANG / FAN MINGYU et al. | European Patent Office | 2021

    Free access

    Scene recognition in autonomous driving environment

    YLENHAMMAR MAGNUS / SVEINKRONA HU KAN | European Patent Office | 2022

    Free access

    Driving scene based path planning for autonomous driving vehicles

    ZHU FAN | European Patent Office | 2021

    Free access


    Awareness of Road Scene Participants for Autonomous Driving

    Prof. Petrovskaya, Anna / Dr. Perrollaz, Mathias / Dr. Oliveira, Luciano et al. | Springer Verlag | 2012