idw – Informationsdienst Wissenschaft

Nachrichten, Termine, Experten

Grafik: idw-Logo
Science Video Project
idw-Abo

idw-News App:

AppStore

Google Play Store



Instance:
Share on: 
11/17/2020 14:54

Using virtual reality to better understand auditory perception

Dr. Corinna Dahm-Brey Presse & Kommunikation
Carl von Ossietzky-Universität Oldenburg

    In the new priority program AUDICTIVE funded by the German Research Foundation, Oldenburg experts want to use virtual reality to better understand complex auditory processes. In three projects, they are going to simulate room acoustics, investigate how the ear assesses distances and movements and study visual stimuli that help to direct attention to the desired sound source.

    Many people find it difficult to hear in classrooms, open-plan offices or near busy streets. In order to understand how acoustic perception works in such complex environments, leading experts of the University of Oldenburg, Germany, count on virtual reality (VR). As part of the new priority programme Audictive ("Auditive Cognition in Interactive Virtual Environments"), they set out to clarify how complicated auditory processes such as perception, attention and memory can be investigated using new, interactive VR methods. The German Research Foundation (DFG) is funding three projects led by acoustics Experts Dr. Stephan Ewert, Prof. Dr. Volker Hohmann, Prof. Dr. Steven van de Par and neuropsychologist Prof. Dr. Stefan Debener with about 830.000 euros over the next three years.

    The first project, led by Prof. Dr. Mathias Blau of the University of Applied Sciences in Oldenburg together with van de Par, aims to find out experimentally whether room acoustics can be realistically simulated in a virtual environment. The team also wants to explore how visual perception influences the acoustic experience. Moreover, the researchers are going to investigate how audio rendering in virtual worlds affects social anxiety. That is, for example, whether the fear of giving a talk in front of a large number of people can be reproduced in a virtual environment and thus be studied more easily. Researchers of the University of Regensburg are also involved in the project.

    The second project, led by Ewert, aims to find out how volunteers assess distances and movements using their sense of hearing. This is particularly important in situations with limited vision or when objects are outside the field of vision and the ear acts as a warning system. To this end, the team is going to place test persons in different virtual environments using VR glasses. The acoustic stimuli will be presented either by headphones or loudspeakers. In the experiments, the researchers will vary the type of acoustic sources, their distance and intensity. Van de Par and Dr. Virginia Flanagin of the University Hospital in Munich are also principal investigators in this project. This research relies on the acoustic simulation system RAZR, developed in the groups of both Ewert and van de Par. It creates virtual acoustic environments that are very close to reality.

    The third project, led by Debener and Hohmann, will also be based on experimental studies in virtual environments. The aim is to understand how healthy people manage to filter out the voice of their current conversation partner from a variety of sound sources. This ability goes beyond just hearing or understanding speech, but requires the ability to control attention adequately. Hohmann, who heads a collaborative research centre on hearing acoustics, and his team has developed a powerful acoustic stimulation system called TASCAR. This platform provides virtual characters not only with a voice, but also with realistic lip movements. It is also capable of reproducing common everyday conversational situations. To find out how the test subjects control their attention, the team wants to measure the electrical activity in the test persons’ brain with a mobile EEG device developed by Debener's research group. The aim of the cooperation is to establish mobile EEG technology in VR and, for example, to investigate the visual stimuli that help to direct attention to the desired sound source.

    AUDICTIVE is one of 14 priority programmes established by the DFG this year. It combines the disciplines of acoustics, cognitive psychology and computer science.


    Contact for scientific information:

    Prof. Dr. Steven van de Par, Tel.: 0441/798-3450, E-Mail: steven.van.de.par@uol.de


    More information:

    https://uol.de/en/neuropsychology Neuropsychology Lab
    https://uol.de/en/mediphysics-acoustics Department of Medical Physics and Acoustics
    http://www.spp2236-audictive.de Priority Programme AUDICTIVE


    Images

    How do healthy people manage to filter out the voice of the person they are talking to from a variety of sound sources? Researchers in Oldenburg are investigating this using EEG measurements and virtual reality.
    How do healthy people manage to filter out the voice of the person they are talking to from a variet ...

    University of Oldenburg/Giso Grimm


    Criteria of this press release:
    Business and commerce, Journalists, Scientists and scholars, Students, Teachers and pupils, all interested persons
    Information technology, Medicine, Psychology
    transregional, national
    Cooperation agreements, Research projects
    English


     

    How do healthy people manage to filter out the voice of the person they are talking to from a variety of sound sources? Researchers in Oldenburg are investigating this using EEG measurements and virtual reality.


    For download

    x

    Help

    Search / advanced search of the idw archives
    Combination of search terms

    You can combine search terms with and, or and/or not, e.g. Philo not logy.

    Brackets

    You can use brackets to separate combinations from each other, e.g. (Philo not logy) or (Psycho and logy).

    Phrases

    Coherent groups of words will be located as complete phrases if you put them into quotation marks, e.g. “Federal Republic of Germany”.

    Selection criteria

    You can also use the advanced search without entering search terms. It will then follow the criteria you have selected (e.g. country or subject area).

    If you have not selected any criteria in a given category, the entire category will be searched (e.g. all subject areas or all countries).