idw – Informationsdienst Wissenschaft

Nachrichten, Termine, Experten

Grafik: idw-Logo
Science Video Project
idw-Abo

idw-News App:

AppStore

Google Play Store



Instanz:
Teilen: 
17.11.2020 14:54

Using virtual reality to better understand auditory perception

Dr. Corinna Dahm-Brey Presse & Kommunikation
Carl von Ossietzky-Universität Oldenburg

    In the new priority program AUDICTIVE funded by the German Research Foundation, Oldenburg experts want to use virtual reality to better understand complex auditory processes. In three projects, they are going to simulate room acoustics, investigate how the ear assesses distances and movements and study visual stimuli that help to direct attention to the desired sound source.

    Many people find it difficult to hear in classrooms, open-plan offices or near busy streets. In order to understand how acoustic perception works in such complex environments, leading experts of the University of Oldenburg, Germany, count on virtual reality (VR). As part of the new priority programme Audictive ("Auditive Cognition in Interactive Virtual Environments"), they set out to clarify how complicated auditory processes such as perception, attention and memory can be investigated using new, interactive VR methods. The German Research Foundation (DFG) is funding three projects led by acoustics Experts Dr. Stephan Ewert, Prof. Dr. Volker Hohmann, Prof. Dr. Steven van de Par and neuropsychologist Prof. Dr. Stefan Debener with about 830.000 euros over the next three years.

    The first project, led by Prof. Dr. Mathias Blau of the University of Applied Sciences in Oldenburg together with van de Par, aims to find out experimentally whether room acoustics can be realistically simulated in a virtual environment. The team also wants to explore how visual perception influences the acoustic experience. Moreover, the researchers are going to investigate how audio rendering in virtual worlds affects social anxiety. That is, for example, whether the fear of giving a talk in front of a large number of people can be reproduced in a virtual environment and thus be studied more easily. Researchers of the University of Regensburg are also involved in the project.

    The second project, led by Ewert, aims to find out how volunteers assess distances and movements using their sense of hearing. This is particularly important in situations with limited vision or when objects are outside the field of vision and the ear acts as a warning system. To this end, the team is going to place test persons in different virtual environments using VR glasses. The acoustic stimuli will be presented either by headphones or loudspeakers. In the experiments, the researchers will vary the type of acoustic sources, their distance and intensity. Van de Par and Dr. Virginia Flanagin of the University Hospital in Munich are also principal investigators in this project. This research relies on the acoustic simulation system RAZR, developed in the groups of both Ewert and van de Par. It creates virtual acoustic environments that are very close to reality.

    The third project, led by Debener and Hohmann, will also be based on experimental studies in virtual environments. The aim is to understand how healthy people manage to filter out the voice of their current conversation partner from a variety of sound sources. This ability goes beyond just hearing or understanding speech, but requires the ability to control attention adequately. Hohmann, who heads a collaborative research centre on hearing acoustics, and his team has developed a powerful acoustic stimulation system called TASCAR. This platform provides virtual characters not only with a voice, but also with realistic lip movements. It is also capable of reproducing common everyday conversational situations. To find out how the test subjects control their attention, the team wants to measure the electrical activity in the test persons’ brain with a mobile EEG device developed by Debener's research group. The aim of the cooperation is to establish mobile EEG technology in VR and, for example, to investigate the visual stimuli that help to direct attention to the desired sound source.

    AUDICTIVE is one of 14 priority programmes established by the DFG this year. It combines the disciplines of acoustics, cognitive psychology and computer science.


    Wissenschaftliche Ansprechpartner:

    Prof. Dr. Steven van de Par, Tel.: 0441/798-3450, E-Mail: steven.van.de.par@uol.de


    Weitere Informationen:

    https://uol.de/en/neuropsychology Neuropsychology Lab
    https://uol.de/en/mediphysics-acoustics Department of Medical Physics and Acoustics
    http://www.spp2236-audictive.de Priority Programme AUDICTIVE


    Bilder

    How do healthy people manage to filter out the voice of the person they are talking to from a variety of sound sources? Researchers in Oldenburg are investigating this using EEG measurements and virtual reality.
    How do healthy people manage to filter out the voice of the person they are talking to from a variet ...

    University of Oldenburg/Giso Grimm


    Merkmale dieser Pressemitteilung:
    Journalisten, Lehrer/Schüler, Studierende, Wirtschaftsvertreter, Wissenschaftler, jedermann
    Informationstechnik, Medizin, Psychologie
    überregional
    Forschungsprojekte, Kooperationen
    Englisch


     

    How do healthy people manage to filter out the voice of the person they are talking to from a variety of sound sources? Researchers in Oldenburg are investigating this using EEG measurements and virtual reality.


    Zum Download

    x

    Hilfe

    Die Suche / Erweiterte Suche im idw-Archiv
    Verknüpfungen

    Sie können Suchbegriffe mit und, oder und / oder nicht verknüpfen, z. B. Philo nicht logie.

    Klammern

    Verknüpfungen können Sie mit Klammern voneinander trennen, z. B. (Philo nicht logie) oder (Psycho und logie).

    Wortgruppen

    Zusammenhängende Worte werden als Wortgruppe gesucht, wenn Sie sie in Anführungsstriche setzen, z. B. „Bundesrepublik Deutschland“.

    Auswahlkriterien

    Die Erweiterte Suche können Sie auch nutzen, ohne Suchbegriffe einzugeben. Sie orientiert sich dann an den Kriterien, die Sie ausgewählt haben (z. B. nach dem Land oder dem Sachgebiet).

    Haben Sie in einer Kategorie kein Kriterium ausgewählt, wird die gesamte Kategorie durchsucht (z.B. alle Sachgebiete oder alle Länder).