• Young woman with EEG cap, virtual cafeteria environment in the background.

    A healthy person can easily filter out the voice of their current conversation partner from a variety of sound sources. Researchers in Oldenburg are using EEG measurements and virtual reality to investigate the role of attention in this process. Photo: University of Oldenburg/Giso Grimm

Using virtual reality to better understand auditory perception

How can virtual reality help to better investigate complex auditory processes? This is what Oldenburg researchers are investigating in three projects of the priority programm AUDICTIVE.

How can virtual reality help to better investigate complex auditory processes? This is what Oldenburg researchers are investigating in three projects of the priority programm AUDICTIVE.

Many people find it difficult to hear in classrooms, open-plan offices or near busy streets. In order to understand how acoustic perception works in such complex environments, leading experts of the University of Oldenburg, Germany, count on virtual reality (VR). As part of the new priority programme Audictive ("Auditive Cognition in Interactive Virtual Environments"), they set out to clarify how complicated auditory processes such as perception, attention and memory can be investigated using new, interactive VR methods. The German Research Foundation (DFG) is funding three projects led by acoustics Experts Dr. Stephan Ewert, Prof. Dr. Volker Hohmann, Prof. Dr. Steven van de Par and neuropsychologist Prof. Dr. Stefan Debener with about 830.000 euros over the next three years.

The sound of virtual rooms

The first project, led by Prof. Dr. Mathias Blau of the University of Applied Sciences in Oldenburg together with van de Par, aims to find out experimentally whether room acoustics can be realistically simulated in a virtual environment. The team also wants to explore how visual perception influences the acoustic experience. Moreover, the researchers are going to investigate how audio rendering in virtual worlds affects social anxiety. That is, for example, whether the fear of giving a talk in front of a large number of people can be reproduced in a virtual environment and thus be studied more easily. Researchers of the University of Regensburg are also involved in the project.

The second project, led by Ewert, aims to find out how volunteers assess distances and movements using their sense of hearing. This is particularly important in situations with limited vision or when objects are outside the field of vision and the ear acts as a warning system. To this end, the team is going to place test persons in different virtual environments using VR glasses. The acoustic stimuli will be presented either by headphones or loudspeakers. In the experiments, the researchers will vary the type of acoustic sources, their distance and intensity. Van de Par and Dr. Virginia Flanagin of the University Hospital in Munich are also principal investigators in this project. This research relies on the acoustic simulation system RAZR, developed in the groups of both Ewert and van de Par. It creates virtual acoustic environments that are very close to reality.

Chatting with virtual characters

The third project, led by Debener and Hohmann, will also be based on experimental studies in virtual environments. The aim is to understand how healthy people manage to filter out the voice of their current conversation partner from a variety of sound sources. This ability goes beyond just hearing or understanding speech, but requires the ability to control attention adequately. Hohmann, who heads a collaborative research centre on hearing acoustics, and his team have developed a powerful acoustic stimulation system called TASCAR. This platform provides virtual characters not only with a voice, but also with realistic lip movements. It is also capable of reproducing common everyday conversational situations. To find out how the test subjects control their attention, the team wants to measure the electrical activity in the test persons’ brain with a mobile EEG device developed by Debener's research group. The aim of the cooperation is to establish mobile EEG technology in VR and, for example, to investigate the visual stimuli that help to direct attention to the desired sound source.

AUDICTIVE is one of 14 priority programmes established by the DFG this year. It combines the disciplines of acoustics, cognitive psychology and computer science.

This might also be of interest to you:

Ein Mann mit Kopfhörern sitzt vor einem Computermonitor, der viele Regler anzeigt. Der Betrachter blickt dem Mann über die Schulter.
Excellence Strategy Top News Medical Physics and Acoustics

Singing has become quieter

Researchers from Oldenburg have discovered that lead singers used to be significantly louder in relation to the rest of their band than they are today…

more
Bernhard Eurich und Mathias Dietz sitzen im Hörlabor vor einem Computer und tragen Kopfhörer.
Hearing Research Excellence Strategy Top News Medical Physics and Acoustics

Hearing by numbers

Researchers have developed a mathematical model that can simulate directional hearing using considerably less computing power than others – and also…

more
Frontal view into the laboratory, the wedges can be seen on all walls. A man is in the center of the room, standing on a net of wires.
Hearing Research Excellence Strategy Research Top News Medical Physics and Acoustics

A room without echoes

The door of the Anoechoic Lab on Wechloy Campus opens onto a world of silence. The key feature of the recently renovated acoustics lab is its huge…

more
Presse & Kommunikation (Changed: 10 Dec 2024)  | 
Zum Seitananfang scrollen Scroll to the top of the page