Back

Neural Tracking of Audiovisual Effects in Noise Using Deep Neural Network-Generated Virtual Humans

Cooper, J. K.; Vanthornhout, J.; van Wieringen, A.; Francart, T.

2025-06-05 neuroscience
10.1101/2025.06.02.656280 bioRxiv
Show abstract

This study investigates the effectiveness of Deep Neural Network (DNN)-generated virtual humans in enhancing audiovisual speech perception in noisy environments, with a focus on using neural measures to quantify these effects. Lip movements are essential for speech comprehension, especially when auditory cues are degraded by noise. Traditional recording methods produce high-quality audiovisual materials but are resource intensive. This research explores the use of DNN avatars as a promising alternative, utilizing a commercially available tool to create realistic virtual humans. The study included both simple sentences and a short story to improve ecological validity. Eleven young, normal-hearing participants proficient in Flemish-Dutch listened to semantically meaningful sentences and a short story with various speaker types: a female FACS avatar, male and female DNN avatars, and a video of a human male speaker. The study included behavioral measures which consisted of an adaptive recall procedure and an adaptive rate procedure and electrophysiological measures consisting of neural tracking. Findings in the adaptive recall procedure showed consistent audiovisual benefits, with the human speaker offering the greatest benefit (-4.75 dB SNR), followed by the DNN avatar (-4.00 dB SNR) and the FACS avatar (-1.55 dB SNR). Additionally in the adaptive rate procedure, the DNN avatar improved speech intelligibility, with average SRTs enhancing from -7.17 dB SNR (audio-only) to -9.02 dB SNR (audiovisual). The results from the neural tracking procedure indicated that most participants experienced audiovisual benefits, particularly in the -9 dB SNR range, revealing that audiovisual cues provided by DNN avatars can enhance speech perception, validating these avatars as effective tools for studying audiovisual effects when using both behavioral measures and electrophysiological measures.

Matching journals

The top 6 journals account for 50% of the predicted probability mass.

1
Trends in Hearing
12 papers in training set
Top 0.1%
14.7%
2
Scientific Reports
3102 papers in training set
Top 5%
10.4%
3
PLOS ONE
4510 papers in training set
Top 18%
10.4%
4
The Journal of the Acoustical Society of America
33 papers in training set
Top 0.1%
7.0%
5
Hearing Research
49 papers in training set
Top 0.1%
7.0%
6
Ear & Hearing
15 papers in training set
Top 0.1%
5.0%
50% of probability mass above
7
Frontiers in Neuroscience
223 papers in training set
Top 0.9%
4.4%
8
Journal of Neural Engineering
197 papers in training set
Top 0.6%
4.1%
9
Frontiers in Human Neuroscience
67 papers in training set
Top 0.4%
3.7%
10
Journal of Neuroscience Methods
106 papers in training set
Top 0.6%
2.5%
11
Behavior Research Methods
25 papers in training set
Top 0.1%
2.1%
12
Journal of the Association for Research in Otolaryngology
11 papers in training set
Top 0.1%
1.9%
13
Philosophical Transactions of the Royal Society B
51 papers in training set
Top 2%
1.9%
14
Frontiers in Psychiatry
83 papers in training set
Top 3%
1.0%
15
IEEE Transactions on Neural Systems and Rehabilitation Engineering
40 papers in training set
Top 0.5%
0.9%
16
Brain Sciences
52 papers in training set
Top 2%
0.8%
17
Frontiers in Psychology
49 papers in training set
Top 1%
0.8%
18
eneuro
389 papers in training set
Top 9%
0.8%
19
Healthcare
16 papers in training set
Top 2%
0.7%
20
European Journal of Neuroscience
168 papers in training set
Top 2%
0.7%
21
NeuroImage
813 papers in training set
Top 7%
0.5%
22
IEEE Transactions on Biomedical Engineering
38 papers in training set
Top 1%
0.5%
23
Translational Vision Science & Technology
35 papers in training set
Top 0.7%
0.5%
24
Journal of Speech, Language, and Hearing Research
10 papers in training set
Top 0.1%
0.5%
25
Journal of Vision
92 papers in training set
Top 0.6%
0.5%
26
Journal of NeuroEngineering and Rehabilitation
28 papers in training set
Top 1%
0.5%
27
Journal of Visualized Experiments
30 papers in training set
Top 1%
0.5%