- Data Note
- Open access
- Published:
Data on neurophysiological and psychological responses to audio immersive experience in stereo and 3D audio formats
BMC Research Notes volume 17, Article number: 356 (2024)
Abstract
Objectives
This dataset presents demographic, psychological, auditory and neurophysiological information of 31 volunteers, who participated in an experiment measuring the auditory immersive experience in two audio formats: stereophonic downmix and three-dimensional audio. This dataset could help understand the objectiveness (based on the nervous system response) behind the subjectiveness of immersion brought about by the audio format (based on the listener evaluation). The final objective of this dataset is to study the psychological and neurophysiological responses of immersive attributes in auditory events in future studies.
Data description
Participants were randomly assigned to an audio format and attended two listening sessions, where soundscapes of a concert hall and urban park were randomly played. Electroencephalography, electrodermal activity, and blood volume pressure data were collected during auditory stimulation, and psychological responses to the auditory immersive experience were collected at the end of the session.
Clinical trial number
Not applicable.
Objective
The concept of immersion has been a compelling topic in the field of extended reality without reaching a consensus among researchers. Immersive audio has been used interchangeably with three-dimensional (3D) audio, assuming that a high-fidelity spatial audio system causes an immersive experience. However, this assumption lacks evidence [1].
Nervous system responses could help researchers objectively understand the attributes of the immersive experience [2]. In electroencephalography (EEG) data, research on the sense of presence has been associated with neural oscillations within clinical bands such as theta and beta [3, 4]. For blood volume pressure (BVP), heart rate variability, and its frequency composition have been related to involvement [5, 6]. For electrodermal activity (EDA), descriptive features in skin conductance response and levels can be extracted to monitor behavior in emotional activation [5, 7].
Therefore, understanding auditory immersive experiences through behavioral questionnaires and neurophysiological responses motivated the data collection, enabling the extraction of neural features and the study of their association with immersive attributes. The audio recordings used are also included. Firstly, participants voluntarily registered for the experiment. Secondly, participants were selected based on exclusion criteria on auditory and neurological health. Thirdly, participants answered the Immersive Tendencies Questionnaire (ITQ) [8] to be assigned to stereophonic downmix (stereo) or 3D audio group. Lastly, participants attended two listening sessions, where the soundscape of a concert hall and an urban park were randomly played. During each listening session, EEG, BVP and EDA recordings of resting state and auditory stimulation were collected while participants were seated in the middle of a critical listening room with their eyes closed. After each soundscape, participants answered an immersive experience questionnaire.
Previous work using this dataset showed statistical significance between the mean heart rate variability and the immersive attribute of involvement [4], indicating that the volume of data may be suitable for further signal and statistical analyses.
Data description
Data corresponds to 31 subjects (stereo: 10 males, five females, mean ITQ = 49.2. 3D: 10 males, six females, mean ITQ = 48.3) aged 18–36 years, all university students. Participants were excluded if they: (1) were outside the age range to minimize the likelihood of presenting hearing loss [9], (2) reported auditory and neurological problems, and (3) reported never having been in a concert hall or an urban park. After registration, participants were randomly and blindly assigned to an audio format group (stereo or 3D) based on ITQ scores. Table 1 provides an overview of data files and data sets available in the repository.
DataSpecs excel file (Data file 1) contains six worksheets: (1) Demographic data: ID, audio format group, ITQ score, Sex, Age, Highest education level, Country of origin, and Handedness; (2) Biofiles: ID, audio format group, listening condition, biosignal, sample rate, number of events and total length in seconds; (3) ITQ: ID, audio format group, and responses to items of ITQ; (4) Responses: ID, listening condition, audio format group, responses to the Self-Assessment Manikin test [10], igroup Presence Questionnaire [11, 12], and system performance questionnaire designed by the authors based on the Wheel of Sound recommendation [13, 14]; (5) Items: questionnaires, item codes if available by the authors of the assessments, the question or statement of the item, type of scale, and minimum and maximum values; and (6) EEGchannels: electrode positions in Theta/Phi coordinates.
LoudspeakerInformation file (Data file 2) explains the alignment and positions of loudspeakers for stereo, PCMA-3D, and ESMA-3D audio playback. PCMA-3D and ESMA-3D are horizontally spaced and vertically coincident microphone techniques dedicated to music and sound field recordings, respectively [15].
RawData folder (Data set 1) has individual subfolders of participants with EEG, EDA, and BVP recordings in three conditions: (1) resting state (Bl), (2) concert hall (Music), and (3) urban park (Park) soundscapes. Neurophysiological data were collected with OpenViBE v3.3.1 and the following devices: (1) EEG: 24-channel electrode cap following the 10/20 international electrode system connected to a Smarting mbrain train amplifier at 500 Hz sample rate; (2) BVP and (3) EDA signals: Empatica E4 wristband connected via Lab Streaming Layer code [16] with 64 Hz and 4 Hz sample rates, respectively. Condition 8 code in GDF files indicates the start of audio playback and can be ignored for Bl files. Code 33031 refers to one second before the end of the experiment and is not indicated in Music and Park files since the recording was stopped a few seconds after the sound finished to avoid audio interruption.
LatencyAdjustment folder (Data set 2) mirrors Raw Data organization. Codes latencies were adjusted due to latencies caused by sound traveling through the air and digital delay. This process is explained in Data file 2.
AudioFiles folder (Data set 3) contains two subfolders: (1) Music: 2-minute-long WAV audio files of concert hall recordings [17]; and (2) Park: 2-minute-long WAV audio files of urban park recordings [18]. All files were generated at 48k Hz sample rate and 24-bit depth. These files are organized for the three loudspeaker configurations detailed in Data file 2. Max 8 software was used for audio playback.
Limitations
-
The data of three participants in specific soundscapes is missing because they did not attend the last session: Music – ID 20 and 40, and Park - ID 25.
-
In some neurophysiological recordings, signal discontinuities and noisy segments may be present due to momentary device disconnection or subject movements, respectively, modifying the actual length of the recording as indicated in Data file 1.
-
Stereophonic versions were created from the downmix process of PCMA-3D and ESMA-3D recordings. As a result, future findings may be applicable only to these specific playback conditions.
Data availability
The data described in this Data note can be freely and openly accessed on Figshare under https://doiorg.publicaciones.saludcastillayleon.es/10.6084/m9.figshare.25421464.v3. Please see Table 1 and reference [12] for details and links to the data.
Abbreviations
- 3D:
-
Three dimensional
- Stereo:
-
Stereophonic downmix
- EEG:
-
Electroencephalography
- BVP:
-
Blood volume pressure
- EDA:
-
Electrodermal activity
- ITQ:
-
Immersive tendencies questionnaire
- Bl:
-
Resting state
- Music:
-
Concert hall
- Park:
-
Urban park
References
Agrawal S, Bech S, de Moor K, Forchhammer S. Influence of changes in Audio Spatialization on Immersion in Audiovisual experiences. AES J Audio Eng Soc. 2022;70:810–23. https://doiorg.publicaciones.saludcastillayleon.es/10.17743/jaes.2022.0034.
Zhang C. The why, what, and how of immersive experience. IEEE Access. 2020;8:90878–88. https://doiorg.publicaciones.saludcastillayleon.es/10.1109/ACCESS.2020.2993646.
Zhang S, Feng X, Shen Y. Quantifying Auditory Presence using Electroencephalography. Appl Sci. 2021;11:10461. https://doiorg.publicaciones.saludcastillayleon.es/10.3390/app112110461.
Naal-Ruiz NE, Lee H, Gonzalez A, Gonzalez R, Duville MM, Alonso-Valerdi LM et al. Towards the neurophysiological characterization of immersive attributes in auditory events. In: 2023 AES International Conference on Spatial and Immersive Audio. 2023. Huddersfield, UK. Audio Engineering Society. p. 9. https://www.aes.org/e-lib/browse.cfm?elib=22167
Kobayashi M, Ueno K, Ise S. The effects of Spatialized sounds on the sense of Presence in Auditory virtual environments: a psychological and physiological study. Presence Teleoperators Virtual Environ. 2015;24:163–74. https://doiorg.publicaciones.saludcastillayleon.es/10.1162/PRES_a_00226.
Geronazzo M, Rosenkvist A, Eriksen DS, Markmann-Hansen CK, Køhlert J, Valimaa M et al. Creating an Audio Story with Interactive Binaural Rendering in Virtual Reality. Wirel Commun Mob Comput. 2019;2019. https://doiorg.publicaciones.saludcastillayleon.es/10.1155/2019/1463204
Cuadrado F, Lopez-Cobo I, Mateos-Blanco T, Tajadura-Jiménez A. Arousing the sound: a field study on the emotional impact on children of arousing Sound Design and 3D Audio spatialization in an Audio Story. Front Psychol. 2020;11:1–19. https://doiorg.publicaciones.saludcastillayleon.es/10.3389/fpsyg.2020.00737.
Witmer BG, Singer MJ. Measuring Presence in virtual environments: a Presence Questionnaire. Presence Teleoperators Virtual Environ. 1998;7:225–40. https://doiorg.publicaciones.saludcastillayleon.es/10.1162/105474698565686.
Jin IK, Lee D, Jeong Y, Seo YJ, Kong TH, Suh MJ, et al. Trends in distributions of hearing threshold levels by ages: a comparison of the ISO 7029 and newly available country-specific data. J Audiol Otol. 2024;28:1–9. https://doiorg.publicaciones.saludcastillayleon.es/10.7874/JAO.2023.00626.
Bradley MM, Lang PJ. Measuring emotion: the self-assessment manikin and the semantic differential. J Behav Ther Exp Psychiatry. 1994;25:49–59. https://doiorg.publicaciones.saludcastillayleon.es/10.1016/0005-7916(94)90063-9.
Oramus MT, Neubauer MP. Comparing perception of Presence between Binaural and Stereo Reproduction. 2022. https://aes2.org/publications/elibrary-page/?id=21872
Schubert T, Friedmann F, Regenbrecht H. The experience of Presence: factor Analytic insights. Presence Teleoperators Virtual Environ. 2001;10:266–81. https://doiorg.publicaciones.saludcastillayleon.es/10.1162/105474601300343603.
Pedersen TH, Zacharov N. The development of a sound wheel for reproduced sound. 138th Audio eng soc Conv 2015. 2015;2:765–77. https://aes2.org/publications/elibrary-page/?id=17734
Pedersen TH. Perceptual characteristics of audio. 2015.
Lee H. Multichannel 3D microphone arrays: a review. AES J Audio Eng Soc. 2021;69:5–26. https://doiorg.publicaciones.saludcastillayleon.es/10.17743/JAES.2020.0069.
Carral HL, Empatica, -E4-LSL. 2019. GitHub. https://github.com/HectorCarral/Empatica-E4-LSL
Lee H. 3D microphone array Recording comparison (3D-MARCo). Zenodo. 2019. https://doiorg.publicaciones.saludcastillayleon.es/10.5281/zenodo.3474285.
Lee H. ESMA-3D immersive soundscape recordings. Zenodo. 2019. https://doiorg.publicaciones.saludcastillayleon.es/10.5281/zenodo.3549050.
Naal-Ruiz NE, Lee H, Alonso-Valerdi LM, Ibarra-Zarate DI. Data on EEG, EDA, BVP, psychological responses and audio files used for the study of 3D Audio Immersive Experience. 2024. Figshare. https://doiorg.publicaciones.saludcastillayleon.es/10.6084/m9.figshare.25421464.v3
Acknowledgements
Thanks to members of the NeuroTechs Research group, Acoustic Innovation Research group, and Applied Psychoacoustics Laboratory for assisting in data management and audio calibration.
Funding
This work was supported by the Tecnologico de Monterrey, CONAHCYT and The University of Huddersfield.
Author information
Authors and Affiliations
Contributions
N.E.N.R collected, revised and organized the data presented in this work. He additionally produced the table, and wrote the original draft of the manuscript. H.L., L.M.A.V. and D.I.I.Z supervised, provided funds for the research, reviewed and edited the final version of the manuscript. All authors contributed to the conceptualization and methodology of data collection.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
The Ethics Committee of the School of Computing and Engineering at the University of Huddersfield granted approval for conducting the experiment and sharing the data (Reference: SCE_SREIC_2022/10/05_M_5.1). Informed consent was obtained from all volunteers before experimentation by signing a consent form. Personal information such as name, email and phone number of participants are not included in the data. All procedures were performed in accordance with the Declaration of Helsinki.
Consent for publication
Not applicable.
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License, which permits any non-commercial use, sharing, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if you modified the licensed material. You do not have permission under this licence to share adapted material derived from this article or parts of it. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by-nc-nd/4.0/.
About this article
Cite this article
Naal-Ruiz, N.E., Lee, H., Alonso-Valerdi, L.M. et al. Data on neurophysiological and psychological responses to audio immersive experience in stereo and 3D audio formats. BMC Res Notes 17, 356 (2024). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s13104-024-07009-x
Received:
Accepted:
Published:
DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s13104-024-07009-x