Loading…
Evaluation of spatial audio [clear filter]
Wednesday, October 20
 

1:30pm EDT

Best Paper: Perceptual Evaluation of Interior Panning Algorithms Using Static Auditory Events
Interior panning algorithms enable content authors to position auditory events not only at the periphery of the loudspeaker configuration, but also within the internal space between the listeners and the loudspeakers. In this study such algorithms are rigorously evaluated, comparing rendered static auditory events at various locations against true physical loudspeaker references. Various algorithmic approaches are subjectively assessed in terms of; Overall, Timbral, and Spatial Quality for three different stimuli, at five different positions and three radii. Results show for static positions that standard Vector Base Amplitude Panning performs equal, or better, than all other interior panning algorithms tested here. Timbral Quality is maintained throughout all distances. Ratings for Spatial Quality vary, with some algorithms performing significantly worse at closer distances. Ratings for Overall Quality reduce moderately with respect to reduced reproduction radius and are predominantly influenced by Timbral Quality.

Speakers
TR

Thomas Robotham

International Audio Laboratories Erlangen
AS

Andreas Silzle

International Audio Laboratories Erlangen
AN

Anamaria Nastasa

Aalto University
avatar for Alan Pawlak

Alan Pawlak

PhD Candidate, University of Huddersfield
Alan Pawlak is a final-year PhD candidate at the Applied Psychoacoustics Laboratory (APL) of the University of Huddersfield, specialising in spatial audio and binaural rendering. During his four-year Music Technology and Audio Systems program, Alan completed a year in industry as... Read More →
avatar for Juergen Herre

Juergen Herre

Chief Executive Scientist, International Audio Laboratories Erlangen
Prof. Dr.-Ing. Herre is a fellow member of the Audio Engineering Society (AES), co-chair of the AES Technical Committee on Coding of Audio Signals and chair of the AES Technical Council. In 1989 he joined the Fraunhofer Institute for Integrated Circuits (IIS) in Erlangen, Germany... Read More →


Wednesday October 20, 2021 1:30pm - 2:00pm EDT
Stream A
 
Thursday, October 28
 

9:00pm EDT

3D Impulse Response Convolution with Multichannel Direct Sound: Assessing Perceptual Equivalency between Room- and Source- Impression for Music Production
A method for representing the three-dimensional radiation patterns of instruments/performers within artificial reverberation using multichannel direct sound files convolved with channel-based spatial room impulse responses (SRIRs) is presented. Two reverb conditions are studied in a controlled listening test: a) all SRIR channel positions are convolved with a single monophonic direct sound file, and b) each SRIR channel position is convolved with a unique direct sound file taken from a microphone array surrounding the performer. Participants were asked to adjust the level of each reverberation condition (relative to a fixed direct sound stream) to three perceptual thresholds relating to source- and room- impression. Results of separate three-way within-subject ANOVAs and post-hoc analysis show significant interactions between instrument / room type, and instrument / reverb condition on each of the three thresholds. Most notably, reverb condition b) required less level than condition a) to yield perceptual equivalency between source- and room- impression, suggesting that the inclusion of multichannel direct sound in SRIR convolution may increase the salience of room impression in the immersive reproduction of acoustic music.

Speakers
avatar for Jack Kelly

Jack Kelly

McGill University
Jack Kelly is. Ph.D. candidate at the Schulich School of Music, McGill University. His thesis research centers on the influence of spatial room impulse response convolution technologies (channel-based and HOA arrays) on the sensation of physical presence in immersive music production. He... Read More →
avatar for Richard King

Richard King

McGill University
Richard King is an Educator, Researcher, and a Grammy Award winning recording engineer. Richard has garnered Grammy Awards in various fields including Best Engineered Album in both the Classical and Non-Classical categories. Richard is an Associate Professor at the Schulich School... Read More →
WW

Wieslaw Woszczyk

McGill University


Thursday October 28, 2021 9:00pm - Friday December 3, 2021 6:00pm EST
On-Demand

9:00pm EDT

Bit Rate Requirements for an Audio Codec for Stereo, Surround and Immersive Formats
This paper describes a comprehensive study on the sound quality of the Opus codec for stereo, surround and immersive audio formats for music and cinematic content. We conducted three listening tests on Opus encoded stereo, 5.1 and 7.1.4 test samples taken from music, cinematic and EBU files encoded at bit rates of 32, 48 and 64 kbps per channel. Preliminary results indicate that a bit rate of 64 kbps per channel or higher is required for stereo, but 48 kbps per channel may be sufficient for surround and immersive audio formats.

Speakers
avatar for Sunil G. Bharitkar

Sunil G. Bharitkar

Samsung Research America
avatar for Allan Devantier

Allan Devantier

Samsung Research America
CT

Carlos Tejeda-Ocampo

Samsung Research Tijuana
CZ

Carren Zhongran Wang

Samsung Research America
WS

Will Saba

Samsung Research America


Thursday October 28, 2021 9:00pm - Friday December 3, 2021 6:00pm EST
On-Demand

9:00pm EDT

Comparison of different techniques for recording and postproduction using main-microphone arrays for binaural reproduction.
We present a subjective evaluation of various 3D main-microphone techniques for three-dimensional binaural music production. Forty-seven subjects participated in the survey, listening on headphones. Results suggest that ESMA-3D, followed by Decca tree with height, work best of the included 3D arrays. However, the dummy head and a stereo AB microphone performed as well if not better than any of the arrays. Though not implemented for this study, our workflow allows the possibility to include individualized HRTF's and head-tracking; their impact will be considered in a future study.

Speakers
avatar for Josua Dillier

Josua Dillier

Zürcher Hochschule der Künste ZHdK
Josua Dillier is a young audio engineer and producer living in Zurich, Switzerland. His works range from CD- or Videoproduction to live mixing. He is specialized in the recording of acoustic instruments.Before his studies as a Tonmeister at University of the Arts Zurich he studied... Read More →
HJ

Hanna Järveläinen

Zürcher Hochschule der Künste ZHdK


Thursday October 28, 2021 9:00pm - Friday December 3, 2021 6:00pm EST
On-Demand

9:00pm EDT

Effects of Near-field Sources on Ambisonics Recording and Playback
Ambisonic recording with spherical microphone arrays (SMAs) is based on a far-field assumption which determines how microphone signals are encoded into Ambisonic signals. In the presence of a near-field source, low-frequency distance-dependent boosts arise in SMAs in similar nature to proximity effects in far-field equalized directional microphones. In this study, the effects of near-field sources on Ambisonic signals are modelled analytically, their interaction with regularization stages is observed, and then traced further across to two basic ambisonic processing operations: virtual microphones, and binaural decoding.

Speakers
avatar for Raimundo Gonzalez

Raimundo Gonzalez

Post-Doctoral Researcher, Aalto University
AP

Archontis Politis

Audio & Speech Processing Group, Tampere University of Technology
TL

Tapio Lokki

Department of Signal Processing and Acoustics, Aalto University


Thursday October 28, 2021 9:00pm - Friday December 3, 2021 6:00pm EST
On-Demand

9:00pm EDT

Implementing and Evaluating a Higher-order Ambisonic Sound System in a Multi-purpose Facility: A Lab Report
Although Ambisonic sound reproduction has an extensive history, it started finding more widespread use in the past decade due to the advances in computer hardware that enable real-time encoding and decoding of Ambisonic sound fields, availability of user-friendly software that facilitate the rendering of such sound fields, and recent developments in immersive media technologies, such as AR and VR systems, that prompt new research into spatial audio. In this paper, we discuss the design, implementation, and evaluation of a third-order Ambisonic system in an academic facility that is built to serve a range of functions including instruction, research, and artistic performances. Due to the multi-purpose nature of this space, there are numerous limitations to consider when designing an Ambisonic sound system that can operate efficiently without interfering with the variety of activities regularly carried out in it. We discuss our approach to working around such limitations and evaluating the resulting system. To that end, we present a user study conducted to assess the performance of this system in terms of perceived spatial accuracy. Based on the growing number of such facilities around the world, we believe that the design and evaluation methods presented here can be of use in the implementation of spatial audio systems in similar multi-purpose environments.

Speakers
avatar for Anıl Çamcı

Anıl Çamcı

Associate Professor of Performing Arts Technology, University of Michigan
SS

Sam Smith

University of Michigan
SH

Seth Helman

University of Michigan


Thursday October 28, 2021 9:00pm - Friday December 3, 2021 6:00pm EST
On-Demand

9:00pm EDT

Spatial auditory masking between real sound signals and virtual sound images
In augmented reality (AR) environment, audio signals of real world and virtual world are simultaneously presented to a listener. It is desirable that a virtual sound content and a real sound source do not interfere each other. In order to make it possible, we have examined spatial auditory masking between maskers and maskees, where maskers are real sound signals emitted from loudspeakers, and maskees are virtual sound images, generated by using head related transfer functions (HRTFs), emitted from headphones. Open-ear headphones were used for the experiment, which allow us to listen to the audio content while hearing the environmental sound. The results are very similar to those of the previous experiment [1, 2] where masker and maskee were both real signals emitted from loudspeakers. That is, with a given masker location, masking threshold levels as a function of maskee locations have symmetric property with respect to the frontal plane of a subject. Masking threshold level is, however, lowered than the previous experiment perhaps because of limitation of sound image localization by HRTFs. The results indicate that spatial auditory masking of human hearing occurs with virtually localized sound images in the same way as real sound signals.

Speakers
avatar for Masayuki Nishiguchi

Masayuki Nishiguchi

Professor, Akita Prefectural University
Masayuki Nishiguchi received his B.E., M.S., and Ph.D. degrees from Tokyo Institute of Technology, University of California Santa Barbara, and Tokyo Institute of Technology, in 1981, 1989, and 2006 respectively.  He was with Sony corporation from 1981 to 2015, where he was involved... Read More →
SI

Soma Ishihara

Akita Prefectural University
KW

Kanji Watanabe

Akita Prefectural University
KA

Koji Abe

Akita Prefectural University
ST

Shouichi Takane

Akita Prefectural University


Thursday October 28, 2021 9:00pm - Friday December 3, 2021 6:00pm EST
On-Demand
 
  • Timezone
  • Filter By Date AES Fall Online Convention Oct 11 -31, 2021
  • Filter By Venue Online
  • Filter By Type
  • Acoustics & Psychoacoustics
  • Applications in Audio
  • Archiving & Restoration
  • Audio Builders Workshop
  • Audio for Cinema
  • Broadcast & Online Delivery
  • Diversity & Inclusion
  • E-Briefs on-demand
  • Education
  • Electronic Dance Music
  • Electronic Instrument Design & Applications
  • Game Audio & XR
  • Hip-Hop/R&B
  • Historical
  • Immersive Music
  • Immersive & Spatial Audio
  • Networked Audio
  • Papers on-demand
  • Recording & Production
  • Sound Reinforcement
  • Special Event
  • Tech Tours
  • Technical Committee Meeting
  • Company
  • Subject
  • Area


Filter sessions
Apply filters to sessions.