Spatial Sound Scene Synthesis and Manipulation for Virtual Reality and Audio Effects.

Ville Pulkki, Archontis Politis, Tapani Pihlajamäki, Mikko-Ville Laitinen

Research output: Chapter in Book/Report/Conference proceedingChapterScientificpeer-review

3 Citations (Scopus)

Abstract

This chapter covers DirAC-based parametric time-frequency domain (TF domain) audio techniques where the virtual sound scene is synthesized based on a geometric description of it around the avatar. Time-frequency domain parametric spatial audio techniques also have applications in audio engineering, in addition to applications in virtual reality. The basic processing block of DirAC-based spatialization in virtual worlds is the DirAC monosynth. When producing audio for virtual reality, often some recorded speech, music, and/or environmental sounds are used as signals for virtual sources. The chapter presents the application of parametric analysis and synthesis to spatial audio effects, using the DirAC framework. The ambience extraction method attempts to make foreground sounds with a clear direction softer, while preserving background sounds or ambience. Spatialization of a monophonic audio channel, either to a multichannel audio reproduction setup or as background sound in virtual reality, is a useful task in many audio engineering applications.
Original languageEnglish
Title of host publicationParametric Time-Frequency Domain Spatial Audio
EditorsVille Pulkki, Symeon Delikaris-Manias, Archontis Politis
PublisherWiley
Chapter14
Pages347-361
ISBN (Electronic)9781119252634
ISBN (Print)9781119252597
DOIs
Publication statusPublished - 15 Dec 2017
MoE publication typeA3 Book section, Chapters in research books

Keywords

  • DirAC-based spatialization
  • parametric analysis
  • spatial audio techniques
  • time-frequency domain
  • virtual reality

Fingerprint

Dive into the research topics of 'Spatial Sound Scene Synthesis and Manipulation for Virtual Reality and Audio Effects.'. Together they form a unique fingerprint.

Cite this