Spatial Sound Scene Synthesis and Manipulation for Virtual Reality and Audio Effects.

Ville Pulkki, Archontis Politis, Tapani Pihlajamäki, Mikko-Ville Laitinen

Research output: Chapter in Book/Report/Conference proceedingChapterScientificpeer-review

3 Citations (Scopus)


This chapter covers DirAC-based parametric time-frequency domain (TF domain) audio techniques where the virtual sound scene is synthesized based on a geometric description of it around the avatar. Time-frequency domain parametric spatial audio techniques also have applications in audio engineering, in addition to applications in virtual reality. The basic processing block of DirAC-based spatialization in virtual worlds is the DirAC monosynth. When producing audio for virtual reality, often some recorded speech, music, and/or environmental sounds are used as signals for virtual sources. The chapter presents the application of parametric analysis and synthesis to spatial audio effects, using the DirAC framework. The ambience extraction method attempts to make foreground sounds with a clear direction softer, while preserving background sounds or ambience. Spatialization of a monophonic audio channel, either to a multichannel audio reproduction setup or as background sound in virtual reality, is a useful task in many audio engineering applications.
Original languageEnglish
Title of host publicationParametric Time-Frequency Domain Spatial Audio
EditorsVille Pulkki, Symeon Delikaris-Manias, Archontis Politis
ISBN (Electronic)9781119252634
ISBN (Print)9781119252597
Publication statusPublished - 15 Dec 2017
MoE publication typeA3 Part of a book or another research book


  • DirAC-based spatialization
  • parametric analysis
  • spatial audio techniques
  • time-frequency domain
  • virtual reality


Dive into the research topics of 'Spatial Sound Scene Synthesis and Manipulation for Virtual Reality and Audio Effects.'. Together they form a unique fingerprint.

Cite this