MetaMGC: a music generation framework for concerts in metaverse

Cong Jin, Fengjuan Wu, Jing Wang*, Yang Liu, Zixuan Guan, Zhe Han

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

18 Citations (Scopus)

Abstract

In recent years, there has been a national craze for metaverse concerts. However, existing meta-universe concert efforts often focus on immersive visual experiences and lack consideration of the musical and aural experience. But for concerts, it is the beautiful music and the immersive listening experience that deserve the most attention. Therefore, enhancing intelligent and immersive musical experiences is essential for the further development of the metaverse. With this in mind, we propose a metaverse concert generation framework — from intelligent music generation to stereo conversion and sound field design for virtual concert stages. First, combining the ideas of reinforcement learning and value functions, the Transformer-XL music generation network is improved and used in training all the music in the POP909 dataset. Experiments show that both improved algorithms have advantages over the original method in terms of objective evaluation and subjective evaluation metrics. In addition, this paper validates a neural rendering method that can be used to generate spatial audio based on a binaural-integrated neural network with a fully convolutional technique. And the purely data-driven end-to-end model performs to be more reliable compared with traditional spatial audio generation methods such as HRTF. Finally, we propose a metadata-based audio rendering algorithm to simulate real-world acoustic environments.

Original languageEnglish
Article number31
JournalEurasip Journal on Audio, Speech, and Music Processing
Volume2022
Issue number1
DOIs
Publication statusPublished - Dec 2022

Keywords

  • Audio digital twin
  • Audio rendering
  • Metaverse concert
  • Neural network
  • Transformer-XL

Fingerprint

Dive into the research topics of 'MetaMGC: a music generation framework for concerts in metaverse'. Together they form a unique fingerprint.

Cite this