Novel full reference perceptual quality metric for audio-visual asynchrony

Yao Du Wei*, Xiang Xie, Jing Ming Kuang, Xin Lu Han

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

2 Citations (Scopus)

Abstract

A full reference model was proposed to evaluate the perceptual quality of audiovisual asynchrony. A standard synchronization process was used to determine the time difference between audio and video. The mapping between the time difference and the perceptual quality was derived by co-inertia analysis. The co-inertia analysis extracted the most related component from audio and video features, and then formed a mapping for each audiovisual sequence. Audiovisual contents were divided into three categories: clean speech, non speech and mixed speech. The clean speech category was further split into two subcategories. Audio and video features were chosen separately for each category. Subjective test results showed that the proposed model conforms well with subjective results.

Original languageEnglish
Pages (from-to)182-190
Number of pages9
JournalTongxin Xuebao/Journal on Communications
Volume33
Issue number2
Publication statusPublished - Feb 2012

Keywords

  • Audiovisual quality assessment
  • Co-inertia analysis
  • Signal processing technique
  • Synchrony

Fingerprint

Dive into the research topics of 'Novel full reference perceptual quality metric for audio-visual asynchrony'. Together they form a unique fingerprint.

Cite this