학술논문

Perception and automated assessment of audio quality in user generated content: An improved model
Document Type
Conference
Source
2016 Eighth International Conference on Quality of Multimedia Experience (QoMEX) Quality of Multimedia Experience (QoMEX), 2016 Eighth International Conference on. :1-6 Jun, 2016
Subject
Signal Processing and Analysis
Algorithm design and analysis
Degradation
Frequency measurement
User-generated content
Acoustic distortion
Data models
audio quality
perception
audio quality of experience
distortion
wind noise
handling noise
Language
Abstract
Technology to record sound, available in personal devices such as smartphones or video recording devices, is now ubiquitous. However, the production quality of the sound on this user-generated content is often very poor: distorted, noisy, with garbled speech or indistinct music. Our interest lies in the causes of the poor recording, especially what happens between the sound source and the electronic signal emerging from the microphone, and finding an automated method to warn the user of such problems. Typical problems, such as distortion, wind noise, microphone handling noise and frequency response, were tested. A perceptual model has been developed from subjective tests on the perceived quality of such errors and data measured from a training dataset composed of various audio files. It is shown that perceived quality is associated with distortion and frequency response, with wind and handling noise being just slightly less important. In addition, the contextual content of the audio sample was found to modulate perceived quality at similar levels to degradations such as wind and rendering those introduced by handling noise negligible.