Options
Analyzing the performance of autoencoder-based objective quality metrics on audio-visual content
Author(s)
Date Issued
2020-01-30
Date Available
2020-04-29T14:57:27Z
Abstract
The development of audio-visual quality models faces a number of challenges, including the integration of audio and video sensory channels and the modeling of their interaction characteristics. Commonly, objective quality metrics estimate the quality of a single component (audio or video) of the content. Machine learning techniques, such as autoencoders, offer as a very promising alternative to develop objective assessment models. This paper studies the performance of a group of autoencoder-based objective quality metrics on a diverse set of audio-visual content. To perform this test, we use a large dataset of audio-visual content (The UnB-AV database), which contains degradations in both audio and video components. The database has accompanying subjective scores collected on three separate subjective experiments. We compare our autoencoder-based methods, which take into account both audio and video components (multi-modal), against several objective (single-modal) audio and video quality metrics. The main goal of this work is to verify the gain or loss in performance of these single-modal metrics, when tested on audio-visual sequences.
Sponsorship
Science Foundation Ireland
Other Sponsorship
Insight Research Centre
Type of Material
Conference Publication
Publisher
Society for Imaging Science and Technology
Copyright (Published Version)
2020 Society for Imaging Science and Technology
Language
English
Status of Item
Peer reviewed
Conference Details
The 2020 IS&T International Symposium on Electronic Imaging (EI2020), Burlingame, California, 26-30 January 2020
ISSN
2470-1173
This item is made available under a Creative Commons License
File(s)
Loading...
Name
insight_publication.pdf
Size
2.95 MB
Format
Adobe PDF
Checksum (MD5)
0f6baad05f6ce868076628127d81a976
Owning collection
Mapped collections