Going the Extra Mile in Face Image Quality Assessment: A Novel Database and Model

Shaolin Su, Hanhe Lin, Vlad Hosu, Oliver Wiedemann, Jinqiu Sun, Yu Zhu, Hantao Liu, Yanning Zhang, Dietmar Saupe

Research output: Working paper/PreprintPreprint

Abstract

Computer vision models for image quality assessment (IQA) predict the subjective effect of generic image degradation, such as artefacts, blurs, bad exposure, or colors. The scarcity of face images in existing IQA datasets (below 10\%) is limiting the precision of IQA required for accurately filtering low-quality face images or guiding CV models for face image processing, such as super-resolution, image enhancement, and generation. In this paper, we first introduce the largest annotated IQA database to date that contains 20,000 human faces (an order of magnitude larger than all existing rated datasets of faces), of diverse individuals, in highly varied circumstances, quality levels, and distortion types. Based on the database, we further propose a novel deep learning model, which re-purposes generative prior features for predicting subjective face quality. By exploiting rich statistics encoded in well-trained generative models, we obtain generative prior information of the images and serve them as latent references to facilitate the blind IQA task. Experimental results demonstrate the superior prediction accuracy of the proposed model on the face IQA task.
Original languageEnglish
Place of PublicationCornell University
PublisherarXiv
DOIs
Publication statusPublished - 11 Jul 2022

Keywords

  • cs.CV
  • Computer Vision and Pattern Recognition

Fingerprint

Dive into the research topics of 'Going the Extra Mile in Face Image Quality Assessment: A Novel Database and Model'. Together they form a unique fingerprint.

Cite this