Online videoconferencing applications have become more popular and occupy a significant portion of consumer Internet traffic. Thus, it is crucial to assess the Quality of Experience (QoE) for these applications to effectively implementing high-quality services. Although the major approach to assess the QoE is the conduction of a subjective quality assessment, this method relies on user's feedback and it is not suitable for real-time management systems. For this reason, in this paper, we propose alternative QoE estimation models based on facial expression features extracted from the face of the users during audiovisual conversations with a WebRTC-based telemeeting service. We first conducted a subjective quality assessment to investigate the QoE of two-party audiovisual conversations impaired by poor network conditions. Then, we used the extracted facial expression features obtained by observing the test participants to train Machine Learning (ML) algorithms and a Fully Convolutional Network (FCN) with the aim to estimate the perceived QoE solely on the base of these features. Finally, we discuss and compare the QoE estimation performance of the proposed models in terms of accuracy, precision, recall, and F1-score metrics. In particular, the ML model (Support Vector Machine) and the FCN achieved the mean estimation accuracy of 0.78 and 0.70, respectively.

QoE Estimation of WebRTC-based Audiovisual Conversations from Facial Expressions

Bingol, G;Porcu, S;Floris, A;Atzori, L
2022-01-01

Abstract

Online videoconferencing applications have become more popular and occupy a significant portion of consumer Internet traffic. Thus, it is crucial to assess the Quality of Experience (QoE) for these applications to effectively implementing high-quality services. Although the major approach to assess the QoE is the conduction of a subjective quality assessment, this method relies on user's feedback and it is not suitable for real-time management systems. For this reason, in this paper, we propose alternative QoE estimation models based on facial expression features extracted from the face of the users during audiovisual conversations with a WebRTC-based telemeeting service. We first conducted a subjective quality assessment to investigate the QoE of two-party audiovisual conversations impaired by poor network conditions. Then, we used the extracted facial expression features obtained by observing the test participants to train Machine Learning (ML) algorithms and a Fully Convolutional Network (FCN) with the aim to estimate the perceived QoE solely on the base of these features. Finally, we discuss and compare the QoE estimation performance of the proposed models in terms of accuracy, precision, recall, and F1-score metrics. In particular, the ML model (Support Vector Machine) and the FCN achieved the mean estimation accuracy of 0.78 and 0.70, respectively.
2022
978-1-6654-6495-6
Quality of Experience
WebRTC
Facial Expressions
Subjective assessment
Machine Learning
Fully Convolutional Network
File in questo prodotto:
File Dimensione Formato  
post-2022-10 SITIS - QoE Estimation of WebRTC-based.pdf

embargo fino al 11/04/2025

Tipologia: versione post-print
Dimensione 308.72 kB
Formato Adobe PDF
308.72 kB Adobe PDF   Visualizza/Apri   Richiedi una copia
pub-2022-10 SITIS - QoE Estimation of WebRTC-based.pdf

Solo gestori archivio

Tipologia: versione editoriale
Dimensione 1.29 MB
Formato Adobe PDF
1.29 MB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11584/380843
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? 1
social impact