Treffer: AI-based remote photoplethysmography : benchmarking on realistic datasets

Title:
AI-based remote photoplethysmography : benchmarking on realistic datasets
Publisher Information:
Institute of Electrical and Electronics Engineers
Publication Year:
2025
Collection:
University of Malta: OAR@UM / L-Università ta' Malta
Document Type:
Konferenz conference object
Language:
English
DOI:
10.1109/EMBC58623.2025.11253069
Rights:
info:eu-repo/semantics/restrictedAccess ; The copyright of this work belongs to the author(s)/publisher. The rights of this work are as defined by the appropriate Copyright Legislation or as modified by any successive legislation. Users may access this work and can make use of the information contained in accordance with the Copyright Legislation provided that the author must be properly acknowledged. Further distribution or reproduction in any format is prohibited without the prior permission of the copyright holder.
Accession Number:
edsbas.EFDE727
Database:
BASE

Weitere Informationen

In this study, we investigate the performance of the Contrast-Phys AI model for remote photoplethysmography on datasets recorded in more realistic conditions, introducing a new dataset named CAMVISIM LAB for this purpose. The videos in CAMVISIM LAB were recorded in a lightly controlled lab environment, with participants lying 2 m from the camera and allowed to move their heads, speak, and act naturally. The Contrast-Phys model was evaluated on the CAMVISIM LAB data and the UBFC-RPPG and PURE datasets, as the latter two are among the most popular datasets for the performance assessment of rPPG methods. The initial evaluation of the heart rate estimation on CAMVISIM LAB was conducted by training a Contrast-Phys model from scratch using k-fold cross-validation, resulting in a mean absolute error (MAE) of 7.7 bpm. This was improved by using a model pre-trained on the UBFC dataset for rPPG estimation, which improved the result to 5.8 bpm. The error was further minimised by fine-tuning the model with the pretrained weights from UBFC and using our dataset for a second training run. This resulted in an improved MAE of 1.0 bpm when fine-tuning and freezing the initial layers of the Contrast-Phys model during training. This result outperformed the score achieved when using the same strategy on the PURE dataset. ; peer-reviewed