Abstract
Heart rate (HR) estimation is crucial for early cardiovascular diagnosis, continuous monitoring, and various health applications. While electrocardiography (ECG) remains the gold standard, its discomfort and impracticality for continuous use have spurred the development of non-contact methods like remote photoplethysmography (rPPG). This systematic review (PROSPERO: CRD 42024592157) examines 70 studies to assess the impact of Region of Interest (ROI) selection on HR estimation accuracy. Most methods (36.8%) use the holistic face, while forehead and cheek areas (24.5% and 21.7%) show superior accuracy. Machine learning-based approaches outperform traditional methods under motion artifacts and poor lighting, achieving Mean Absolute Error and Root Mean Square Error below 1.0 for some datasets. Combining multiple patches improves performance, though increasing ROIs beyond 60 patches results in diminishing returns and higher computational complexity. These findings highlight the significance of ROI optimization for robust rPPG-based HR estimation.
Similar content being viewed by others
Introduction
Heart rate (HR) and blood pressure (BP) estimation are essential for early detection of cardiovascular diseases, continuous health monitoring, emotion detection, and assessing other vital parameters. Although electrocardiography (ECG) is the gold standard for HR detection due to its high accuracy, it has notable drawbacks, including high equipment costs, user discomfort, and challenges with continuous, everyday monitoring. To address these limitations, photoplethysmography (PPG) was developed, which accurately measures HR by detecting light reflected from a wearable device, such as an oximeter. Traditional PPG still relies on contact-based devices, such as finger sensors and wearables, which, while effective in controlled environments, present significant limitations in real-world applications. These devices can cause discomfort during prolonged use and may fail in scenarios involving delicate or compromised skin conditions, such as burns, eczema, or post-surgical recovery. Furthermore, populations such as neonates, elderly patients, and individuals in intensive care units often face additional risks or challenges with contact-based monitoring. In neonates, especially those in intensive care, traditional contact-based heart rate monitoring can be problematic due to their delicate skin and the risk of injury. Studies have explored alternative methods, such as forehead monitoring of heart rate in neonatal intensive care, to address these challenges1. In these contexts, remote photoplethysmography (rPPG) offers a transformative solution, enabling non-invasive and continuous heart rate monitoring using standard cameras, which could significantly enhance healthcare delivery in both clinical and at-home settings.
The rPPG was introduced to overcome these challenges. Unlike traditional PPG, rPPG estimates HR using standard RGB cameras by detecting subtle light changes on a subject’s face, eliminating the need for direct physical contact, however bringing comparable results2. The most effective regions for capturing these skin color changes are areas with lower skin density, such as the cheeks and forehead3,4. Despite its advantages, rPPG faces several challenges, including sensitivity to varying lighting conditions, movement artifacts, false face detection, and obstructions caused by masks, hair, or clothing. To filter noisy signals, a Signal Quality Index (SQI) was introduced, with a threshold value NSQI < 0.2935. Achieving optimal accuracy in rPPG also requires maintaining a stable position, an appropriate distance from the camera, and minimizing motion.
Early traditional rPPG methods, such as Principal Component Analysis (PCA)6, Independent Component Analysis (ICA)7, the chrominance-based CHROM8 method, the GREEN channel9 approach, and the Plane-Orthogonal-to-Skin (POS)10 algorithm, focused primarily on noise reduction. These methods typically began by detecting the face and defining a Region of Interest (ROI) based on empirical knowledge, followed by extracting pulse signals from the RGB channels within the ROIs. HR was then derived from the signals using techniques like Discrete Fourier Transform (DFT) and peak detection.
With the rise of machine learning (ML), new approaches11 have emerged that utilize neural network models, particularly convolutional neural networks (CNNs), to extract spatiotemporal features from facial videos. Examples include Dual-Gan12, TranPulse13, and MAR-rPPG14. These methods aim to develop robust end-to-end network structures capable of accurately estimating physiological parameters under real-world conditions by enhancing temporal correlations and minimizing redundant information. Unlike traditional approaches, which rely on handcrafted features derived from skin texture, rPPG signals, and manually defined ROIs, ML methods improve HR estimation quality and significantly reduce reliance on clean input data. On the other hand, ML-based methods require labeled data for training and are more computationally difficult, when the traditional methods could be efficient and low-complexity15. Apart from HR and BP16 an estimation of HRV (Heart Rate Variability), SpO2 (Blood Oxygen Saturation) and anxiety detection17 are possible. In this paper, we conduct a systematic review of 70 studies from the past decade, assessing different algorithms for HR estimation. We examine 20 commonly used datasets and evaluate the impact of ROI selection on HR estimation through rPPG.
Figures 1 and 2 provide a synthesized flow diagram of the typical remote photoplethysmography processing pipeline, based on our literature review. This chart presents a high-level overview of commonly employed rPPG-based physiological monitoring stages. While this is a generalized representation that may not capture every method in its entirety, it highlights critical steps within the rPPG pipeline. The flow diagram emphasizes the importance of ROI segmentation, a critical step in rPPG, as it significantly impacts the quality and reliability of the extracted signal. ROIs typically include facial areas with high blood perfusion, such as the forehead and cheeks regions, to provide more stable rPPG signals. Common techniques such as superpixel segmentation, skin segmentation, and triangulation help isolate relevant skin areas and enhance spatial smoothness during data processing.
a Face Detection algorithms with Color Channel Enhancement and Photometric Normalization (Illumination Equalization, Color Normalization)—the light-blue area with a dashed border; followed by ROI Selection methods and Pre-Processing (Bandpass Filter, Baseline Correction, Signal Filtering Gaussian/Median) for (b) rPPG Signal detection and filtering—the light-purple area with a dashed border. Next, (c) HR detection—the light-yellow area with a dashed border. Either using ML-based methods(Option 1: the light-red area with a dashed border) with labeled data to estimate HR, or using traditional methods (Option 2: the light-brown area with dashed border) without training data for HR prediction, e.g., via PSD or FFT.
After ROI segmentation, the flow chart illustrates several pre-processing steps commonly used to enhance signal quality, including color channel selection, photometric normalization, and signal filtering. These steps help mitigate noise from environmental factors (e.g., lighting changes) and movement artifacts. The flow diagram also differentiates between traditional methods (e.g., CHROM, ICA, PCA, GREEN) and ML-based approaches, providing a snapshot of the variety of available techniques for physiological signal extraction.
Traditional methods, such as Principal Component Analysis (PCA), Independent Component Analysis (ICA), and Fast Fourier Transform (FFT), rely on predefined algorithms and handcrafted features, making them computationally efficient and cost-effective for consumer devices. However, these methods often struggle with noise and variability in real-world conditions. ML-based methods, such as convolutional neural networks (CNNs), automatically learn complex spatiotemporal patterns from data, excelling in challenging scenarios involving motion artifacts or varying lighting. Despite their accuracy, ML approaches are computationally intensive, requiring substantial resources for training and deployment, which may increase costs for healthcare systems and limit their scalability in consumer-grade devices. Hybrid approaches that blend the efficiency of traditional methods with ML’s adaptability could offer a promising middle ground.
Furthermore, contact-based PPG devices, while effective in controlled settings, often show reduced performance across diverse skin tones due to varying melanin levels that affect light absorption and reflection. Additionally, conditions that induce peripheral vasoconstriction, such as hypothermia or cardiovascular diseases, can impair signal quality by reducing blood flow in extremities18. In contrast, remote PPG (rPPG) leverages facial regions with relatively stable blood perfusion, such as the forehead and cheeks, making it less affected by these limitations and better suited for diverse populations and clinical conditions.
Overall, this flow diagram does not represent a comprehensive, step-by-step process but rather serves as a conceptual overview of the rPPG processing pipeline, with an emphasis on ROI selection, pre-processing techniques, and physiological parameter estimation.
Results
Publications
This literature review identified a total of 70 studies evaluating various algorithms for heart rate estimation using remote photoplethysmography (rPPG), as illustrated in Fig. 3. The search yielded 39 articles from PubMed, 80 from IEEE Xplore, and 14 from Embase. After an initial screening, 11 duplicate studies and 3 articles that were inaccessible or incompatible were excluded. Of the remaining articles, 49 were deemed ineligible: 32 did not focus on HR or BP detection, 8 utilized additional technologies such as infrared or near-infrared (NIR) imaging or employed contact-based devices, 5 were theoretical papers, and 4 did not primarily collect data from facial regions. This selection process identified a notable trend: an increasing focus on ROI selection, with studies reporting optimal results for algorithms that incorporate multiple ROIs.
A notable trend observed in these studies is the increasing focus on ROI selection, with studies reporting optimal results for algorithms that incorporate multiple ROIs19 as shown in Fig. 4 with a total of 33 values for MAE and 29 values for RMSE. Table 1 highlights key variables influencing ROI performance, such as the predominance of participants with lighter skin tones and the geographic concentration of datasets. This lack of diversity underscores the need for future studies to incorporate participants with varied ethnicities and skin tones, particularly from underrepresented regions, to better generalize ROI-based HR estimation techniques. Another key observation is the shift towards machine learning (ML) techniques in HR estimation algorithms. ML-based approaches have shown enhanced accuracy, especially for challenging cases involving motion, using publicly available datasets. Recent top-performing methods12,20,21 commonly employ combinations of multiple facial ROIs, in contrast to earlier studies, which predominantly relied on single-region analysis.
Characteristics of participants in datasets
Most of the reviewed studies utilize publicly available datasets, with 20 studies22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41 relying exclusively on in-house datasets for testing, and 8 studies42,43,44,45,46,47,48,49 combining both public and in-house datasets. Two studies50,51 do not specify details about the testing subjects. The datasets predominantly feature healthy male and female participants, typically aged 18 to 60. Most videos in these datasets are 5 minutes or shorter, though the BIDMC dataset52 includes videos up to 8 minutes. Video resolutions range from 1920 × 1080 to 640 × 480 pixels, and filming conditions vary widely in terms of subject movement, head rotations, physical activities, and lighting conditions.
The most frequently used datasets are UBFC-rPPG53, PURE54, COHFACE55, VIPL-HR56, and MAHNOB-HCI57, with 27, 24, 13, 11, and 6 mentions, respectively. UBFC-rPPG demonstrates relatively low error rates due to its steady setup, with controlled indoor illumination and variable sunlight exposure, as subjects engage in a quiz. The PURE dataset incorporates six different scenarios: steady, talking, slow translation, fast translation, slow rotation, and medium rotation. VIPL-HR generally reports higher error rates, likely due to its nine distinct conditions that include various head movements and lighting conditions, making it a valuable benchmark for assessing algorithm stability. It is important to note that most datasets, whether publicly available or in-house, contain a larger proportion of male subjects compared to female subjects. This gender imbalance may introduce biases and reduce accuracy in real-world applications.
Types of camera devices and reference systems
All dataset videos are recorded using standard RGB cameras (non-IR/NIR), with ground-truth HR measurements provided by finger pulse oximeters or wrist-worn devices. Table 1 includes detailed information on the specific filming settings and reference devices used across studies. It highlights critical variables relevant to ROI selection, including skin color, ethnicity, and geographic distribution of participants. Notably, most datasets either lack explicit reporting of participants’ skin tones or primarily include lighter-skinned individuals. This limitation can bias the effectiveness of ROI-based HR estimation, as skin color significantly impacts light absorption and reflection, which are critical for rPPG signal quality. Expanding datasets to include participants from diverse ethnic backgrounds and geographic locations is essential to better understand how ROI performance varies across skin tones, lighting conditions, and cultural contexts. Such efforts would ensure the development of rPPG technologies that are robust and equitable across global populations.
ROIs used for HR estimation
Multiple studies emphasize the critical role of Region of Interest (ROI) selection in the accuracy of heart rate (HR) estimation, noting that inappropriate ROI choices can introduce significant errors. Kwon et al.58 divided the face into seven regions to evaluate each for signal quality, while Poh et al.59 recommended an ROI that spans 60% of the full face’s width and its entire height. Later, Zhao et al.60 focused on an ROI below the eye line, covering skin areas around the nose, mouth, and cheeks.
According to Dae-Yeol Kim et al.4, facial areas with larger surface areas and thinner skin, such as the cheeks and forehead, tend to yield more reliable results and are less affected by light reflections. In contrast, the nose area is considered less reliable for capturing skin color changes. Kim et al. proposed five specific facial regions (TOP-5) characterized by lower variability and noise, with an average skin thickness of 1191.11 μm and a pixel count of 2431. Detailed information regarding ROIs, their size, skin thickness and visual representation is shown in Table 2. According to Kim et al.4, using those facial patches improves the MAE on the UBFC dataset for POS and CHROM methods from 1,87 to 1,85 and from 2.67 to 1.5 accordingly. Similarly, MAE for PPGI dataset is improved from 4.04 to 3.61 for POS and from 4.04 to 2.93 for CHROM methods.
Similarly, Li et al.61 analyzed 28 facial regions defined anatomically and identified the most effective ROIs: the glabella, medial forehead, left and right lateral forehead, left and right malar regions, and upper nasal dorsum. Among these, the glabella demonstrated the best overall performance across both motion and cognitive datasets.
Furthermore, due to the symmetry of cheek areas, it might be more robust to the noise and can enable dynamic substitution of ROIs if some part of the face is obscured due to the head rotation62. Furthermore, the forehead and cheeks have a larger flat area, which has a positive impact on the Signal-to-noise ratio (SNR). Apart from that, there could be benefits of using those areas, as they are mostly free of facial hair, accessories and facial expression change.
In real-world applications, certain facial regions may be obscured by facial hair, accessories, shadows, or masks, necessitating more precise ROI selection techniques. Some studies37,63,64 have adopted superpixel segmentation to isolate facial ROIs, excluding non-skin regions such as the mouth and eyes. This technique allows for the segmentation of skin areas with irregular shapes, unlike block-based segmentation. As shown in Fig. 5, the forehead and cheeks are the most frequently selected regions, used 26 and 23 times, respectively. The nose and chin appear less frequently, in 7 and 5 articles, respectively. However, a majority of studies (39) utilize the entire holistic face. Yaran Duan et al.62 introduced a self-adaptive ROI pre-tracking and signal selection method to mitigate motion artifacts using 18 facial patches. The ROIs are continuously tracked, with their visibility dynamically assessed based on the motion state. Similarly,24 implemented a symmetry substitution approach, where data from visible areas of the left and right cheeks are symmetrically replicated when facial rotation (30-45 degrees) obscures certain regions.
The top row illustrates the percentage of studies utilizing specific ROIs, including the whole face (36.8%), forehead (24.5%), cheeks (21.7%), nose (6.6%), chin (4.7%), and other regions (5.7%). Gaussian probability density plots show the normalized intensity distributions within each ROI across studies. The bottom row depicts the spatial intensity maps (viridis colormap) of pixels within each ROI, highlighting areas of high and low intensity. Asymmetric ROIs, such as those from the nose and eyes, are shaped by the Mediapipe algorithm, which uses triangular regions that may reduce accuracy. Environmental factors, such as lighting, reflections, and head rotation, also impact ROI selection algorithms.
Haoyuan Gao et al.19 recommend an optimal range of 30 to 40 triangular ROIs on the face, cautioning that an excessive number of ROIs may lead the model to behave more as a face detection system than as an HR estimation tool. In their study, they utilized Delaunay triangulation to create 898 triangular ROIs. Recent works19,20,21,43 favor using multiple ROI combinations rather than averaging values from facial patches. Figure 6 illustrates the increasing trend in the number of ROIs used per study, which correlates with improved performance evaluations. This suggests that employing a larger number of facial ROI combinations can enhance algorithmic accuracy, though further data is needed for more robust statistics.
Because this work is a systematic literature review rather than an experimental study, we did not implement a common rPPG-to-HR extraction pipeline across datasets. Instead, we summarized the published performance that each primary study achieved with its own signal processing chain and then isolated the variable of interest, the number, and arrangement of ROIs, when comparing results. Consequently, absolute error values still reflect the underlying algorithmic diversity (e.g., CHROME, POS, Dual-GAN, etc.), but the review keeps the discussion centered on how ROI granularity modulates those outcomes. Future benchmark work in which identical extraction code is reapplied to every ROI configuration would be valuable, yet it lies beyond the scope of the present survey, whose aim is to map current practice rather than reimplement it.
Performance evaluation
Figure 4 shows a trend toward an increasing number of ROIs across studies, along with improved Mean Absolute Error (MAE) and Root Mean Square Error (RMSE) values on datasets such as PURE54, COHFACE55, and UBFC-rPPG53. To capture non-linear data trends, we included a LOESS (Locally Estimated Scatterplot Smoothing) function. Unlike linear regression, LOESS fits a series of localized regressions across subsets of the data, generating a smooth curve that adapts to pattern changes, thus providing a more accurate visualization of the relationship between ROI quantity and error metrics (MAE and RMSE). The Pearson Correlation Coefficient (PCC) between the number of ROIs and error shows a moderate negative linear relationship, calculated as follows:
where xi and yi are points at lag i of the rPPG and PPG signals, respectively. \(\bar{x}\) and \(\bar{y}\) represent their means. N is the number of points of the discrete signals.
Equations (2) Mean Absolute Error (MAE) and (3) Root Mean Square Error (RMSE) are commonly used metrics for evaluating error in models. are expressed as follows:
where N is the number of points and xi, yi are the points at lag i of the rPPG and contact PPG signals, respectively.
From Figs. 4 and 6, it is evident that more recent, high-accuracy studies often use a larger number of smaller facial patches or combinations thereof. Taking more patches (e.g. more than 60) could slightly increase or even have the same performance but with much more computational effort. However, establishing a strong correlation between ROI quantity and error metrics (MAE/RMSE) remains challenging due to the limited number of studies with clearly documented ROI selection on public datasets. Additionally, as noted by Gao et al.19, an excessive number of ROIs can reduce accuracy. However, their analysis focused on the entire face rather than specific regions like the forehead and cheeks, which recent research has frequently emphasized.
As of October 2024, the best results on publicly available datasets, in terms of Mean Absolute Error (MAE) and Root Mean Square Error (RMSE), have been achieved by the following studies:
-
Qian et al.20, utilized up to 64 distinct ROI patches (ROI combinations for noise reduction), distributed symmetrically across the face to ensure spatial diversity and enhance rPPG prediction. Introduced Spatial TokenLearner for identifying the most informative ROIs while suppressing noisy regions, and Temporal TokenLearner to mitigate disturbances such as motion artifacts and illumination changes.
-
Zhao et al.14, Focused on precise ROI localization using attention regularization techniques. Improved robustness to motion artifacts and inconsistencies in ROI localization by utilizing MediaPipe, Masked Attention Regularization (MAR) and Enhanced rPPG Expert Aggregation (EREA).
-
Hao Lu et al.12, proposed a Dual-GAN architecture to jointly model BVP predictors and noise distributions. Designed an ROI Alignment and Fusion (ROI-AF) block to align features across ROIs and address inconsistencies in noise distributions and used adversarial learning to disentangle BVP and noise components, enhancing robustness to environmental and physiological noise.
-
Si-Qi Liu et al.65, introduced a lightweight spatiotemporal convolutional network (STConv) for rPPG estimation with a noise-disentangling module to separate environmental noise from physiological signals, guided by background features. Also, applied adaptive ROI selection for robustness across datasets and scenarios.
-
Yaran Duan et al.62, divided the facial region into 18 small circular sub-ROIs, and grouped symmetrically into 9 main ROIs for robust tracking. Introduced a self-adaptive tracking system to discard occluded or noisy ROIs and retain their visible counterparts and applied spectral analysis to select the best-quality signal from available ROIs dynamically.
Table 3 provides specific MAE and RMSE values for the top five datasets. Overall, there is a clear trend toward employing a growing number of ROI combinations to improve heart rate estimation accuracy, particularly through machine learning-based techniques.
It is important to note that these approaches have been tested on a variety of datasets with differing levels of complexity. This variability can result in lower performance on datasets containing motion artifacts, particularly in terms of MAE and RMSE. Nonetheless, the findings consistently show that recent machine learning methods with multiple ROIs outperform earlier techniques across all publicly available datasets.
Influence of light and motions to accuracy
Several well-known challenges complicate heart rate estimation via remote photoplethysmography (rPPG). Light reflections on the subject’s facial skin can cause inaccurate signal readings, leading to errors in heart rate predictions. Additionally, poor video quality-resulting from factors such as the subject’s distance from the camera, insufficient lighting, improper face detection, occluded facial regions, and excessive movement-can significantly impact precision.
Many datasets used in rPPG research were collected in controlled environments53,55, predominantly featuring young to middle-aged white male subjects. As a result, models trained on these datasets often struggle to accurately predict heart rates across diverse populations or under real-world conditions. In contrast, the VIPL dataset56 presents a more challenging scenario by incorporating various conditions, including head movements, rotations, speech, facial expressions, and both low and high lighting levels. These factors contribute to the dataset’s complexity and are evident in performance analyses, underscoring the need for models capable of handling diverse environmental and demographic variables.
Across the reviewed studies, commonalities in data exclusion practices were observed, particularly in frames where face-detection algorithms (e.g., Dlib, MediaPipe, Viola-Jones) failed to accurately capture the face. Additionally, most studies excluded regions such as the eyes, mouth, and areas with facial hair from analysis. When certain facial patches became obscured due to head rotation or other factors, they were often excluded from rPPG estimation or substituted with their symmetrical counterparts to maintain data integrity.
Discussion
The objective of this review was to assess the existing literature on challenges associated with region of interest (ROI) selection for heart rate estimation in remote photoplethysmography (rPPG) algorithms. We conducted a comprehensive search to identify relevant studies, focusing on how the choice and quantity of facial patches influence the performance of rPPG-based heart rate estimation.
This review identifies several limitations that warrant consideration. First, despite conducting an extensive search across PubMed, IEEE Xplore, and Embase, some relevant studies may have been inadvertently missed. Second, certain studies included in this review did not specify details such as the exact number of ROIs analyzed, the datasets used, or the performance evaluation methods applied. These gaps may affect the completeness of the presented data and could influence the interpretation of results. Although these limitations do not detract from the overall value of this review, they highlight the need for future research to provide more comprehensive reporting to deepen insights into this field.
As illustrated in Fig. 4, using a greater number of ROIs generally improves estimation quality, although it remains unclear whether additional patches reduce error or simply add noise. While the diversity of algorithms and ROI selection strategies complicates establishing a definitive trend, our findings suggest a potential negative correlation between the number of ROIs and error metrics, such as Mean Absolute Error (MAE) and Root Mean Square Error (RMSE). However, we recommend conducting further studies with a larger number of ROIs and across diverse datasets to validate or refute the presence of a linear relationship. Nevertheless, findings suggest that using more than four ROIs often enhances accuracy by mitigating noise. Conversely, an excessive number of patches (e.g., more than 100) may introduce additional noise, potentially diminishing accuracy. So, the error for 6 ROIs and 60 ROIs is relatively the same, however, for a second case we require more timing and computational resources. For optimal results, multiple sub-ROIs from areas like the forehead and cheeks, or combinations thereof, may provide the most reliable outcomes.
Based on this review, we recommend the following directions for future research:
-
A key research priority for future studies is to investigate whether dividing the face into additional sub-ROIs enhances rPPG signal quality or inadvertently introduces more noise. Addressing this question could provide critical insights into optimizing ROI segmentation strategies for improved accuracy in heart rate estimation. We suggest performing research on diverse datasets by consistently dividing facial ROIs into sub-ROIs to find an optimal distribution threshold, based on the work of Li et al.61.
-
Prioritize the selection and segmentation of facial ROIs, with a specific focus on the forehead and cheeks. These areas, particularly the forehead and cheeks, can be further divided into smaller sub-ROIs, as studies have shown that increasing the number of well-defined patches improves signal quality and enhances HR estimation accuracy, particularly under challenging conditions like motion artifacts or poor lighting. For instance, recent research utilizing more than three ROIs19 or employing up to 2n − 1 ROI combinations20,21 has consistently reported lower error rates on benchmark datasets. This highlights the potential of sub-ROI segmentation to improve robustness in real-world applications, though further validation is needed to confirm these findings across diverse datasets.
-
Improve rPPG signal quality by dynamic ROIs selection and enhancement based on SQI, SNR, histogram and find an optimal threshold value to all those parameters, similar to 0.293 value for SQI5. Additionally, resample shared datasets to overcome limited sample sizes and model overfitting.
-
Increase the diversity of testing and validation datasets by including more female subjects from varied age groups and racial backgrounds. Additionally, incorporating factors, such as motion, make-up, sweat, facial accessories, distance from camera and light source, multiple subjects and partial facial occlusion, and diverse lighting conditions would contribute to more robust and realistic outcomes in real-world applications.
-
Adaptive ROI Orchestration: a promising extension of this review is to quantify and optimize the robustness-versus-flexibility trade-off that arises as the number of ROIs grows. We propose evaluating dynamic ROI-selection frameworks in which the analysis engine continuously ranks each patch by objective quality indicators (e.g., SQI, SNR, or PCC) and, at run time, activates only the subset that exceeds a reliability threshold. Such an adaptive scheduler would:
-
1.
Fallback for occlusions. Seamlessly switch to alternative patches dynamically when a region is covered by hair, glasses, or head motion. Noise suppression. Exclude low-quality ROIs (based on a single or multiple quality indicators) on a frame-by-frame basis, preventing the dilution of the composite rPPG signal.
-
2.
Resource awareness. Limit the active ROI set when computational or energy budgets are tight, then re-expand it when resources allow.
-
3.
Exploring these mechanisms, ideally across datasets with controlled occlusion and motion scenarios, will help establish principled guidelines on when to favor holistic, sparse, or dense ROI configurations and how to transition between them automatically.
-
1.
Methods
Registration and protocol
This review was registered in the PROSPERO database (ID: CRD42024592157) before its initiation. During the preparation of this work, we used ChatGPT (version GPT-4o, OpenAI) to optimize the readability. After using this tool, authors reviewed and edited the content of the manuscript as required and took full responsibility for the publication and its content.
Literature search and selection criteria
Following PRISMA guidelines for systematic reviews, we conducted a comprehensive literature search across IEEE Xplore, PubMed, and Embase, focusing on studies published from January 2014 to October 2024 to capture recent advancements in heart rate (HR) estimation via remote photoplethysmography (rPPG). The search terms included ‘ROI,’ ‘region of interest,’ ‘patch,’ ‘video,’ ‘image,’ ‘camera,’ ‘smartphone,’ ‘rPPG,’ ‘remote photoplethysmogram,’ ‘transdermal optical imaging,’ ‘TOI,’ ‘heart rate,’ ‘HR,’ ‘blood pressure,’ ‘face,’ and ‘facial,’ with Boolean operators applied to expand the scope of relevant studies. The initial search was conducted by one author (M.B.), with a second author (M.E.) independently verifying the results to ensure accuracy. Studies were included if they reported on HR or BP estimation using facial regions captured by a standard RGB camera. Exclusion criteria encompassed duplicate publications, inaccessible articles (i.e., those without full-text availability), studies irrelevant to HR or BP prediction, review articles, theoretical papers, studies using non-RGB cameras (e.g., NIR cameras), and studies employing non-contactless devices or analyzing non-facial regions.
Data analysis and statistical approach
In this review, we systematically analyzed the impact of ROI selection on heart rate estimation accuracy by assessing algorithmic errors, specifically Mean Absolute Error (MAE) and Root Mean Squared Error (RMSE), across publicly available datasets. Given the variety of datasets used for training and testing, establishing a definitive optimal approach is challenging. However, statistical evidence indicates that recent machine learning-based methods utilizing combinations of multiple ROIs generally achieve superior performance.
To enhance the selection of ROIs beyond statistical data, we recommend incorporating a Signal Quality Index (SQI) analysis for each ROI’s rPPG signal. By applying a threshold value of 0.293, ROIs and corresponding signals with excessive noise can be excluded. This approach ensures that only high-quality signals contribute to heart rate (HR) detection, thereby improving the overall accuracy and robustness of the measurements.
Data availability
No datasets were generated or analyzed during the current study.
Code availability
No custom code or software was developed for this study. All analyses were conducted using publicly available tools and standard computational methods.
References
Stockwell, S. J., Kwok, T. C., Morgan, S. P., Sharkey, D. & Hayes-Gill, B. R. Forehead monitoring of heart rate in neonatal intensive care. Front. Physiol. 14 https://www.frontiersin.org/journals/physiology/articles/10.3389/fphys.2023.1127419 (2023).
Ontiveros, R. C., Elgendi, M., Missale, G. & Menon, C. Evaluating rgb channels in remote photoplethysmography: a comparative study with contact-based ppg. Front. Physiol. 14 https://www.frontiersin.org/journals/physiology/articles/10.3389/fphys.2023.1296277 (2023).
Haugg, F., Elgendi, M. & Menon, C. Effectiveness of remote ppg construction methods: a preliminary analysis. Bioengineering 9, 485 (2022).
Kim, D. Y., Lee, K. & Sohn, C. B. Assessment of roi selection for facial video-based rPPG. Sens. (Basel) 21, 7923 (2021).
Elgendi, M., Martinelli, I. & Menon, C. Optimal signal quality index for remote photoplethysmogram sensing. npj Biosensing 1 (2024).
Lewandowska, M., Rumiński, J., Kocejko, T. & Nowak, J. Measuring pulse rate with a webcam—a non-contact method for evaluating cardiac activity. In: Ganzha, M., Maciaszek, L. A. & Paprzycki, M. (eds.) 2011 Federated Conference on Computer Science and Information Systems (FedCSIS) 405–410 (IEEE Computer Society Press, Los Alamitos, CA, 2011).
Poh, M.-Z., McDuff, D. J. & Picard, R. W. Non-contact, automated cardiac pulse measurements using video imaging and blind source separation. Opt. Express 18, 10762–10774 (2010).
de Haan, G. & Jeanne, V. Robust pulse rate from chrominance-based rPPG. IEEE Trans. Biomed. Eng. 60, 2878–2886 (2013).
Verkruysse, W., Svaasand, L. O. & Nelson, J. S. Remote plethysmographic imaging using ambient light. Opt. Express 16, 21434–21445 (2008).
Wang, W., den Brinker, A. C., Stuijk, S. & de Haan, G. Algorithmic principles of remote ppg. IEEE Trans. Biomed. Eng. 64, 1479–1491 (2017).
Ontiveros, R. C., Elgendi, M. & Menon, C. A machine learning-based approach for constructing remote photoplethysmogram signals from video cameras. Commun. Med. 4, 109 (2024).
Lu, H., Han, H. & Zhou, S. K. Dual-gan: Joint bvp and noise modeling for remote physiological measurement. In: Forsyth, D.,Gkioxari, G., Tuytelaars, T., Yang, R. & Yu, J. et al. (eds.) 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 12399–12408, https://doi.org/10.1109/CVPR46437.2021.01222 (IEEE Computer Society, Piscataway, NJ, 2021).
Shao, H. et al. Tranpulse: Remote photoplethysmography estimation with time-varying supervision to disentangle multiphysiologically interference. IEEE Trans. Instrum. Meas. 73, 1–11 (2024).
Zhao, P. et al. Toward motion robustness: a masked attention regularization framework in remote photoplethysmography. In: Farhadi, A. et al. (eds.) Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 7829–7838, https://doi.org/10.1109/CVPRW63382.2024.00779 (IEEE Computer Society, Piscataway, NJ, 2024).
Haugg, F., Elgendi, M. & Menon, C. Grgb rPPG: an efficient low-complexity remote photoplethysmography-based algorithm for heart rate estimation. Bioengineering 10, 243 (2023).
Frey, L., Menon, C. & Elgendi, M. Blood pressure measurement using only a smartphone. npj Digital Med 5, 86 (2022).
Ritsert, F., Elgendi, M., Galli, V. & Menon, C. Heart and breathing rate variations as biomarkers for anxiety detection. Bioengineering 9, 711 (2022).
Fine, J. et al. Sources of inaccuracy in photoplethysmography for continuous cardiovascular monitoring. Biosensors 11 https://www.mdpi.com/2079-6374/11/4/126 (2021).
Gao, H., Zhang, C., Pei, S. & Wu, X. Region of interest analysis using delaunay triangulation for facial video-based heart rate estimation. IEEE Trans. Instrum. Meas. 73, 1–12 (2024).
Qian, W. et al. Dual-path tokenlearner for remote photoplethysmography-based physiological measurement with facial videos. IEEE Trans. Comput. Soc. Syst. 11, 4465–4477 (2024).
Qiu, Z., Liu, J., Sun, H., Lin, L. & Chen, Y. W. Costhr: a heart rate estimating network with adaptive color space transformation. IEEE Trans. Instrum. Meas. 71, 1–10 (2022).
Lin, B. et al. Estimation of vital signs from facial videos via video magnification and deep learning. iScience 26, 107845 (2023).
Zheng, K. et al. Heart rate prediction from facial video with masks using eye location and corrected by convolutional neural networks. Biomed. Signal Process Control 75, 103609 (2022).
Zheng, K., Ci, K., Cui, J., Kong, J. & Zhou, J. Non-contact heart rate detection when face information is missing during online learning. Sensors (Basel) 20 https://doi.org/10.3390/s20247021 (2020).
Wei, B., He, X., Zhang, C. & Wu, X. Non-contact, synchronous dynamic measurement of respiratory rate and heart rate based on dual sensitive regions. Biomed. Eng. Online 16, 17 (2017).
Yang, Y. et al. Motion robust remote photoplethysmography in cielab color space. J. Biomed. Opt. 21, 117001 (2016).
Su, T. J. et al. Application of independent component analysis and nelder-mead particle swarm optimization algorithm in non-contact blood pressure estimation. Sensors (Basel) 24, https://doi.org/10.3390/s24113544 (2024).
Le, D. Q., Lie, W. N., Nhu, Q. N. Q. & Nguyen, T. T. A. Heart rate estimation based on facial image sequence. In: Huang, Y.-P. et al. (eds) 2020 5th International Conference on Green Technology and Sustainable Development (GTSD) 449–453, https://doi.org/10.1109/GTSD50082.2020.9303142 (2020).
Bhattacharjee, A. & Yusuf, M. S. U. A facial video based framework to estimate physiological parameters using remote photoplethysmography. In: Tiwari, S. (ed.) 2021 International Conference on Advances in Electrical, Computing, Communication and Sustainable Technologies (ICAECT) 1–7, https://doi.org/10.1109/ICAECT49130.2021.9392503 (IEEE, Piscataway, NJ, 2021).
Favilla, R., Zuccalá, V. C. & Coppini, G. Heart rate and heart rate variability from single-channel video and ica integration of multiple signals. IEEE J. Biomed. Health Inform. 23, 2398–2408 (2019).
Saikia, T., Birla, L., Gupta, A. K. & Gupta, P. Hreadai: Heart rate estimation from face mask videos by consolidating eulerian and lagrangian approaches. IEEE Trans. Instrum. Meas. 73, 1–11 (2024).
Sharma, N., Kumar, V., Shakya, K. & Sardana, V. A real-time framework to find optimal roi for contactless heart rate variability detection. In: Anita, E. A. M., Misra, P. & Kumar, S. (eds.) 2024 IEEE International Conference on Contemporary Computing and Communications (InC4) Vol. 1, 1–6, https://doi.org/10.1109/InC460750.2024.10649083 (IEEE, Piscataway, NJ, 2024).
Cruz, J. C. D., Pangan, M. J. P. G. & Wong, T. H. D. Non-contact heart rate and respiratory rate monitoring system using histogram of oriented gradients. In: Chang, C.-Y., Lin, K.-P. & Kuo, S. M. (eds.) 2021 5th International Conference on Electrical, Telecommunication and Computer Engineering (ELTICOM). Vol. 5, 122–127, https://doi.org/10.1109/ELTICOM53303.2021.9590137 (IEEE, Piscataway, NJ, 2021).
Chen, R., Chen, J., Cheng, L. & Huang, X. Heart rate detection with the off-the-shelf camera: Static to non-static. In: Li, B., Chen, L. & Li, T. (eds.) 2020 IEEE 5th International Conference on Signal and Image Processing (ICSIP) 704–708, https://doi.org/10.1109/ICSIP49896.2020.9339456 (IEEE, Piscataway, NJ, 2020).
Wei, W., Vatanparvar, K., Zhu, L., Kuang, J. & Gao, A. Remote photoplethysmography and heart rate estimation by dynamic region of interest tracking. In: James, C., Patton, J. & Summers, R. (eds.) 2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC) 3243–3248, https://doi.org/10.1109/EMBC48229.2022.9871722 (IEEE, Piscataway, NJ, 2022).
Tang, C., Lu, J. & Liu, J. Non-contact heart rate monitoring by combining convolutional neural network skin detection and remote photoplethysmography via a low-cost camera. In: Forsyth, D et al. (eds.) 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 1390–13906, https://doi.org/10.1109/CVPRW.2018.00178 (IEEE, Piscataway, NJ, 2018).
Bobbia, S., Benezeth, Y. & Dubois, J. Remote photoplethysmography based on implicit living skin tissue segmentation. In: Davis, L et al. (eds.) 2016 23rd International Conference on Pattern Recognition (ICPR) 361–365, https://doi.org/10.1109/ICPR.2016.7899660 (IEEE, Piscataway, NJ, 2016).
Kossack, B., Wisotzky, E., Hilsmann, A. & Eisert, P. Automatic region-based heart rate measurement using remote photoplethysmography. In Damen, D et al. (eds.) 2021 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), 2755–2759, https://doi.org/10.1109/ICCVW54120.2021.00309 (IEEE, Piscataway, NJ, 2021).
Wiede, C., Richter, J. & Hirtz, G. Signal fusion based on intensity and motion variations for remote heart rate determination. In: Giakos, G. C. (ed.) 2016 IEEE International Conference on Imaging Systems and Techniques (IST) 526–531, https://doi.org/10.1109/IST.2016.7738282 (IEEE, Piscataway, NJ, 2016).
Chou, Y. C., Ye, B. Y., Chen, H. R. & Lin, Y. H. A real-time and non-contact pulse rate measurement system on fitness equipment. IEEE Trans. Instrum. Meas. 71, 1–11 (2022).
Speth, J., Vance, N., Flynn, P., Bowyer, K. & Czajka, A. Remote pulse estimation in the presence of face masks. In: Gupta, M., Patel, V. M. & Souvenir, R. (eds.) 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2085–2094, https://doi.org/10.1109/CVPRW56347.2022.00226 (IEEE, Piscataway, NJ, 2022).
Li, B., Jiang, W., Peng, J. & Li, X. Deep learning-based remote-photoplethysmography measurement from short-time facial video. Physiol Meas 43 https://doi.org/10.1088/1361-6579/ac98f1 (2022).
Akamatsu, Y., Umematsu, T. & Imaoka, H. Calibrationphys: Self-supervised video-based heart and respiratory rate measurements by calibrating between multiple cameras. IEEE J. Biomed. Health Inform. 28, 1460–1471 (2024).
Qiao, D., Ayesha, A. H., Zulkernine, F., Jaffar, N. & Masroor, R. Revise: Remote vital signs measurement using smartphone camera. IEEE Access 10, 131656–131670 (2022).
Ahmadi, N. et al. Development and evaluation of a contactless heart rate measurement device based on rPPG. In: Heidari, H. & Blokhina, E. (eds.) 2022 29th IEEE International Conference on Electronics, Circuits and Systems (ICECS), 1–4, https://doi.org/10.1109/ICECS202256217.2022.9971006 (IEEE, Piscataway, NJ, 2022).
Liu, L., Xia, Z., Zhang, X., Feng, X. & Zhao, G. Illumination variation-resistant network for heart rate measurement by exploring rgb and msr spaces. IEEE Trans. Instrum. Meas. 73, 1–13 (2024).
Wu, B. F., Wu, Y. C. & Chou, Y. W. A compensation network with error mapping for robust remote photoplethysmography in noise-heavy conditions. IEEE Trans. Instrum. Meas. 71, 1–11 (2022).
Chen, X., Yang, G., Li, Y., Xie, Q. & Liu, X. Heart rate measurement based on spatiotemporal features of facial key points. Biomed. Signal Process. Control 96, L2033380572 2024-07-23 2024-07-25 (2024).
Song, R., Sun, X., Cheng, J., Yang, X. & Chen, X. Video-based heart rate measurement against uneven illuminations using multivariate singular spectrum analysis. IEEE Signal Process. Lett. 29, 2223–2227 (2022).
Feng, L., Po, L. M., Xu, X., Li, Y. & Ma, R. Motion-resistant remote imaging photoplethysmography based on the optical properties of skin. IEEE Trans. Circuits Syst. Video Technol. 25, 879–891 (2015).
Hebbar, S. & Sato, T. Motion robust remote photoplethysmography via frequency domain motion artifact reduction. In: Thewes, R. (ed.) 2021 IEEE Biomedical Circuits and Systems Conference (BioCAS), 1–4, https://doi.org/10.1109/BioCAS49922.2021.9644650 (IEEE, Piscataway, NJ, 2021).
Pimentel, M. A. F. et al. Toward a robust estimation of respiratory rate from pulse oximeters. IEEE Trans. Biomed. Eng. 64, 1914–1923 (2017).
Bobbia, S., Macwan, R., Benezeth, Y., Mansouri, A. & Dubois, J. Unsupervised skin tissue segmentation for remote photoplethysmography. Pattern Recognit. Lett. 124, 82–90 (2019).
Stricker, R., Müller, S. & Gross, H.-M. Non-contact video-based pulse rate measurement on a mobile service robot. In: Loureiro, R. et al. (eds.) 2014 RO–MAN: The 23rd IEEE International Symposium on Robot and Human Interactive Communication, 1056–1062, https://doi.org/10.1109/ROMAN.2014.6926392 (IEEE, Piscataway, NJ, 2014).
Heusch, G., Anjos, A. & Marcel, S. A reproducible study on remote heart rate measurement. Preprint at https://arxiv.org/abs/1709.00962 (2017).
Niu, X., Han, H., Shan, S. & Chen, X. Vipl-hr: A multi-modal database for pulse estimation from less-constrained face video. In: Jawahar, C. V et al. (eds.) Computer Vision – ACCV 2018, vol. 11365 of Lecture Notes in ComputerScience, 562–576, https://doi.org/10.1007/978-3-030-20873-8_36 (Springer, Cham, 2019).
Soleymani, M., Lichtenauer, J., Pun, T. & Pantic, M. A multimodal database for affect recognition and implicit tagging. IEEE Trans. Affect. Comput. 3, 42–55 (2012).
Kwon, S., Kim, J., Lee, D. & Park, K. Roi analysis for remote photoplethysmography on facial video. In: Patton, J. (ed.) 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC) 4938–4941, https://doi.org/10.1109/EMBC.2015.7319499 (IEEE,Piscataway, NJ, 2015).
Poh, M.-Z., McDuff, D. J. & Picard, R. W. Advancements in noncontact, multiparameter physiological measurements using a webcam. IEEE Trans. Biomed. Eng. 58, 7–11 (2011).
Zhao, C., Mei, P., Xu, S., Li, Y. & Feng, Y. Performance evaluation of visual object detection and tracking algorithms used in remote photoplethysmography. In: Sato, Y. & Yu, J. (eds.) 2019 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), 1646–1655, https://doi.org/10.1109/ICCVW.2019.00204 (IEEE, Piscataway, NJ, 2019).
Li, S., Elgendi, M. & Menon, C. Optimal facial regions for remote heart rate measurement during physical and cognitive activities. npj Cardiovascular Health 1 (2024).
Duan, Y., He, C. & Zhou, M. Anti-motion imaging photoplethysmography via self-adaptive multi-roi tracking and selection. Physiol Meas 44 https://doi.org/10.1088/1361-6579/ad071f (2023).
Jaiswal, K. B. & Meenpal, T. Continuous pulse rate monitoring from facial video using rPPG. In: 2020 11th International Conference on Computing, Communication and Networking Technologies (ICCCNT), 1–5, https://doi.org/10.1109/ICCCNT49239.2020.9225371 (IEEE, Piscataway, NJ, 2020).
Bobbia, S. et al. Real-time temporal superpixels for unsupervised remote photoplethysmography. In: Forsyth, D. et al. (eds.) 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 1454–1460, https://doi.org/10.1109/CVPRW.2018.00182 (IEEE, Piscataway, NJ, 2018).
Liu, S. Q. & Yuen, P. C. Robust remote photoplethysmography estimation with environmental noise disentanglement. IEEE Trans. Image Process 33, 27–41 (2024).
Zhang, Z. et al. Multimodal spontaneous emotion corpus for human behavior analysis. In: Proceedings of the IEEE Conferenceon Computer Vision and Pattern Recognition (CVPR) (2016).
Pilz, C. S., Zaunseder, S., Krajewski, J. & Blazek, V. Local group invariance for heart rate estimation from face videos in the wild. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) 1335–13358 (2018).
Nowara, E. M., Marks, T. K., Mansour, H. & Veeraraghavan, A. Near-infrared imaging photoplethysmography during driving. IEEE Trans. Intell. Transp. Syst. 1–12 (2020).
Magdalena Nowara, E., Marks, T. K., Mansour, H. & Veeraraghavan, A. Sparseppg: Towards driver monitoring using camera-based vital signs estimation in near-infrared. In: Forsyth, D. et al. (eds.) 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 1272–1281, https://doi.org/10.1109/CVPRW.2018.00174 (IEEE, Piscataway, NJ, 2018).
Hsu, G.-S., Ambikapathi, A. & Chen, M.-S. Deep learning with time-frequency representation for pulse estimation from facial videos. In: Beveridge, R. et al. (eds.) 2017 IEEE International Joint Conference on Biometrics (IJCB), 383–389, https://doi.org/10.1109/BTAS.2017.8272721 (IEEE,Piscataway, NJ, 2017).
Geng, J., Zhang, C., Gao, H., Lv, Y. & Wu, X. Motion resistant facial video based heart rate estimation method using head-mounted camera. In: Zheng, J. X. et al. (eds.) 2021 IEEE International Conferences on Internet ofThings (iThings) and IEEE Green Computing Communications (GreenCom) and IEEE Cyber, Physical Social Computing (CPSCom) and IEEE Smart Data (SmartData) and IEEE Congress on Cybermatics (Cybermatics), 229–237, https://doi.org/10.1109/iThings-GreenCom-CPSCom-SmartData-Cybermatics53846.2021.00047 (IEEE, Piscataway, NJ, 2021).
Zhao, C., Zhou, M., Han, W. & Feng, Y. Anti-motion remote measurement of heart rate based on region proposal generation and multi-scale roi fusion. IEEE Trans. Instrum. Meas. 71, 1–13 (2022).
Li, X. et al. The obf database: a large face video database for remote physiological signal measurement and atrial fibrillation detection. In: Bhanu, B. et al. (eds.) 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), 242–249, https://doi.org/10.1109/FG.2018.00043 (IEEE, Piscataway, NJ, 2018).
Meziatisabour, R., Benezeth, Y., Oliveira, P., Chappé, J. & Yang, F. UBFC-phys: a multimodal database for psychophysiological studies of social stress. IEEE Trans. Affect. Comput. 14, 622–636 (2021).
Goldberger, A. L. et al. Physiobank, physiotoolkit, and physionet. Circulation 101, e215–e220 (2000).
Zhang, Z. et al. Multimodal spontaneous emotion corpus for human behavior analysis. In: Agapito, L. et al. (eds.) Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 3438–3446, https://doi.org/10.1109/CVPR.2016.374 (IEEE, Piscataway, NJ, 2016).
Maki, Y., Monno, Y., Yoshizaki, K., Tanaka, M. & Okutomi, M. Inter-beat interval estimation from facial video based on reliability of BVP signals. In: Barbieri, R. (ed.) 2019 41st Annual International Conference of the IEEE Engineering in Medicine Biology Society (EMBC), 6525–6528, https://doi.org/10.1109/EMBC.2019.8857081 (IEEE, Piscataway, NJ, 2019).
Tang, J. et al. MMPD: Multi-domain mobile video physiology dataset. In: Mitsis, G. (ed.) 2023 45th Annual InternationalConference of the IEEE Engineering in Medicine & Biology Society (EMBC), 1–5, https://doi.org/10.1109/EMBC40787.2023.10340857 (IEEE, Piscataway, NJ, 2023).
Kopeliovich, M., Mironenko, Y. & Petrushan, M. V. Architectural tricks for deep learning in remote photoplethysmography. In: Sato, Y. & Yu, J. (eds.) 2019 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW), 1688–1696, https://doi.org/10.1109/ICCVW.2019.00209 (IEEE, Piscataway, NJ, 2019).
Acknowledgements
Open access funding was provided by the Swiss Federal Institute of Technology Zurich.M.B. acknowledges the DAAD (German Academic Exchange Service) and the Technical University of Munich for support through the PROMOS scholarship for international mobility, funded by the Federal Ministry of Education and Research (BMBF).M.E. acknowledges funding from Khalifa University (grant number FSU-2025-001).
Funding
Open access funding provided by Swiss Federal Institute of Technology Zurich.
Author information
Authors and Affiliations
Contributions
M.E. designed and led the study. M.B., M.E., and C.M. conceived the study. The literature search was carried out by M.B. Both reviewers, M.B. and M.E. collaborated in constructing the protocol and developing the search terms. M.B. conducted the initial literature search. M.E. directly supervised the work of M.B. All authors have read and agreed to the published version of themanuscript.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Bondarenko, M., Menon, C. & Elgendi, M. The role of face regions in remote photoplethysmography for contactless heart rate monitoring. npj Digit. Med. 8, 479 (2025). https://doi.org/10.1038/s41746-025-01814-9
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41746-025-01814-9
This article is cited by
-
Demographic bias in public remote photoplethysmography datasets
npj Digital Medicine (2025)