Expand this Topic clickable element to expand a topic
Skip to content
Optica Publishing Group

Capturing the pulse: a state-of-the-art review on camera-based jugular vein assessment

Open Access Open Access

Abstract

Heart failure is associated with a rehospitalisation rate of up to 50% within six months. Elevated central venous pressure may serve as an early warning sign. While invasive procedures are used to measure central venous pressure for guiding treatment in hospital, this becomes impractical upon discharge. A non-invasive estimation technique exists, where the clinician visually inspects the pulsation of the jugular veins in the neck, but it is less reliable due to human limitations. Video and signal processing technologies may offer a high-fidelity alternative. This state-of-the-art review analyses existing literature on camera-based methods for jugular vein assessment. We summarize key design considerations and suggest avenues for future research. Our review highlights the neck as a rich imaging target beyond the jugular veins, capturing comprehensive cardiac signals, and outlines factors affecting signal quality and measurement accuracy. Addressing an often quoted limitation in the field, we also propose minimum reporting standards for future studies.

Published by Optica Publishing Group under the terms of the Creative Commons Attribution 4.0 License. Further distribution of this work must maintain attribution to the author(s) and the published article's title, journal citation, and DOI.

1. Introduction

Heart failure (HF) is a significant health problem worldwide. In high income countries, HF is the most common diagnosis for hospital admissions for those over 65 [1]. The 10-year survival probability for those with the diagnosis is less than 30% [2], and the reported physical quality of life for those individuals is poor [3]. Beyond the individual effects, the economic burden for the condition is high, with over $\$$108 billion spent globally, and the current upward trajectory of global disease burden is expected to continue with the aging population [46]. The dominant burden is due to hospitalisations, costing both time for the patient, and money for the health system [4]. As a chronic, progressive disease, the accumulated burdens on the body over time result in frequent rehospitalisation for those diagnosed with HF, with 25% likely to be re-admitted after 30 days, and 50% likely to be re-admitted within 6 months, with a large percentage of these being avoidable [7,8]. Elevated central venous pressure (CVP) has been associated with rehospitalised patients both when presenting at the hospital [9], as well as the increase in time-integrated pressure out of hospital [10], suggesting this may be used as a marker for early intervention, reducing rehospitalisations [11]. Development of an at-home, non-invasive measurement of this pressure may be beneficial in this domain, as some individuals have shown interest in the use of mobile health applications to manage their own symptoms [12].

Technological advancements in the remote monitoring of patients includes wearable technologies such as flexible sensors that can be used to detect cardiorespiratory signal [13,14]; implanted devices such as the continuous glucose monitors [15]; and the use of cameras and computers in telehealth consults. The use of cameras in physiological monitoring is a popular area of research, and companies are emerging using artificial intelligence (AI) and smartphone technologies to provide consumers with information to help guide self-management of their conditions [16,17]. One such measure, concerning HF, is the potential use of cameras in the replication of a common bedside technique in the inspection of the jugular venous pulsations (JVP). The jugular veins act as a manometer to the right heart, and an increase in the height of a visible pulsation on the neck can indicate an increase in CVP [18]. CVP measurements are pivotal in assessing volume status and right heart function in HF patients, guiding diuretic and vasodilator therapy. Persistently elevated CVP may indicate the need for advanced interventions, such as ventricular assist devices or heart transplantation. Regular monitoring of this pressures aids in optimizing treatment efficacy, prognostication, and ensuring timely therapeutic adjustments. In hospitals, clinicians rely on CVP measurements (acquired either through the invasive gold standard method using a central catheter or via bedside examination) to tailor patient therapy. Bedside examinations are difficult to perform, even for highly trained clinicians, with the subtle pulsations often being difficult to visualise, especially in patients with higher body mass index (BMI) [19]. After discharge, utilizing the invasive method becomes impractical. Instead, patients primarily rely on self-reported weight measurements to determine fluid overload levels. To enhance remote monitoring capabilities and promote patient independence, recent innovations have turned to video assessments. Two studies have demonstrated that the accuracy of clinicians’ CVP estimations via recorded videos are comparable with traditional bedside examinations [20,21]. Despite this, the role of video assessment for the purpose of CVP estimation is not yet standard practice in either remote or clinical settings.

The field of physiological assessment by video is broad, and in the process of establishment. From recent large-scale surveys and systematic reviews [2229], a consistently mentioned challenge was the lack of standardization in studies and reports. To address this, this review outlines factors likely to affect the accuracy of estimations, from both system and study design, suggesting reporting guidelines for future investigations. Furthermore, much of the literature is centred around the face as a region of interest (ROI), for the purposes of measuring heart rate. This review highlights the benefits of imaging the neck, which provides a richer source of information as it contains information relating to both the right and left sides of the heart through the carotid arteries and jugular veins. Additionally, the neck, as compared with the face, contains fewer personally identifiable features, reducing the difficulties in the compilation of large publicly available datasets.

In this review, we begin with a brief introduction to the mathematical model for cameras used to motivate the variety of different approaches taken. We then provide systematic review of the approaches taken in the literature that use video and signal processing in the assessment of the jugular veins and the associated results. The discussion consists of a meta-review, identifying the diversity of approaches in camera-based physiology assessments, and the general trend of these effects, providing some guidance in hardware selection and considers the study design for experiments such as this, and factors that need to be considered. We then summarise the literature in terms of general design implementation and provide future direction regarding reporting of these results for future reviews and meta-analyses.

2. Modelling the camera signal

The typical model for the detected signal is based on the skin reflection model and is found in numerous publications [26,30,31]. Variations in surface angles and underlying tissue over the physiological signals of interest give rise to different methods of signal acquisition. Regions where motion is expected, such as pulsatile motion in the neck or wrist due to the pumping of the heart or the rise and fall of the chest and shoulders due to breathing, may lead some investigators to design imaging systems that emphasise either the specular or diffuse components of reflection, depending on the signal of interest. With this in mind:

The signal as measured by each pixel at time $t$ can be expressed as:

$$I_{\text{pixel}}\left(t\right)=\int_{-\infty}^{\infty}{Q_{\text{efficiency}}\left(\lambda\right)I_{\text{sensor}}\left(t,\lambda\right)d\lambda}+N_{\text{sensor}}\left(t\right)$$
Where $I_p$ is the pixel value, $Q_{\text {efficiency}}\left (\lambda \right )$ is the quantum efficiency of the pixel with wavelength $\lambda$, $I_{sensor}$ is the irradiance (power received by the pixel sensor) and $N_{\text {sensor}}$ is the noise of the sensor. If the signal is being received from a surface at distance $d$ from the pixel, assuming no losses of intensity due to the air, the relationship between the irradiance of the surface, and that detected by the pixel is:
$$I_{\text{sensor}}(t,\lambda)=\frac{A}{4{\pi d}^2}\int_{\theta_1}^{\theta_1}{I_{\text{surface}}(t,\lambda,\theta)d\theta}$$
Where A is the size of the sensor, and $I_{\text {surface}}$ is the emitted irradiance of the surface that corresponds with the light received over the region $A$, as determined by the angle $\theta$. Based on Shafer’s dichromatic reflectance model [32], we can decompose the emitted light from the surface in terms of the types of reflected light, either specular $r_{\text {specular}}$ or diffuse $r_{\text {diffuse}}$.
$$I_{\text{surface}}(t,\lambda,\theta)=I_{\text{incident}}(t,\lambda,\theta)(r_{\text{specular}}(t,\lambda,\theta)+r_{\text{diffuse}}(t,\lambda,\theta))$$
Specular reflection is like that of a mirror, where the angle of incidence is equal to the angle of the reflected light, when measured from the normal of the surface. As such, a rough surface will reflect incident light in a variety of directions compared with the macro-surface orientation. On the aggregate, this can be considered diffuse reflection, but in this instance, we will consider this the diffuse component of specular reflection. The overall amount of light reflected in a specular fashion is relatively low, between 2 - 4% [33]. The diffuse component of this is even smaller, between 2.67 - 9.91% of total reflected light depending on factors such as skin roughness and the angle of incidence of the incoming light [34]. Although the total proportion of specular reflection is small, relative to that of diffuse reflection, the solid angle over which it is spread is small, and so may contribute a larger part of the signal as measured by the camera.

The diffuse reflective component at a point, $r_{\text {diffuse}}(t,\lambda,\theta )$, is due to the initial transmission of light beneath the surface. After transmission below the tissue surface, the light interacts with matter either by absorption or scattering. Some proportion of light is then scattered back out from the surface. The angle of propagation from the normal of the surface is random, and so is more accurately described as diffuse. The likelihood of a photon transmitting through the skin surface and being reflected back is dependent on the incident wavelength of the photon. Longer wavelengths generally correspond with increased likelihood of transmission.

For jugular venous pulsations, the measured signal has a large time-varying component due to changes in both specular and diffuse reflections from the underlying blood vessels. The pulsating vein gives rise to a time dependence of the angle of the surface, and the varying amount of venous blood with the cardiac cycle results in a time varying diffuse reflection component. Design considerations are expanded on in the discussion section and include the use of specific light wavelengths known to be less penetrating to maximise specular reflection, or polarizing filters to maximise diffuse reflection.

3. Methods

This systematic review adheres to the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines [35]. The PRISMA flowchart, which outlines the progression of studies through the review, can be found in Fig. 1. Below, we detail the study selection process to facilitate the reproduction of results in future reviews.

 figure: Fig. 1.

Fig. 1. PRISMA Systematic Review Flowchart

Download Full Size | PDF

3.1 Eligibility criteria

As is typical in PRISMA systematic reviews, we have used the PICO Framework to structure the eligibility of studies included in the review [36]. All Participants were included. The Interventions were restricted to the use of cameras for the assessment of the jugular veins. The review did not limit studies based on Comparison devices or Outcome measures. Studies were from database inception through to October 2022, limited to English, including all conference papers and journal articles. All included manuscripts are published and peer reviewed. As most published literature for this topic consists of proof-of-concept works, and the literature is relatively sparse (n = 7), all studies were included.

3.2 Information sources

Data was acquired through searches of the OVID, Web of Science, IEEE and Pubmed databases, from inception up to October 2022. References from prior systematic reviews of camera based measurement of physiology and vital signs by McDuff et al. [23], Pham et al. [24], Selvaraju et al. [25] and Zaunseder et al. [26] were included.

3.3 Search strategy

All databases were searched with ("Camera" OR "iPPG" OR "PPGi" OR "Photoplethysmography") AND ("Carotid" OR "Jugular" OR "CVP" OR "JVP" OR "JVD" OR "Neck") with the search augmented for each database accordingly.

Searches included the carotid artery as any assessment of the vasculature in the neck may have included jugular venous assessments.

3.4 Selection process

The collected items were collated and added to Endnote for de-duplication. From the identification of the first investigation in the use of cameras for Photoplethysmography (PPG), also known as remote Photoplethysmography (rPPG) by Wu, Blazek and Schmitt [37], in 2000, a first date cut-off was used. Non-English items were then removed. Titles and abstracts were screened chronologically, to determine the first published author using cameras for the neck pulsation assessment [38], and a date cut-off for all other articles was then applied, removing all items prior to 2015. Articles were then removed if they did not contain the word "camera" in the abstract, and then again if they did not contain "jugular", "carotid" or "neck". Titles were then screened for eligibility. Abstracts were then manually screened, and accepted articles retrieved for synthesis.

3.5 Data items and collection process

For each article, the primary interest of the review is around both study and system design. Study design includes participant demographics, setting and setup information in the data acquisition process, and the reference standard used. System design was separated into hardware, and software, with software further divided into image and signal processing methods.

4. Experimental setup and aims

Each study has been analysed in two parts – study design and system design (Fig. 2). Study design includes the demographic, location, setup, and standard reference information used and overall accuracy of the method. System design is primarily concerned with the hardware and software components used as a means of assessing the state-of-the-art system. We fabricate an illustrative in Table 5 with a figure showing a potential setup in Fig. 3. The path from the acquisition of the data through the software is not always consistent and may bypass signal processing methods altogether. We first give an overall description of the relevant demographic and setup of the studies. As the sample size of studies included in the review is relatively small (n = 7), with each employing unique setups, hardware and processing algorithms, we do not dwell on specific experimental design choices made by the investigators. In sections 6.3 and 6.5, we discuss in detail the potential pros and cons of specific design decisions to assist investigators in future studies.

 figure: Fig. 2.

Fig. 2. Overview of Study Design. The design of a study consists of both the system itself, and the environment where it is applied.

Download Full Size | PDF

 figure: Fig. 3.

Fig. 3. An illustration of the setup described Table 5. The setting of the study was in the emergency department of the hospital in Argentina. The participant is in a seated position, facing forward. The camera on the phone is situated at a distance of 30 cm from the neck, at an angle of 45°from the front plane of the person. The flash of the light from the phone is used to illuminate the neck. Adapted from vecteezy.com

Download Full Size | PDF

4.1 Study design

4.1.1 Study aims

The jugular veins provide two significant sources of clinically relevant information. The highest point on the neck of visible pulsation, relative to the sternal angle, correlates well with the CVP, and the waveform is associated with a range of cardiovascular diseases [39]. As such, investigators have tended to concentrate their efforts on augmenting videos as an assistive device for the clinicians [20,38], or visualising the waveform itself [31,4042]. Moço et al. investigated the importance of ROI location and patient positioning [41]. Amelard et al. investigated the relationship between attenuation of infrared light in the jugular venous region, rather than concentrating on pulse wave extraction, although this was still included in the video processing system [43].

4.1.2 Demographics

A summary of demographic data is shown in Table 1. All investigations consist of small pilot studies, ranging from 4 to 48 participants. The age range is from 9 – 75+ years, though only one study had a reported mean participant age over 30 [20]. In the context of HF, where JVP assessment is most common, patients are frequently in the older demographic. As such, future investigations may do well to include older participants to ensure accuracy on the cohort that is most likely to benefit from the technology. No study reported the skin tone or ethnicity of the participants, though one study described it as "varied". It’s important to highlight that skin tone is of particular importance when it comes to ensuring an equitable system. This is discussed further in section 6.3. Amelard et al. noted that the signal recorded for participants who are older, and with higher BMIs have a lower signal-to-noise ratio for the measured waveform [31]. This suggests that BMI (or other body morphology metrics) are of importance, and should be noted during experiments.

Tables Icon

Table 1. Summary of Demographic Information for the included studies.

4.1.3 Setting and setup

Setting and setup information is summarised in Table 2. Herein, we describe the range of experimental design decisions. The potential implications of these particular design considerations is elaborated on in sections 6.3 and 6.5.

Tables Icon

Table 2. Summary of Experimental Setup Information for the included studies.a

All studies were conducted in either laboratory or clinical settings. Cameras were positioned at a variety of distances, from 30 cm up to 1.5 m. Not all studies reported the distance of the camera to the neck. All investigations used artificial lighting, though the specific spectra were not always clearly defined.

Most studies investigated participants in the supine position, mimicking the approach taken by clinicians by the bedside. One study investigated the effect of participant position and ROI selection on the neck, and so a small number of participants were assessed in various positions [41]. Another study assessed the change in the attenuation of the signal with experimentally induced pressure changes. These include tilting the body below the horizontal, increasing venous blood-flow to the head and torso, forced exhalation via the Valsalva manoeuvre to increase intra-thoracic pressure, and the application of lower body negative pressure, mimicking hypovolemia [43].

5. Results

5.0.1 Comparison and results

Results from the literature can be characterised as either quantitative or qualitative. Quantitative results provide insight into the timing and signal strength of the signals measured by the imaging system. Qualitative results provide hints towards improvement in setup and future processing algorithms. The results have been summarised in Table 3.

Tables Icon

Table 3. Summary of the quantitative and qualitative for the included studies.a

Quantitative results

Amelard et al. [31] visualised the dual pulsating regions in the neck and showed two regions with high Pearson correlation with a synchronously acquired finger pulse. The arterial (0.85 $\pm$ 0.8), and venous (-0.73 $\pm$ 0.17) waveforms respectively. Lam Po Tang et al. [40] utilised a displacement measurement algorithm to track skin motion and yielded comparable Pearson correlation between estimated motion and a laser distance meter (0.93 $\pm$ 0.05). Furthermore, temporal spacing between electrocardiogram (ECG) and CVP waveform characteristics were first established. Saiko et al. [42] used a novel imaging technique based on increasing specular reflection to show a 6-fold increase in SNR for the waveform measurements as compared with Amelard et al.’s initial PPG based approach [31]. SNR was calculated by the mean difference between the peak and the trough of the waveforms generated.

Abnousi et al. [20] demonstrated the feasibility of using both raw and augmented videos to estimate the CVP of participants by neck examination. Augmenting the video by enhancing motion within a specific frequency band decreased mean measurement error from -2.9 cmH2O at the bedside to -0.8 cmH2O with augmented videos. Amelard et al. [43] demonstrated high mean Pearson correlation (> 0.85) with DC variations of the PPG signal under experimentally induced pressure changes.

Qualitative results

Dang et al. [38] showed that it is possible to identify pulsating regions in the neck, as confirmed by clinicians. Amelard et al. [31] revealed the waveform of the jugular veins by remote photoplethysmography imaging. They also showed that the arterial and jugular pulsations in the neck are highly correlated with the arterial finger pulse, and that this can be used to distinguish between pulsating regions on the neck, confirmed by ultrasound. Positive correlation is associated with the carotid artery, and negative correlation with the jugular vein. Lam-Po-Tang et al. [40] used a technique previously employed to measure carotid artery waveforms, and showed that the CVP waveform can be acquired from measurement of skin displacement. Moço et al. [41] demonstrated the importance of patient positioning and region of interest selection for neck based assessments such that the same location on the neck can be dominated by either the carotid artery pulse or the jugular venous pulse depending on if the participant is seated, recumbent or supine.

5.2 System design

The ideal system will vary depending on the intended use-case. An ideal hardware system is dependent on the the system setup and the processing algorithms. Similarly, the existence of a benchmarking dataset for the purpose of defining a best set of processing algorithms is also yet to be compiled. As such, we cannot comment on whether the studies included in this analysis are better or worse than each other. However, we can draw on the lessons learned from face-specific rPPG approaches on the expected benefits for each of the setups. More detail regarding the lessons drawn from facial rPPG approaches is outlined in section 6.5, but we briefly relay some of the potential decisions behind design choices.

Hardware

As shown in Fig. 2, the hardware used can be divided into the camera, artificial illumination sources, and any other optical devices present. The hardware systems used are summarised in Table 4.

Tables Icon

Table 4. Summary of Hardware Information for the included studies.a

Camera

A range of cameras have been used, including red, green and blue (RGB), infrared (IR) and monochromatic cameras, although not all colour channels were always used. It’s likely that the IR camera used by Amelard et al. [31,43] will extract signals from deeper within the neck, and so may increase the relative signal strength when compared with other colour channels. However, it’s noted that RGB cameras used by other investigators all managed to yield pulsatile signals. The frame rates for the videos were between 20 and 250 FPS. From Whittaker-Nyquist-Shannon sampling theorem [4446], a frame rate of a minimum of 8 FPS is required to accurately reconstruct a cardiac cycle. Higher frame rates will likely yield better results, and so higher frame rates are preferred where possible. Resolutions ranged from 500 x 500 to 2048 x 2048. Given a fixed distance from the participant, an increase in resolution will result in a larger number of pixels within the ROI, likely increasing the overall SNR. A near infrared (NIR) wavelength bandpass filter was used by Amelard et al. [31].

Illumination

All studies used artificial illumination sources. This included light emitting diode (LED) arrays in the blue and NIR ranges, fluorescent and halogen lamps, and otherwise unspecified lighting. Lighting is of particular importance when considering the skin reflection model illustrated in section 2. Increasing the incidence of photons on the neck will yield a larger measured signal by the camera. Blue LEDs are used to emphasise the surface of the skin (specular reflection), and longer wavelengths may emphasise the pulsatile blood via diffuse reflection. The best lighting for each situation will depend on the intended method of pulse extraction, as is evident by the contrast in approaches taken by Saiko et al. [42], and Amelard et al.[31,43], who primarily focus on the extraction of specular and diffuse reflections respectively.

Other hardware

Polarizing filters were used by Amelard et al.[43] to enhance PPG signals at depth. Diffusers were used for illumination sources in some instances [31,42].

5.1.2 Software

Image processing techniques consist primarily with spatial processing algorithms for selection and refinement of ROIs. Spatial processing algorithms involve denoising and image segmentation. Signal processing is typically concerned with temporal signal processing, including time averaging, denoising and the use of frequency filters. Signal processing techniques can be used to refine ROIs beyond spatial techniques.

Image processing

Pre-processing: The averaging of a region of pixels (spatial averaging) reduces the effect of sensor noise. Spatial averaging is performed in nearly all approaches, typically with a Gaussian filter, but mean filters are also used [31,38,4143]. One approach uses large overlapping regions, with the centre of the region being used as a "control point", for which signal processing if later performed [40]. Dang used a motion tracking algorithm to reject videos if undesired motion was detected above a threshold [38]. A novel calibration tool was used in conjunction with an IR camera to emphasise diffuse reflectance and the PPG signal, with specular reflective components minimised [43].

ROI Selection and Refinement: Manual segmentation of a region of interest was performed in 4/7 investigations [38,4042]. Amelard [31] established the use of temporal correlation of pixel signals with that arterial pulsations in the finger and the technique was used in two subsequent investigations [31,47]. The phase lag between the arterial and venous pulsations identified by Amelard was used to justify a growing spatial clustering algorithm [48], with the venous waveform lagging by approximately 400 ms after the carotid wave [42,43]. Based on frequency domain operations with signal processing, ROIs were refined by the selection of pixels with high power within specific frequency bands [40].

Signal processing

De-trending: A detrending method [49] was used to remove variations in environmental illumination [31]. One-second means are removed from the pulsatile signal in another approach [42].

De-noising: Band-pass and low-pass filters have been used to remove noise, and emphasise the signals within a specific frequency range including ideal bandpass filters and Butterworth filters [20,40,47]. Kalman filters have also been used, based on Newtonian smoothness priors [43,50]. Moving average filters have also been used [42].

Other processing: A phased-based Savitzky-Golay gradient correlation algorithm was used for sub-pixel displacement estimation [51]. Eulerian amplification was used on the whole video, with the use of Laplacian pyramids and spatio-temporal bandpass filtering used to emphasise motions within specific frequency ranges [20,52,53].

6. Discussion and future direction

All approaches indicate a variety of methods will yield pulsatile signals from the neck. This suggests a camera-based technology is feasible as an assistive device in clinic, or for self-monitoring. Each investigation varied in both hardware and software. As such, a "state-of-the-art" declaration of the technologies so far is futile. This has been a consistent theme in reviews of camera-based technologies thus far [2426,54,55], though techniques based on neck assessments are more scarce than those based on faces.

6.1 Challenges

A multitude of challenges exist for rPPG of the neck, with significant overlap with facial approaches.

6.1.1 ROI selection

Unlike the face, the neck has not had the same attention when it comes to feature detection. As such, segmentation and ROI selection is not as well researched. Compounding this, the neck contains two cardiac pulsatile signals - one from the jugular vein, the other from the carotid artery. The methods used to separate the ROIs included here all rely on the use of external signals from either PPG or ECG to determine heart rate. No method has currently relied solely on the video processing and proved that it is possible to reliably distinguish between the two ROIs.

6.1.2 Motion artifacts

It is generally presumed that the time-varying pixel signal is a direct result of the time-varying light attenuation signal, thereby suggesting a linear relationship. However, this association becomes invalid when the pixel does not measure the exact same area over time, typically due to unrelated subject motion [23,26,56]. Motion artefacts may be due to from various physiological phenomena, such as swallowing, breathing, and head-turning. The magnitude of the effect will likely depend on the region being examined and the processing techniques employed. Swallowing detection may be a necessary problem to solve for this purpose.

6.1.3 Illumination

Lighting conditions may shift over time, leading to the introduction of errors due to shadows or non-homogeneous lighting [23,26,5759]. Additionally, the relative distance between the camera, participant and light source, significantly impacts the overall signal-to-noise ratio, where SNR decreases with increasing distance [60]. This aspect is particularly relevant when considering the application; for instance, a clinical setting allows for controlled lighting and optimal camera-to-subject distance, thus reducing noise and motion artifacts compared to a system designed for ambient monitoring. However, as the neck’s vessels are deeper than the forehead’s, there may be a requirement for stronger lighting or a closer camera-to-subject distance to capture a discernible pulsatile signal.

6.1.4 Lack of reporting standards

Generally, the development rPPG would benefit from a more systematic approach, including guidelines for experimental design, system design and reporting, as well as publicly available datasets for the benchmarking of processing algorithms. Some guidance regarding this can be drawn from prior reviews concerned with cardiac pulse extraction from other anatomic regions, but the neck consists of two pulsatile signals, the carotid artery and the jugular vein, and so general signal processing approaches may not be adequate for this compared with those derived for the face.

The IDEAL-D and GEP-HI frameworks provide some structure in the development diagnostic systems, including study design and the broader context considerations required for development of novel tools used in medical diagnosis and patient monitoring [6163]. Using the IDEAL-D framework, the 5/7 investigations thus far are in the Idea phase, Stage 1, with 2/7 investigations in Stage 2a, with iterative developments being used to assess the effects of changes such as patient position or experimental perturbations of pressure.

6.2 Design goals

The IDEAL-D framework suggests the development of diagnostic tools are tailored to a specific purpose. As such, hardware and algorithm choices can be chosen to optimise specific outcomes. For example, remote monitoring and telehealth applications will need to develop technology on typical hardware devices such as webcams or smartphone cameras, limiting the utility of complex optical devices (eg. polarizing filters and specialised lighting sources). Software built for this purpose will also likely need to be tolerant to motion and a variety of illumination environments. Conversely, if the system is being developed as a clinical aid, there may not be as much motion or variation in lighting conditions.

6.3 Reporting

Regardless of the design goal, standard reporting information should be included. The STARE-HI guidelines provide a good starting point for reporting in journal articles [61,62] and conference papers [64]. We want to develop systems that are equitable and accurate. Equity will include evaluation across all demographics, with accuracy reported. To do so, we need to consider metrics that could impair the accuracy of the system and perform specific evaluation studies on those populations. The reporting of evaluation studies should include detailed demographic and system information for future replication studies to be carried out for populations beyond those previously assessed. Reported demographic information should include any characteristic likely to be seen as either inequitable or could theoretically affect accuracy of the system. For example, sex, age, skin tone, pathologies, body morphology, pharmacology etc. System information should include at least minimum detail for accurate replication, including the make and model of cameras and illumination equipment, participant setup information and ideally the open-source algorithms used, or at least a detailed description of it. Finally, an oft overlooked component to reporting is the reference standard with which the developed system is compared. Reference standards have their own limitations, including racial bias [65,66] and large errors in measurement if the procedure is not performed correctly [66,67]. The gold standard reference for CVP is the right heart catheter. However, the technique for insertion and zeroing can vary, inducing errors [6870]. For all references used, investigators should, as best as possible, clearly define how the reference device is setup, and quantify uncertainties in the measurement process. In Table 5 we provide our list of recommendations, as well as an fabricated example for illustrative purposes, in a similar fashion to the STARE-HI elaboration publication [62]. Fig. 3 is an illustration of the example provided in Table 5.

Tables Icon

Table 5. Recommendations for minimum reporting guidelines. The example provided is fabricated, for the illustration of how this may be reported. The use of [citation] and [table] indicate that a citations and tables should be included, providing a link to the model information if possible.

6.4 Public datasets

A commonly referenced shortcoming in the previous reviews is the limited publication of public datasets, particularly for people with pathologies [2226,55]. The primary region of interest in most literature is the face. An unfortunate shortcoming with facial videos is that they by necessity contain identifying features. The neck provides a region of interest with strong pulsatile signals, information regarding both the right and left side of the heart by means of the jugular veins and the carotid arteries, without the limitation of easily identifiable features. This provides strong motivation for the future of large-scale datasets containing the neck as a region of interest, instead of the face.

6.5 Benchmarking and design choices

Benchmarking consists of the establishment of a baseline with which future investigations can be compared. A benchmark system can be established in two forms: hardware and processing. Hardware benchmarks establish minimum system requirements for the cameras and lighting used and can be compared with via benchmark algorithms. Benchmark algorithms establish themselves on multiple datasets to ensure accuracy in a variety of applications. The number of investigations into the neck as a whole limit the ability of the authors to establish benchmarks for systems or algorithms, however, guidance can be provided from parallel research into other camera-based investigations into vital sign estimation. The results are summarised in Table 6.

Tables Icon

Table 6. Hardware guidance based on general remote photoplethysmography analyses.a

6.5.1 Hardware

Camera

Camera Resolution and Camera-Subject Distance: Visual realisation of pulsatile signals in the neck are inherently difficult for individuals, as evidenced by the frequent inability of clinicians with no vision defects, to realise the pulsatile signal at short ranges. It is therefore likely that this would benefit from superhuman fidelity via the use of cameras. Signal extraction from videos relies on the number of pixels expressing that signal. As blood pulse signals are relatively noisy within a single pixel, increasing the number of pixels for signal extraction can significantly reduce the amount of noise. Investigations by Song et al. into the effect of camera resolution showed that for any distance greater than 0.5 m, increasing the camera resolution reduces the error in heart rate measurement across a variety of signal extraction techniques [72]. For distances less than 1.5 m, resolution of at least 720p can result in heart rate accuracy within 4 BPM (beats per minute).

Frame rate: Cardiac pulsatility can be assumed to occur for most people between 30 – 240 BPM (0.5 - 4 Hz). The Whittaker-Nyquist-Shannon sampling theorem [4446], suggests a minimum sampling frequency of twice the maximum frequency is required to reconstruct the original signal and avoid aliasing, so cameras with > 8 FPS should be able to accurately reconstruct the cardiac cycle. Indeed, Blackford and Estepp found no degradation in signal processing when down-sampling from 120 FPS to 30 FPS [73]. This finding was challenged by Speth et al. using a convolutional neural network, where the mean absolute error increased from <4.18 BPM to >6 BPM for frame rates less than 11.3 FPS [74].

Colour Channel and Colour Filter Array: Monte-Carlo modelling of light for reflectance PPG devices such as smart watches show that longer wavelengths have increased depth penetration as compared with shorter wavelengths [75]. As such, the signals from IR cameras like those used by Amelard et al. [31,43] are a result of light interactions with greater tissue volumes, and at greater tissue depth than blue or green light. The arrangement of the bandpass filters over the charge-coupled device (CCD) used to get RGB images is termed the colour filter array. Typically a Bayer array is used, where the ratio of G:R:B filters is 5:2:2, mimicking the human eye and its increased sensitivity to green light [76]. From use of standard RGB sensors, it has been shown that green light has the highest signal to noise ratio [58]. McDuff et al. investigated the use of 5 colour channel cameras and their impact on pulse measurement and breathing rate [77]. The 5 channels were red, green, blue, cyan, and orange (RGBCO). They found that by using CGO increased the measurement accuracy of heart rate variability and that no benefit was obtained from the red or blue channels, and that the addition of the blue channel degraded performance. The addition of an orange colour channel was noted to be particularly informative. This could be due to the combination of increased tissue depth penetration of green and orange light compared with blue, and the peak absorption of the orange spectra at 580 nm, corresponding closely with one of the peaks of the absorption spectra of the two dominant chromatophores in blood, oxyhemoglobin and deoxyhemoglobin [78]. To the best of our knowledge there have been no investigations into the effects of colour filter array patterns. However, this information flows downstream in terms of variation in illumination and spread of measured values, and so assessments of this may not be necessary.

Bit Depth: Bit depth refers to the detail in the recorded intensity. The parallel could be made to significant figures, or resolution of a measurement device. Cameras typically capture values for each channel in 8-bit depth, giving values between 0-255. Bit depths of 10, 12 and 14 have been used in previous investigations [37,7982]. Due to the variation in processing and site selection, the effect of increasing bit depth has not yet been defined. However, as identified by McDuff et al. [77], under non-ideal conditions, most cameras successfully can ascertain heart rate, however, peak detection is limited, and so parameters such as heart rate variability are more difficult to measure accurately. By increasing bit depth, this may improve the temporal accuracy of peak detection, and may warrant investigation, especially with the association between heart-rate variability and blood pressure [83].

Image Compression: Cameras typically compress the captured videos for faster file transfer and reduced storage space. Several investigations have been conducted into the various video compression techniques and video encodings [8486]. All investigations showed that with increasing video compression, accuracy of heart rate extraction decreased. A variety of compression algorithms exist, and the magnitude of the effect will likely be different for each of them. For instance, as the PPG signal is primarily due to chromaticity variations, chroma-subsampling as performed in the YUV colourspace is likely to reduce the SNR of the measured signal for each pixel. Similarly, micro-motion of the skin contributes to the signal via specular reflection, and algorithms that may stabilise or remove this motion will further deteriorate the signal [85].

Illumination

The measured signal is highly dependent on the illumination of the surface. As such, illumination is likely to play a significant role in the accuracy of the measured signal. Illumination in studies thus far is either defined as indoor or artificial light such as those in a lab, hospital, or office, natural/ambient as provided by the sun, or specifically designed to elucidate specific features that may be useful for specific wavelengths, such as blue light used by Moço et al. [41] when attempting to enhance the surface motion [26].

Other optical devices

Filters: If an artificial lighting source is used to exaggerate specific signal properties, filters can act to further highlight those properties. By using bandpass filters that correspond with the spectrum of the light, we can then probe specific depths within the tissue. For instance, if we use an infrared filter, the reflected signal will be a result of interactions at deeper tissue depths compared with that of a blue filter. This is particularly useful when attempting to visualise noisy signals at depth, such as those of venous pulsations, and was employed by Amelard et al. [31,43]. Monochromatic cameras often allow higher frame rates, at the expense of loss of colour information, and so a monochromatic camera may have higher frame rate and resolution than an RGB camera at the same price point. By using filters with monochromatic cameras, one can selectively remove wavelengths which may be less useful (eg. red signals may have higher noise due to motion artifacts [87]), while retaining the benefit of the increased frame rates and reduced camera cost.

Polarizers: Polarizing filters on both the light source and sensor reduce specular reflection component in the signal. When light is oriented orthogonal to a surface with a linear polarizing filter it results in negligible specular reflection. By combining this with a circular polarizing filter on the image sensor, one can ensure that the signal measured is due to blood volume changes (i.e., PPG) effects below the surface. Thus, using polarizing filters, we can increase specific signals either due to PPG effects, or surface effects [26,88].

6.5.2 Software

Signal processing approaches

Traditional signal processing approaches have been investigated significantly in parallel literature and other reviews [26,55]. This includes ROI selection by automated segmentation, ROI tracking algorithms for motion tolerance, spatial filtering to reduce quantization noise, detrending and bandpass filters to emphasise specific desired frequencies, wavelet transforms and more. The algorithms proposed have been assessed on a range of publicly available datasets, which is unable to be performed in this instance, and so a "best approach" recommendation is not possible. A publicly available dataset consisting of neck based videos and corresponding heart rate wave forms would be a beneficial to this end.

Artificial intelligence and machine learning

Artificial intelligence (AI), machine learning and deep learning algorithms are emerging as being more accurate than traditional signal processing approaches for heart rate detection among faces [28]. Furthermore, AI can be used to automatically segment and define the ROI for traditional approaches [89], or traditional approaches can supply a signal to an AI algorithm for future processing [41]. This promising application is yet to be investigated on any neck-based assessments, however, this has been used widely in contact-based finger, and wrist based photoplethysmography systems for the purposes of measuring arterial blood pressure [8890]. A recent comparison of several algorithms was recently conducted, and found that Support Vector Regression (SVRs) may marginally outperform other architectures [84].

6.6 Limitations of review

Due to a limitation in resources, this systematic review was limited to a single-reviewer model for the screening and review of articles, which, although methodical, introduces the potential for selection bias. The scope of this review is further constrained by the focus on English-language publications, potentially omitting significant findings reported in other languages. The review was limited to investigations of the jugular vein, and so by doing neglects potential insights drawn from investigations concentrating on the carotid arteries. Due to the shared challenge of ROI selection, future reviews would do well to include both. This presents an analytical challenge for distinguishing between these two sources of signals. To address this, future studies could benefit from a dedicated review of literature focusing on the discrete analysis of signals from both vascular structures.

It is worth noting that the fields of artificial intelligence, deep learning, and neural networks are rapidly advancing. As such, techniques related to ROI selection and segmentation that employ these computational models are likely in nascent stages of development. Although the current body of literature in this domain is limited, we anticipate that future work will explore these avenues, providing more sophisticated methods for signal analysis.

7. Conclusion

This review details the approaches used to measure the jugular venous pulsations in the neck. The utility of the jugular veins is well established and continues to provide clinical utility. Several techniques have been used to accurately show pulsatile signals and large pressure changes in the neck. We have summarised the design of previous studies by their demographics (Table 1) and, experimental setup (Table 2), with results of these investigations shown (Table 3). The technology is now at a stage where it would benefit from some of the resources that exist for general camera-based assessment of vital signs, including published datasets [9194] and a repository of benchmarking algorithms [55,92] to facilitate future investigations.

This review showed that a variety of approaches are all feasible with variations in both hardware (Table 4) and processing techniques, but that a comparison between the approaches is not yet possible. To allow comparisons in the future, the field requires standardisation in reporting, and publicly available datasets. We provide guidance regarding minimum reporting information for future systems with an illustrative example (Table 5). Hardware suggestions are provided to assist new investigators in the field (Table 6). Finally, we envisage the role of AI and machine learning algorithms in the development of automatic ROI segmentation, diagnosis of arrhythmias and the absolute estimation of pressure from imaging systems.

Funding

National Health and Medical Research Council (MRFF 2021 Cardiovascular Health Mission).

Disclosures

The authors declare no conflicts of interest.

Data availability

No data were generated or analyzed in the presented research.

References

1. E. Braunwald, “The war against heart failure: the Lancet lecture,” The Lancet 385(9970), 812–824 (2015). [CrossRef]  

2. C. J. Taylor, J. M. Ordó nez-Mena, A. K. Roalfe, et al., “Trends in survival after a diagnosis of heart failure in the United Kingdom 2000-2017: population based cohort study,” BMJ 364, l223 (2019). [CrossRef]  

3. M. Moradi, F. Daneshi, R. Behzadmehr, et al., “Quality of life of chronic heart failure patients: a systematic review and meta-analysis,” Heart Fail. Rev. 25(6), 993–1006 (2020). [CrossRef]  

4. A. A. Shafie, Y. P. Tan, and C. H. Ng, “Systematic review of economic burden of heart failure,” Heart Fail. Rev. 23(1), 131–145 (2018). [CrossRef]  

5. C. Cook, G. Cole, P. Asaria, et al., “The annual global economic burden of heart failure,” Int. J. Cardiol. 171(3), 368–376 (2014). [CrossRef]  

6. G. Lippi and F. Sanchis-Gomar, “Global epidemiology and future trends of heart failure,” AME Med. J. 5, 15 (2020). [CrossRef]  

7. A. S. Desai and L. W. Stevenson, “Rehospitalization for Heart Failure,” Circulation 126(4), 501–506 (2012). [CrossRef]  

8. C. v. Walraven, C. Bennett, A. Jennings, et al., “Proportion of hospital readmissions deemed avoidable: a systematic review,” CMAJ 183(7), E391–E402 (2011). [CrossRef]  

9. H. Uthoff, C. Thalhammer, M. Potocki, et al., “Central venous pressure at emergency room presentation predicts cardiac rehospitalization in patients with decompensated heart failure,” Eur. J. Heart Fail. 12(5), 469–476 (2010). [CrossRef]  

10. R. M. Khandwalla, K. T. Birkeland, R. Zimmer, et al., “Usefulness of Serial Measurements of Inferior Vena Cava Diameter by VscanTM to Identify Patients With Heart Failure at High Risk of Hospitalization,” The Am. J. Cardiol. 119(10), 1631–1636 (2017). [CrossRef]  

11. R. Nagata, T. Harada, K. Omote, et al., “Right atrial pressure represents cumulative cardiac burden in heart failure with preserved ejection fraction,” ESC Heart Failure 9, 1454–1462 (2022). [CrossRef]  

12. A. Sohn, W. Speier, E. Lan, et al., “Assessment of Heart Failure Patients’ Interest in Mobile Health Apps for Self-Care: Survey Study,” JMIR Cardio 3(2), e14332 (2019). [CrossRef]  

13. C. Pang, J. H. Koo, A. Nguyen, et al., “Highly Skin-Conformal Microhairy Sensor for Pulse Signal Amplification,” Adv. Mater. 27, 634–640 (2015). _eprint: https://onlinelibrary.wiley.com/doi/pdf/10.1002/adma.201403807. [CrossRef]  

14. B. Zha, Z. Wang, L. Li, et al., “Wearable cardiorespiratory monitoring with stretchable elastomer optical fiber,” Biomed. Opt. Express 14(5), 2260–2275 (2023). [CrossRef]  

15. D. Rodbard, “Continuous Glucose Monitoring: A Review of Successes, Challenges, and Opportunities,” Diabetes Technol. & Ther. 18(S2), S2-3–S2-13 (2016). [CrossRef]  

16. “Smart Beat Video Baby Monitor | Best Video Baby Monitor,”.

17. “Oxevision - Oxehealth,”.

18. T. Lewis, “Remarks on Early Signs of Cardiac Failure of the Congestive Type,” BMJ 1(3618), 849–852 (1930). [CrossRef]  

19. B. Lipton, “Estimation of central venous pressure by ultrasound of the internal jugular vein,” The Am. J. Emerg. Med. 18(4), 432–434 (2000). [CrossRef]  

20. F. Abnousi, G. Kang, J. Giacomini, et al., “A novel noninvasive method for remote heart failure monitoring: the EuleriAn video Magnification apPLications In heart Failure studY (AMPLIFY),” npj Digit. Med. 2(1), 80 (2019). [CrossRef]  

21. S. A. Kelly, K. B. Schesing, J. T. Thibodeau, et al., “Feasibility of Remote Video Assessment of Jugular Venous Pressure and Implications for Telehealth,” JAMA Cardiol. 5(10), 1194–1195 (2020). [CrossRef]  

22. X. Liu, S. Patel, and D. McDuff, “RGB Camera-based Physiological Sensing: Challenges and Future Directions,” arXivarXiv:2110.13362 (2021) [CrossRef]   .

23. D. McDuff, “Camera Measurement of Physiological Vital Signs,” ACM Computing Surveys 55, 176 (2021). [CrossRef]  

24. C. Pham, K. Poorzargar, M. Nagappa, et al., “Effectiveness of consumer-grade contactless vital signs monitors: a systematic review and meta-analysis,” J. Clin. Monit. Comput. 36(1), 41–54 (2022). [CrossRef]  

25. V. Selvaraju, N. Spicher, J. Wang, et al., “Continuous Monitoring of Vital Signs Using Cameras: A Systematic Review,” Sensors 22(11), 4097 (2022). Place: Switzerland Publisher: Multidisciplinary Digital Publishing Institute (MDPI). [CrossRef]  

26. S. Zaunseder, A. Trumpp, D. Wedekind, et al., “Cardiovascular assessment by imaging photoplethysmography - a review,” Biomedizinische Tech. Biomed. Eng. 63(5), 617–634 (2018). [CrossRef]  

27. N. Molinaro, E. Schena, S. Silvestri, et al., “Contactless Vital Signs Monitoring From Videos Recorded With Digital Cameras: An Overview,” Front. Physiol. 13, 801709 (2022). [CrossRef]  

28. A. Ni, A. Azarang, and N. Kehtarnavaz, “A Review of Deep Learning-Based Contactless Heart Rate Measurement Methods,” Sensors 21(11), 3719 (2021). [CrossRef]  

29. F.-T.-Z. Khanam, A. Al-Naji, and J. Chahl, “Remote Monitoring of Vital Signs in Diverse Non-Clinical and Clinical Scenarios Using Computer Vision Systems: A Review,” Appl. Sci. 9(20), 4474 (2019). Number: 20 Publisher: Multidisciplinary Digital Publishing Institute. [CrossRef]  

30. W. Wang, A. C. den Brinker, S. Stuijk, et al., “Algorithmic Principles of Remote PPG,” IEEE Trans. Biomed. Eng. 64(7), 1479–1491 (2017). Conference Name: IEEE Transactions on Biomedical Engineering. [CrossRef]  

31. R. Amelard, R. L. Hughson, D. K. Greaves, et al., “Non-contact hemodynamic imaging reveals the jugular venous pulse waveform,” Sci. Rep. 7(1), 40150 (2017). [CrossRef]  

32. S. A. Shafer, “Using color to separate reflection components,” Color. Res. & Appl. 10, 210–218 (1985). [CrossRef]  

33. A. Douplik, G. Saiko, I. Schelkanova, et al., “3 - The response of tissue to laser light,” in Lasers for Medical Applications, H. Jelínková, ed. (Woodhead Publishing, 2013), Woodhead Publishing Series in Electronic and Optical Materials, pp. 47–109.

34. V. Barun and A. Ivanov, “Light scattering by a rough surface of human skin. II. Diffuse reflectance,” Quantum Electron. 43(10), 979 (2013). [CrossRef]  

35. M. J. Page, J. E. McKenzie, P. M. Bossuyt, et al., “The PRISMA 2020 statement: An updated guideline for reporting systematic reviews,” PLoS Med. 18(3), e1003583 (2021). [CrossRef]  

36. W. S. Richardson, M. C. Wilson, J. Nishikawa, et al., “The well-built clinical question: a key to evidence-based decisions,” ACP J. club 123(3), A12–13 (1995). [CrossRef]  

37. T. Wu, V. Blazek, and H. Schmitt, “Photoplethysmography imaging: a new noninvasive and noncontact method for mapping of the dermal perfusion changes,” Proceedings of SPIE - The International Society for Optical Engineering (2000).

38. T. T. Dang, C. Huynh, A. Dinh, et al., “Recognizing area of pulsations on the neck via video camera systems,” in 2015 International Conference on Advanced Technologies for Communications (ATC), (2015), pp. 139–144. ISSN: 2162–1020.

39. R. Bonow, D. Mann, D. Zipes, et al., Braunwald’s Heart Disease: A Textbook of Cardiovascular Medicine, 9th edition (Elsevier Science, 2011).

40. E. J. Lam Po Tang, A. HajiRassouliha, M. P. Nash, et al., “Non-contact Quantification of Jugular Venous Pulse Waveforms from Skin Displacements,” Sci. Rep. 8(1), 17236 (2018). [CrossRef]  

41. A. Moço, P. Hamelmann, S. Stuijk, et al., “The Importance of Posture and Skin-Site Selection on Remote Measurements of Neck Pulsations: An Ultrasonographic Study,” in 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), (2018), pp. 5918–5921.

42. G. Saiko, T. Burton, and A. Douplik, “Feasibility of Specular Reflection Imaging for Extraction of Neck Vessel Pressure Waveforms,” Front. Bioeng. Biotechnol. 10, 830231 (2022). [CrossRef]  

43. R. Amelard, A. D. Robertson, C. A. Patterson, et al., “Optical Hemodynamic Imaging of Jugular Venous Dynamics During Altered Central Venous Pressure,” IEEE Trans. Biomed. Eng. 68(8), 2582–2591 (2021). Conference Name: IEEE Transactions on Biomedical Engineering. [CrossRef]  

44. E. T. Whittaker, “XVIII.-On the Functions which are represented by the Expansions of the Interpolation-Theory,” Proc. R. Soc. Edinb. 35, 181–194 (1915). [CrossRef]  

45. H. Nyquist, “Certain Topics in Telegraph Transmission Theory,” Trans. Am. Inst. Electr. Eng. 47(2), 617–644 (1928). [CrossRef]  

46. C. Shannon, “Communication in the Presence of Noise,” Proc. IRE 37(1), 10–21 (1949). [CrossRef]  

47. A. V. Moço, L. Z. Mondragon, W. Wang, et al., “Camera-based assessment of arterial stiffness and wave reflection parameters from neck micro-motion,” Physiol. Meas. 38(8), 1576–1598 (2017). [CrossRef]  

48. R. van Luijtelaar, W. Wang, S. Stuijk, et al., “Automatic RoI Detection for Camera-Based Pulse-Rate Measurement,” in Computer Vision - ACCV 2014 Workshops, C. JawaharS. Shan, eds. (Springer International Publishing, 2015), Lecture Notes in Computer Science, pp. 360–374.

49. M. Tarvainen, P. Ranta-aho, and P. Karjalainen, “An advanced detrending method with application to HRV analysis,” IEEE Trans. Biomed. Eng. 49(2), 172–175 (2002). Conference Name: IEEE Transactions on Biomedical Engineering. [CrossRef]  

50. R. Amelard, C. Scharfenberger, F. Kazemzadeh, et al., “Feasibility of long-distance heart rate monitoring using transmittance photoplethysmographic imaging (PPGI),” Sci. Rep. 5(1), 14637 (2015). [CrossRef]  

51. A. HajiRassouliha, A. J. Taberner, M. P. Nash, et al., “Subpixel phase-based image registration using Savitzky-Golay differentiators in gradient-correlation,” Comput. Vis. Image Underst. 170, 28–39 (2018). [CrossRef]  

52. H.-Y. Wu, M. Rubinstein, E. Shih, et al., “Eulerian Video Magnification for Revealing Subtle Changes in the World,” ACM Trans. Graph. 31(4), 1–8 (2012). [CrossRef]  

53. N. Wadhwa, M. Rubinstein, F. Durand, et al., “Phase-Based Video Motion Processing,” ACM Trans. Graph. 32(4), 1–10 (2013). [CrossRef]  

54. D. J. McDuff, J. R. Estepp, A. M. Piasecki, et al., “A survey of remote optical photoplethysmographic imaging methods,” in 2015 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), (2015), pp. 6398–6404.

55. G. Boccignone, D. Conte, V. Cuculo, et al., “pyVHR: a Python framework for remote photoplethysmography,” PeerJ Comput. Sci. 8, e929e929 (2022). [CrossRef]  

56. A. Sedrakyan, B. Campbell, J. G. Merino, et al., “IDEAL-D: a rational framework for evaluating and regulating the use of medical devices,” BMJ 353, i2372 (2016). [CrossRef]  

57. C. Brüser, C. H. Antink, T. Wartzek, et al., “Ambient and Unobtrusive Cardiorespiratory Monitoring Techniques,” IEEE Rev. Biomed. Eng. 8, 30–43 (2015). [CrossRef]  

58. W. Verkruysse, L. O. Svaasand, and J. S. Nelson, “Remote plethysmographic imaging using ambient light,” Opt. Express 16(26), 21434–21445 (2008). [CrossRef]  

59. L. P. Malasinghe, N. Ramzan, and K. Dahal, “Remote patient monitoring: a comprehensive study,” J. Ambient Intell. Humaniz. Comput. 10(1), 57–76 (2019). [CrossRef]  

60. R. Ashrafi, M. Azarbayjani, and H. Tabkhi, “Charlotte-ThermalFace: A Fully Annotated Thermal Infrared Face Dataset with Various Environmental Conditions and Distances,” Infrared Phys. & Technol. 124, 104209 (2022). [CrossRef]  

61. J. Talmon, E. Ammenwerth, J. Brender, et al., “STARE-HI–Statement on reporting of evaluation studies in Health Informatics,” Int. J. Med. Informatics 78(1), 1–9 (2009). [CrossRef]  

62. J. Brender, J. Talmon, N. de Keizer, et al., “STARE-HI - Statement on Reporting of Evaluation Studies in Health Informatics: explanation and elaboration,” Appl. Clin. Informatics 04(03), 331–358 (2013). [CrossRef]  

63. P. Nykänen, J. Brender, J. Talmon, et al., “Guideline for good evaluation practice in health informatics (GEP-HI),” Int. J. Med. Informatics 80(12), 815–827 (2011). [CrossRef]  

64. N. F. de Keizer, J. Talmon, E. Ammenwerth, et al., “Systematic prioritization of the STARE-HI reporting items. An application to short conference papers on health informatics evaluation,” Methods Inf. Med. 51(02), 104–111 (2012). [CrossRef]  

65. M. W. Sjoding, R. P. Dickson, T. J. Iwashyna, et al., “Racial Bias in Pulse Oximetry Measurement,” New Engl. J. Med. 383(25), 2477–2478 (2020). [CrossRef]  

66. R. Padwal, N. R. Campbell, A. E. Schutte, et al., “Optimizing observer performance of clinic blood pressure measurement: a position statement from the Lancet Commission on Hypertension Group,” J. Hypertens. 37(9), 1737–1745 (2019). [CrossRef]  

67. T. G. Pickering, J. E. Hall, L. J. Appel, et al., “Recommendations for Blood Pressure Measurement in Humans and Experimental Animals,” Circulation 111(5), 697–716 (2005). Publisher: American Heart Association. [CrossRef]  

68. S. Magder, “Right Atrial Pressure in the Critically Ill: How to Measure, What Is the Value, What Are the Limitations?” Chest 151(4), 908–916 (2017). [CrossRef]  

69. K. Figg and E. Nemergut, “Error in Central Venous Pressure Measurement,” Anesth. analgesia 108(4), 1209–1211 (2009). [CrossRef]  

70. C. Sjödin, S. Søndergaard, L. Johansson, et al., “Variability in alignment of central venous pressure transducer to physiologic reference point in the intensive care unit-A descriptive and correlational study,” Aust. Crit. Care 32(3), 213–217 (2019). [CrossRef]  

71. G. de Haan and V. Jeanne, “Robust pulse rate from chrominance-based rPPG,” IEEE Trans. Biomed. Eng. 60(10), 2878–2886 (2013). [CrossRef]  

72. R. Song, S. Zhang, J. Cheng, et al., “New insights on super-high resolution for video-based heart rate estimation with a semi-blind source separation method,” Comput. Biol. Med 116, 103535 (2020). [CrossRef]  

73. E. B. Blackford and J. R. Estepp, “Effects of frame rate and image resolution on pulse rate measured using multiple camera imaging photoplethysmography,” Proc. SPIE 9417, 94172D (2015). [CrossRef]  

74. J. Speth, N. Vance, P. Flynn, et al., “Unifying frame rate and temporal dilations for improved remote pulse detection,” Comput. Vis. Image Underst. 210, 103246 (2021). [CrossRef]  

75. S. Chatterjee, K. Budidha, and P. A. Kyriacou, “Investigating the origin of photoplethysmography using a multiwavelength Monte Carlo model,” Physiol. Meas. 41(8), 084001 (2020). [CrossRef]  

76. B. E. Bayer, “Color imaging array,” US3971065A (1976).

77. D. McDuff, S. Gontarek, and R. W. Picard, “Improvements in Remote Cardiopulmonary Measurement Using a Five Band Digital Camera,” IEEE Trans. Biomed. Eng. 61(10), 2593–2601 (2014). [CrossRef]  

78. S. Prahl, “Optical absorption of hemoglobin,” OMLC, 1999, https://omlc.org/spectra/hemoglobin/.

79. M. J. Butler, J. A. Crowe, B. R. Hayes-Gill, et al., “Motion limitations of non-contact photoplethysmography due to the optical and topological properties of skin,” Physiol. Meas. 37(5), N27–N37 (2016). [CrossRef]  

80. S. Rasche, A. Trumpp, T. Waldow, et al., “Camera-based photoplethysmography in critical care patients,” Clin. Hemorheol. Microcirc. 64(1), 77–90 (2016). [CrossRef]  

81. M. Huelsbusch and V. Blazek, “Contactless mapping of rhythmical phenomena in tissue perfusion using PPGI,” in Medical Imaging 2002: Physiology and Function from Multidimensional Images, vol. 4683 (SPIE, 2002), pp. 110–117.

82. G. Lempe, S. Zaunseder, T. Wirthgen, et al., “ROI Selection for Remote Photoplethysmography,” in Bildverarbeitung für die Medizin 2013, H.-P. Meinzer, T. M. Deserno, H. Handels, T. Tolxdorff, eds. (Springer, Berlin, Heidelberg, 2013), Informatik aktuell, pp. 99–103.

83. E. B. Schroeder, D. Liao, L. E. Chambless, et al., “Hypertension, Blood Pressure, and Heart Rate Variability,” Hypertension 42(6), 1106–1111 (2003). Publisher: American Heart Association. [CrossRef]  

84. L. Cerina, L. Iozzia, and L. Mainardi, “Influence of acquisition frame-rate and video compression techniques on pulse-rate variability estimation from vPPG signal,” Biomedizinische Tech. Biomed. Eng. 64(s1), 53–58 (2019). [CrossRef]  

85. D. J. McDuff, E. B. Blackford, and J. R. Estepp, “The Impact of Video Compression on Remote Cardiac Pulse Measurement Using Imaging Photoplethysmography,” in 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), (2017), pp. 63–70.

86. M. Rapczynski, P. Werner, and A. Al-Hamadi, “Effects of Video Encoding on Camera-Based Heart Rate Estimation,” IEEE Trans. Biomed. Eng. 66(12), 3360–3370 (2019). [CrossRef]  

87. J. Lee, K. Matsumura, K.-I. Yamakoshi, et al., “Comparison between red, green and blue light reflection photoplethysmography for heart rate monitoring during motion,” Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. 2013, 1724–1727 (2013). [CrossRef]  

88. A. Trumpp, P. L. Bauer, S. Rasche, et al., “The value of polarization in camera-based photoplethysmography,” Biomed. Opt. Express 8(6), 2822–2834 (2017). [CrossRef]  

89. W. Chen and D. McDuff, “DeepPhys: Video-Based Physiological Measurement Using Convolutional Attention Networks,” in Computer Vision - ECCV 2018, V. Ferrari, M. Hebert, C. Sminchisescu, and Y. Weiss, eds. (Springer International Publishing, 2018), Lecture Notes in Computer Science, pp. 356–373.

90. G. Slapnicar, N. Mlakar, and M. Luštrek, “Blood Pressure Estimation from Photoplethysmogram Using a Spectro-Temporal Deep Neural Network,” Sensors 19(15), 3420 (2019). Place: Switzerland Publisher: Multidisciplinary Digital Publishing Institute (MDPI). [CrossRef]  

91. C. El-Hajj and P. A. Kyriacou, “A review of machine learning techniques in photoplethysmography for the non-invasive cuff-less measurement of blood pressure,” Biomed. Signal Process. Control. 58, 101870 (2020). [CrossRef]  

92. A. Paviglianiti, V. Randazzo, S. Villata, et al., “A Comparison of Deep Learning Techniques for Arterial Blood Pressure Prediction,” Cogn. Comput. 14(5), 1689–1710 (2022). [CrossRef]  

93. S. González, W.-T. Hsieh, and T. P.-C. Chen, “A benchmark for machine-learning based non-invasive blood pressure estimation using photoplethysmogram,” Sci. Data 10(1), 149 (2023). [CrossRef]  

94. X. Niu, Z. Yu, H. Han, et al., “Video-Based Remote Physiological Measurement via Cross-Verified Feature Disentangling,” in Computer Vision - ECCV 2020, A. Vedaldi, H. Bischof, T. Brox, J.-M. Frahm, eds. (Springer International Publishing, 2020), Lecture Notes in Computer Science, pp. 295–310.

Data availability

No data were generated or analyzed in the presented research.

Cited By

Optica participates in Crossref's Cited-By Linking service. Citing articles from Optica Publishing Group journals and other participating publishers are listed here.

Alert me when this article is cited.


Figures (3)

Fig. 1.
Fig. 1. PRISMA Systematic Review Flowchart
Fig. 2.
Fig. 2. Overview of Study Design. The design of a study consists of both the system itself, and the environment where it is applied.
Fig. 3.
Fig. 3. An illustration of the setup described Table 5. The setting of the study was in the emergency department of the hospital in Argentina. The participant is in a seated position, facing forward. The camera on the phone is situated at a distance of 30 cm from the neck, at an angle of 45°from the front plane of the person. The flash of the light from the phone is used to illuminate the neck. Adapted from vecteezy.com

Tables (6)

Tables Icon

Table 1. Summary of Demographic Information for the included studies.

Tables Icon

Table 2. Summary of Experimental Setup Information for the included studies.a

Tables Icon

Table 3. Summary of the quantitative and qualitative for the included studies.a

Tables Icon

Table 4. Summary of Hardware Information for the included studies.a

Tables Icon

Table 5. Recommendations for minimum reporting guidelines. The example provided is fabricated, for the illustration of how this may be reported. The use of [citation] and [table] indicate that a citations and tables should be included, providing a link to the model information if possible.

Tables Icon

Table 6. Hardware guidance based on general remote photoplethysmography analyses.a

Equations (3)

Equations on this page are rendered with MathJax. Learn more.

I pixel ( t ) = Q efficiency ( λ ) I sensor ( t , λ ) d λ + N sensor ( t )
I sensor ( t , λ ) = A 4 π d 2 θ 1 θ 1 I surface ( t , λ , θ ) d θ
I surface ( t , λ , θ ) = I incident ( t , λ , θ ) ( r specular ( t , λ , θ ) + r diffuse ( t , λ , θ ) )
Select as filters


Select Topics Cancel
© Copyright 2024 | Optica Publishing Group. All rights reserved, including rights for text and data mining and training of artificial technologies or similar technologies.