Next Article in Journal
State-of-the-Art Sensors Research in Ireland
Previous Article in Journal
Carbon Black/PDMS Based Flexible Capacitive Tactile Sensor for Multi-Directional Force Sensing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Contactless Measurement of Vital Signs Using Thermal and RGB Cameras: A Study of COVID 19-Related Health Monitoring

1
Health Devices Research Group (HDRG), School of Electrical Engineering and Computer Science, University of Ottawa, Ottawa, ON K1N 6N5, Canada
2
Institute for Biomedical Engineering, Science and Technology (iBEST), St. Michael’s Hospital, Ryerson University, Toronto, ON M5B 1T8, Canada
3
WelChek Inc., Mississauga, ON L4W 4Y8, Canada
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2022, 22(2), 627; https://doi.org/10.3390/s22020627
Submission received: 26 October 2021 / Revised: 7 January 2022 / Accepted: 10 January 2022 / Published: 14 January 2022
(This article belongs to the Section Physical Sensors)

Abstract

:
In this study, a contactless vital signs monitoring system was proposed, which can measure body temperature (BT), heart rate (HR) and respiration rate (RR) for people with and without face masks using a thermal and an RGB camera. The convolution neural network (CNN) based face detector was applied and three regions of interest (ROIs) were located based on facial landmarks for vital sign estimation. Ten healthy subjects from a variety of ethnic backgrounds with skin colors from pale white to darker brown participated in several different experiments. The absolute error (AE) between the estimated HR using the proposed method and the reference HR from all experiments is 2.70 ± 2.28 beats/min (mean ± std), and the AE between the estimated RR and the reference RR from all experiments is 1.47 ± 1.33 breaths/min (mean ± std) at a distance of 0.6–1.2 m.

1. Introduction

With the outbreak of the COVID-19 pandemic, the worldwide healthcare systems are facing challenges that have never been met before. Most common symptoms of COVID-19 include fever, dry cough, and shortness of breath or difficulty breathing. Some researchers have found that COVID-19 can cause unexplained tachycardia (rapid heart rate) [1]. It is important and useful to monitor the vital signs of people before they get a diagnosis from the hospital as well as to track the vital signs during their recovery. The vital signs, such as body temperature (BT), respiratory rate (RR) and heart rate (HR), could be used as evidence for prescreening suspected COVID-19 cases. In order to assist in detecting COVID-19 symptoms, we developed a contactless vital sign monitoring system, which can detect the BT, HR and RR at a safe social distance (60–120 cm). The proposed system and algorithms can be potentially applied for quick vital signs measurement at the entry of buildings, such as shopping malls, schools and hospitals.
Compared to previous work [2,3,4], our method improves the contactless vital signs measurement range up to 2 m for RR estimation and 1.2 m for HR estimation while preserving the high estimation accuracy. Besides, our proposed approach addresses the challenge when the subject is wearing the mask and the face is partly occluded using the CNN based face detector. In addition, the face detector in our method enables us to measure multiple subjects’ vital signs such as RR and HR at the same time, which, to the best of our knowledge, has never been studied before. To summarize, our method makes the following contributions:
  • We present a system that estimates multiple subjects’ vital signs including HR and RR using thermal and RGB cameras. To the best of our knowledge, it is the first study that includes different face masks in contactless RR estimation and our results indicate that the proposed system is feasible for COVID 19-related applications;
  • We propose signal processing algorithms that estimate the HR and RR of multiple subjects under different conditions. Examples of novel approaches include increasing the contrast of the thermal images to improve the SNR of the extracted signal for RR estimation as well as a sequence of steps including independent component analysis (ICA) and empirical mode decomposition (EMD) to enhance heart rate estimation accuracy from RGB frames. Robustness is improved by performing a signal quality assessment of the physiological signals and detecting the deviation in the orientation of the head from the direction towards the camera. By applying the proposed approaches, the system can provide accurate HR and RR estimations with normal indoor illuminations and for subjects with different skin tones;
  • Our work addresses some of the issues reported in other works such as the small distance required between the cameras and the subjects and the need to have a large portion of the face exposed to the camera. Therefore, our system is robust at larger distances, and can simultaneously estimate the vital signs of two people whose faces might be partially covered with face masks or not pointed directly towards the cameras.

2. Related Works

Conventionally, contact based devices are widely used in hospitals and nursing homes for the measurement of vital signs. For example, electrocardiography (ECG) and photoplethysmography (PPG) sensors are usually used in HR monitoring [5] and the breathing belt is used to measure the RR of the subject. However, these devices need to be attached to the subjects’ body thereby constraining or causing them discomfort and thus potentially affecting the measurement. It is also hard to make many people wear this kind of sensor and collect their data when they are in public places.
To resolve these problems, contactless monitoring systems have been designed and applied in the aforementioned scenario. Contactless monitoring systems rely on a number of sensors such as radars, sonars, different kind of camera and so on. In this paper, we will limit our analysis to camera-based systems. Xiaobai Li and Jie Chen et al. proposed a method that can measure HR remotely from face videos [2]. After that, many researchers used the green channel of face images collected from a webcam or smartphone to measure the heart rate [6,7]. On the other side, with the decline of the cost and size of optical imaging devices such as infrared cameras and thermal cameras, the respiratory signal can also be measured by remote monitoring [8,9,10]. Youngjun Cho et al. proposed a respiratory rate tracking method based on mobile thermal imaging [3]. One low-cost and small-sized thermal camera connected to a smartphone was used as a thermographic system to detect the temperature changes of the nostril area caused by the inhaling and exhaling. There have been some other works [4,11,12] related to monitoring the respiratory signal using the thermal camera. However, the detection distance between the human face and the thermal camera is limited to less than 50 cm in Cho’s work [3] and the estimation error increases with the distance between the subject and the camera. In addition, some face detection algorithms [3,4,11] perform poorly in detecting occluded faces if the subjects are wearing face masks and make it improper for COVID symptoms’ detection as a face mask is always mandatory in public places.
In our task of monitoring a subject’s HR and RR using cameras, the subject’s face should be detected from the video frames where the face detection and tracking algorithms need to be developed and applied. The Kanade–Lucas–Tomasi (KLT) feature tracker originally proposed in 1991 is generally used to track the faces and the region of interest (ROI) in videos [5]. In addition, there are many face detectors in OpenCV and Matlab toolboxes such as the Viola–Jones face detector [13] and the Discriminative Response Map Fitting (DRMF) detector [14]. However, these detectors have a problem handling the occluded faces so they cannot be used in our scenario. Face detection also receives attention in the deep learning community where many convolutional neural network frameworks have been developed in recent years to address the challenges. Multi-task Cascaded Convolutional Networks (MTCNN), proposed by Kaipeng Zhang et al. in 2016, which adopt a cascaded structure of deep convolutional networks to detect the face and landmarks [15], were widely used in the task of detecting faces. However, MTCNN cannot resolve the problem when the face is covered with a mask. The Single Shot Scale-invariant Face Detector (S3FD) [16] proposed in 2017 also does not perform well in occluded face detection. PyramidBox, proposed in 2018 [17], is capable of handling the hard face detection problems such as blurred and partially occluded faces in an uncontrolled environment. Though PyramidBox has already been applied to detecting faces occluded by the mask in the application of respiratory infections detection using RGB-Infrared sensors [18], it cannot be applied to scenarios where the facial landmarks or ROI detection are required. Besides the face detection, fitting methods for face alignment [19] are also significant in our task, because we need to detect the interested region based on facial landmarks. However, this landmarks detection relies on complete faces without occlusion.

3. Materials and Protocols

3.1. Data Acquisition System

In this study, a thermal camera (Rakinda Technology, FT20) and an RGB webcam (Logitech©, C270) were employed to collect thermal and RGB videos from human subjects. The thermal camera operates at 25 fps and its resolution is 256 × 192, the RGB camera operates at 30 fps and its resolution is 640 × 480. These cameras are placed side by side horizontally on a pole and the height is 1.6 m. An LED camera light (Neewer, model 176 LED) was placed close to the cameras to enhance illuminations on subjects’ faces and to avoid shadows (normal indoor illumination level). The setup of the devices is shown in Figure 1. To evaluate the performance of the proposed system in vital sign estimation, an optical PPG sensor (Sparkfun, heart rate sensor), a force-based chest respiration belt (Vernier, Go Direct) and a handheld infrared forehead thermometer were used as reference measurements of heart rate, respiration rate and body temperature, respectively.
A desktop (Intel Core i9, Nvidia GeForce RTX 2080 Ti, 64 GB RAM, Ubuntu 20 OS) was used for video acquisitions from the thermal and RGB cameras, reference signal acquisition, visualization, data storage and analysis. Face detection and tracking, signal processing and the proposed algorithms were developed using Python 3.7.

3.2. Experimental Protocols

In this study, ten healthy subjects (seven males, three females, 24–35 years old) participated in several different experiments. The subjects came from a variety of ethnic backgrounds with skin colors from pale white to darker brown. The protocol was approved by the University of Ottawa Research Ethics Board.
One subject (male, 27 years) participated in RR estimation with different face masks at different detection ranges. At each detection range (75 cm, 120 cm and 200 cm), the subject was asked to stand still in front of the thermal and RGB cameras without face mask and with three different face masks (medical/cloth, N90, N95) for two minutes respectively. The subject performed normal breaths during the test. These experiments aim to test the performance of RR monitoring using a thermal camera when a subject has a face cover.
Ten healthy subjects (seven males, three females, 24–35 years old) participated in one-subject vital sign estimation experiments. Each subject was asked to keep stationary and breathe normally in front of the cameras at four different detection ranges (60 cm, 80 cm, 100 cm and 120 cm) separately. At each distance, simultaneous thermal and RGB videos were collected from the subject with and without a face mask (medical/cloth mask) for two minutes respectively. The simultaneous reference respiration waveform was collected from the subject’s chest and the reference PPG signal was collected from the subject’s fingertip. The reference body temperature was measured by the infrared thermometer from the subject’s forehead every 30 s.
Six subjects (four males, two females, 24–28 years old) participated in two-subjects vital sign estimation experiments. In each experiment, two subjects were asked to keep stationary and breathe normally in front of the cameras at different distances (70 cm, 100 cm) for two minutes respectively. Subjects were not wearing face masks in Experiments 1–3 and they were wearing face masks in Experiments 4–5 as demonstrated in Section 5.2.3. The reference respiration waveforms were collected from subjects’ chests and the reference PPG signals were collected from the subject’s fingertips simultaneously.

4. Methods

The steps of remote estimation of heart rate, respiration rate and body temperature are shown in Figure 2. In our system, RGB and thermal videos are collected by RGB and thermal cameras simultaneously, and the sampling rate of these two cameras is 25 frames per second (fps). The RGB camera is used to detect and track human subjects’ faces, and thus locate three different regions of interest (ROIs) for vital signs estimation. The forehead area from the RGB video sequence is extracted to estimate HR. In addition, with the help of image alignment process, a point from the forehead area and the nostril area can be located on the simultaneous thermal video frames for BT and RR estimation.

4.1. Face Detection

We applied the RetinaFace [20] framework to detect faces from the RGB video frames. RetinaFace was proposed by Deng et al. and ranks second in the competition of WIDER FACE (Hard). The WIDER FACE dataset [21] is a face detection benchmark dataset with 32,203 images and 393,703 labeled faces which has a high degree of variability in scale, pose and occlusion. RetinaFace is a practical single-stage deep learning based face detector working on each frame of the video. Consequently, we chose it because only the RetinaFace detector can detect both face and facial landmarks with the confidence score over 0.9 comparing to some widely used face detectors including MTCNN, S3FD and PyramidBox. An example of face and facial landmarks detection using these face detectors when a subject was wearing a face mask is illustrated in Figure 3.
Additionally, RetinaFace can detect multiple faces from one image as illustrated in Figure 4, which makes it possible for our system to detect multiple subjects’ vital signs even when their faces are partially occluded by face masks at the same time.

4.2. Regions of Interest (ROIs)

In this study, the proposed system aims to monitor three vital signs (BT, HR and RR) remotely. Therefore after detecting subject’s face, three different regions of interest (ROIs) needs to be located from RGB or thermal video frames. The locations of the ROIs are illustrated in Figure 5.

4.2.1. ROI for BT Estimation

Like most of the non-contact thermometers measurements, a single point ( R O I B T ) on the forehead is identified as the interested point ( T x , T y ) for body temperature estimation in our system. After detecting the bounding box of the face and locating the positions of the eyes, T y is defined as H 1 / 3 above the middle point of the left and right eyes, where H 1 is the difference between the upper edge of the bounding box and the middle point of eyes on y-axis. T x equals the mean value of x coordinates of the left and right eyes. A demonstration of locating R O I B T using the bounding box and facial landmarks is shown in Figure 6a.

4.2.2. ROI for RR Estimation

In this study, a nostril is determined as the R O I R R for RR estimation as shown in Figure 5b. Warm air from inside the lungs is released through the respiratory system and it increases the temperature in the nasal region during exhalation, whereas cool air from the external environment is breathed in and it lowers the temperature in the nasal region during inhalation. Therefore, the respiration waveform can be obtained by using an infrared thermal camera to measure such nasal-region temperature changes associated with respiration. Consequently, the nostril areas are chosen as an ROI for respiratory signal detection. The nose tip landmark is chosen as the center-point of the nostril area. The width of R O I R R is defined as 90 percent of the mouth width W 2 , and the height of R O I R R is 60 percent of the vertical difference ( H 2 ) between the mouth middle point and the nose tip. An illustration of R O I R R can be found in Figure 6a.

4.2.3. ROI for HR Estimation

Phenomena exploited in rBVP (remote blood volume pulse) for HR estimation are closely related to the cardiac cycle. During each cycle, blood is moved from the heart to the head through the carotid arteries. This periodic blood inflow affects both optical properties of facial skin and the mechanical movement of the head, enabling researchers to measure HR remotely [22].
Several different ROIs used for remote HR estimation are described in existing research. Some researchers used the bounding box of the face given by the face detection algorithm as the ROI [6,23,24,25,26]. Other researchers used the forehead area of the detected face as the ROI for HR estimation [6,23,27]. Through our experiments, the forehead area always performed better than the face area in HR estimation and the forehead will not be occluded by a face mask, therefore the forehead was determined as R O I H R . The center point of the R O I H R is R O I B T ( T x , T y ) . The height of the R O I H R is 0.4 of the height of the bounding box and the width of the R O I H R is 0.5 of the width of the bounding box detected by the RetinaFace detector. A demonstration of R O I H R location can be found in Figure 6b.

4.3. Head Movement Detection

The facial landmarks are important for the division of the face and detection of our interested regions as illustrated in Figure 7a. To achieve accurate estimations of the vital signs, the subjects are supposed to keep stationary and look at the cameras during the data collection. However, there may be some scenarios where the subjects move their heads to other directions as shown in Figure 7b. If the movement is too large, we assume that it is not possible to reliable extract the vital signs and therefore, there is no longer need to extract the data from the ROIs. We calculate the ratio of the distance between the midpoint of the two detected eyes and the midpoint of the face along the x axis to the width of the face box:
F a c e d e f l e c t i o n = | T x F a c e x | F a c e w i d t h ,
where T x represents the x coordinate of the midpoint of the left eye and the right eye, F a c e x represents the x coordinate of the midpoint of the face box. F a c e w i d t h is the width of the face box, and F a c e d e f l e c t i o n denotes the ratio of the distance between the eyes’ midpoint x coordinate and the face midpoint x coordinate to the face width. The ratio F a c e d e f l e c t i o n increases with the face moving away from the camera and the variation of the value describes the status of the subject’s face whether still or moving. We set the ratio F a c e d e f l e c t i o n threshold to 0.17 when the face is not towards the camera as illustrated in Figure 7b and we do not need to process the data to estimate vital signs under this condition anymore. It is noted that we only consider horizontal face inclines rather than the vertical inclines because our device is mounted at a certain height and the subject is required to face the camera during the experiments.

4.4. Frame Registration

After locating the ROIs from the RGB frame, the R O I R R and R O I B T also needs to be located on the thermal image as the RR and BT estimations are based on temperature readings. The resolution of the thermal camera is also lower than the RGB camera, hence an image alignment method needs to be applied. In our system, two cameras are placed side by side in the same plane. Consequently, the alignment of frames from two cameras is simplified to an affine transformation problem. Affine transformation is a linear mapping method that preserves points, straight lines, and planes. Sets of parallel lines remain parallel after an affine transformation. We only consider translation and scaling in our two-camera system.
Transformation matrix T is defined as:
T = s x 0 0 0 s y 0 t x t y 1 ,
where t x specifies the displacement along the x axis, t y specifies the displacement along the y axis, s x specifies the scale factor along the x axis, and s y specifies the scale factor along the y axis.
Translation transformation is caused by the difference between the position of the two parallel cameras. Since the imaging planes of the RGB camera and the thermal camera are within the same plane, the displacements are along the x axis and the y axis. Scale transformation is caused by the focal length. The lens of the RGB camera has a shorter focal length compared to the lens of the thermal camera, and so the RGB camera captures a much wider field of view while producing a smaller picture. Figure 8a shows the original frame with 640 × 480 captured by the RGB camera, and Figure 8b,c show the synchronous and registered thermal frame and RGB frame with 256 × 192 using the transformation matrix T.

4.5. Vital Signs Estimation

4.5.1. Body Temperature Measurement

Thermal cameras can measure a subject’s surface skin temperature without being physically close to the person being evaluated. The FT20 thermal imaging module provides us with 256 × 192 temperature values in each frame. Among them, the device outputs the data including the temperature of the center, the highest temperature and location, the lowest temperature and location, and the overall temperature in each pixel location within the frame. Consequently, we estimate the forehead skin temperature by directly extracting from the location of the forehead center point ( T x , T y ) in the thermal frame.
All of the 256 × 192 temperature values in one thermal frame are raw data generated by the thermal camera following the sequence of a pixel array. However, the raw temperature data are processed and transformed using the palette to be visible as pixels in the output thermal frame. Thus, the coordinate of the forehead center point ( T x , T y ) functions as a key of the dictionary to extract forehead temperature from the table of raw data.
The calibration of the thermal camera is done by using a black body device and has been tested by the camera manufacturer before the delivery. However, the thermal system measures surface skin temperature, which is usually lower than a temperature measured orally, and the experiments also show that the detection range and environmental temperature could have an impact on the measurement. So there are some difference between the measured value and the real body temperature. We also compare our temperature measurements with the handheld thermometer measurements, which only shows a subtle difference of less than 0.5 Celsius degrees when measuring the forehead skin temperature.

4.5.2. Heart Rate Estimation

The steps of HR estimation proposed in this study are shown in Figure 9. After locating the R O I H R from an RGB video frame by frame, the raw color channel signals can be extracted by calculating the mean values of red, green and blue channels within R O I H R . A 3rd order Butterworth bandpass filter with the passband of 0.8–2 Hz (corresponding to 48–120 beats/min) is applied on raw color signals for noise reduction.
Several different approaches have been considered and tested. The independent component analysis (ICA) method and the empirical mode decomposition (EMD) based signal selection method which perform well in HR estimation for subjects with different skin tones and in different illuminations are proposed for rBVP waveform recovery. The ICA, which was originally employed in [7], is widely used in vision-based HR estimation. It is a blind source separation technique used to separate multivariate signals from underlying sources. ICA can decompose the RGB signals into three independent source signals. Any of them can carry rBVP information. For R O I H R extracted at time point t, the denoised signals from red, green and blue channels are denoted as x 1 ( t ) , x 2 ( t ) and x 3 ( t ) respectively. Then, these RGB traces are normalized as follows:
x i ( t ) = x i ( t ) μ i σ i ,
where μ i and σ i for i = 1 , 2 , 3 are the mean and standard deviation of x i ( t ) , respectively. The normalization transforms x i ( t ) to x i ( t ) , which is zero-mean and has unit variance.
The normalized traces are then decomposed into three independent source signals using ICA. The joint approximate diagonalization of the eigenmatrices (JADE) algorithm developed by Cardoso [28] is applied, and the second component, which always contains the strongest plethysmographic signal, is selected as the desired source signal. Then the EMD, which can decompose the signal into a series of IMFs (Intrinsic Mode Functions), is applied to the extracted signal and the power of each IMF is calculated. The IMF with the maximum power is selected as the recovered rBVP signal for HR estimation.
The systolic peaks of the extracted rBVP signal were detected using a peak detection algorithm. Then the mean time interval T I H R between two consecutive systolic peaks was calculated. Therefore, the HR estimated using an RGB camera can be obtained as H R R G B = 60 / T I H R (beats/min). An example of the extracted rBVP waveform using the proposed approaches and the simultaneous reference PPG waveform is shown in Figure 10b.

4.5.3. Respiration Rate Estimation

After locating the nasal region from the thermal video, the respiratory signal is extracted by averaging the values of all pixels within R O I n o s t r i l frame by frame. However, the respiratory signal extracted directly from the original thermal frames has a small magnitude when the respiration-induced thermal variance is weak, for example, during shallow breathing or at a long detection distance. Consequently, we applied a histogram equalization to solve the weak signal problem. Here we use the Contrast Limited Adaptive Histogram Equalization (CLAHE) algorithm [29] to increase the contrast by spreading out the most frequent intensity values. As a result, we enhance the variation of ROI within thermal frames as illustrated in Figure 11 before we attempt to extract the respiratory signal.
Then, a 3rd order Butterworth bandpass filter with a lower cutoff frequency of 0.15 Hz and a higher cutoff frequency of 0.5 Hz (corresponding to 9–30 breaths/min) was applied to the extracted respiratory signal for noise removal.
The peaks of the extracted respiration signal were detected using a peak detection algorithm. Then the mean time interval T I R R between two consecutive peaks was calculated. Therefore, the RR estimated using a thermal camera can be obtained as R R t h e r m a l = 60 / T I R R (breaths/min). An example of the extracted respiratory waveform using the proposed approaches and the simultaneous reference respiration waveform from a breath belt is shown in Figure 10a. It is worth mentioning that the reference breath belt is a force sensor based device and the force value increases during inhalation and decreases during exhalation; however, the temperature at the nostril area decreases during inhalation and increases during exhalation.

4.6. Signal Quality Evaluation

An ensemble averaging based signal quality assessment method was implemented to evaluate the stationarity of the extracted respiratory waveform and rBVP waveform. First, the peaks of the target signal are detected. The template window length is determined as the median value of the durations between two successive peaks. Individual pulses are then aligned with their peaks centered in the middle of the template window and these pulses are averaged to form the signal template. The Pearson correlation coefficients between each individual pulse and the signal template are calculated and their mean value is extracted as the signal quality index (SQI) of this signal segment. Finally, the signal segment will be used for RR and HR estimation if S Q I R R > 0.7 and S Q I H R > 0.8 respectively. An example of the proposed signal quality assessment method applied on a 15 s extracted rBVP signal is shown in Figure 12.

5. Results

5.1. Facial ROIs Detection

Our ROI detection method based on the RetinaFace detector, is deployed both on a desktop computer and an embedded device Nvidia Jetson Board for testing. The classification scores showing the probability of the detection of a human face that are tested on all of the four subjects at different positions with or without mask reach 99.95% or even higher using Retinaface detector. Besides, there are no difference between the subject wearing the mask and the one without mask when we evaluate the face classification probabilities. In addition, to evaluate the performance of facial landmarks detection which is the base of our facial ROI detection, we use the normalised mean error (NME) metric:
N M E = 1 N i = 1 N Δ x i 2 + Δ y i 2 d ,
where N denotes the number of facial landmarks, which is five in our situation and d denotes the normalized distance. Δ x i 2 and Δ y i 2 are deviations between the ith predicted landmark and ground truth in x axis and y axis. We employ the face box size ( W × H ) as the d in the evaluation. We found that the average of NME is 1.5% when the subject is not wearing the mask, while the NME reaches 2.4% when the subject is wearing the mask.
In terms of execution time, we applied the whole RetinaFace detector on the embedded device NVIDIA Jetson TX2 board. Jetson TX2 is built around an NVIDIA Pascal™-family GPU and loaded with 8 GB of memory and 59.7 GB/s of memory bandwidth. The RetinaFace detector takes average 5 s to load the whole network model. Besides it takes around 0.7 s to detect each frame, which means that the fps is more than 1. Due to the limited computing power, the embedded device is still insufficient to run the face detector in real time.
On the other side, RetinaFace detector has a much shorter running time on a desktop computer with more computing power. We applied the whole network on a machine with a NVIDIA GeForce RTX 2080 Ti GPU that has a total memory of 10.76 GB. This graphical processing unit has 7.5 times larger computing capacity than the one in the embedded device mentioned above. The RetinaFace detector takes an average of 0.025 s to detect one frame and the running speed reaches around 40 fps, which is far better than the real time detection on an embedded platform. However, the detection time will grow with the increase of the number of subjects in the image. For example, when there is only one subject in the image, the detector takes less than 0.02 s to detect the faces in one frame.

5.2. Vital Sign Estimation

To simulate the most potential use case of this study where people’s vital signs are remotely measured at the entrances (i.e., stores, buildings), the collected signals from our experiments were divided into 15 s’ segments using the sliding window method with 50% overlap between two consecutive segments for analysis.

5.2.1. Respiration Rate Estimation

The mean absolute error (MAE) between the estimated RR and reference RR of each respiration cycle was employed to evaluate the performance of the proposed system. The respiratory signal examples extracted from R O I R R and a reference breath belt of a subject with different face masks and without face mask can be found in Figure 13. A demo video of respiration monitored by a thermal camera when a subject was wearing different face masks can be found in Supplementary Material Video S1. The MAE between the estimated RR and reference RR of this subject with different face masks and without face masks at different detection ranges is shown in Table 1.
For one-subject RR estimation with and without face masks at different detection ranges, the mean absolute RR estimation error for each respiration cycle were calculated for each subject and can be found in Table 2. The correlation plots and the Bland–Altman plots of the estimated RR and the reference RR for experiments when the subjects with and without face masks are shown in Figure 14. The MAE for all one-subject RR experiments is 1.52 breaths/min. For one-subject experiments, when subjects were not wearing face masks, the MAE is 1.83 breaths/min and the Pearson correlation coefficient between the estimated RR and reference RR is 0.83 as shown in Figure 14a. The mean bias is −0.012 breaths/min and the limits of agreement are −4.63 breaths/min and 4.40 breaths/min as demonstrated in Figure 14b. For one-subject experiments, when subjects were wearing face masks, the MAE is 1.21 breaths/min and the Pearson correlation coefficient between the estimated RR and reference RR is 0.91 as shown in Figure 14c. The mean bias is −0.24 breaths/min and the limits of agreement are −3.45 breaths/min and 2.96 breaths/min as demonstrated in Figure 14d.
It can be observed that the overall performance of RR estimation for experiments with face masks is better than experiments without face masks. This is because the area of temperature changes around the nostril is larger and the temperature changes caused by respiration are more obvious when subjects are wearing face masks.

5.2.2. Heart Rate Estimation

The absolute difference ( e H R ) between the estimated HR and reference HR of each cardiac cycle was employed to evaluate the performance of the proposed system. The mean ( M e H R ) and standard deviation ( S D e H R ) of the absolute HR estimation error for each cardiac cycle are calculated for each subject different distances, as shown in Table 3. The correlation plots and the Bland-Altman plots of the estimated HR and the reference HR for subjects with different skin tones are shown in Figure 15. The MAE for all one-subject HR experiments is 2.79 beats/min. For subjects who have a pale skin tone, the MAE is 2.87 beats/min and the Pearson correlation coefficient between the estimated HR and reference HR is 0.96 as shown in Figure 15a. The mean bias is 0.17 beats/min and the limits of agreement are −7.21 beats/min and 7.56 beats/min as demonstrated in Figure 15b. For subjects who have medium or dark skin tones, the MAE is 2.66 beats/min and the Pearson correlation coefficient between the estimated HR and reference HR is 0.97 as shown in Figure 15c. The mean bias is 0.25 beats/min and the limits of agreement are −6.53 beats/min and 7.03 beats/min as demonstrated in Figure 15d.

5.2.3. Two-Subjects RR and HR Estimation

The feasibility of estimating multiple subjects’ RR and HR at the same time using our proposed methods was verified. Two subjects with and without face masks were required to stand in front of the cameras for two minutes at a distance of 70 cm and 100 cm, respectively. The breathing belts were used to collect reference respiratory waveform from their chests and PPG sensors were used to measure reference HR from their fingertips. The mean and standard deviation of the absolute errors between the estimations and references for each experiment are demonstrated in Table 4.

6. Conclusions

This paper presented a technique for contactless measurements of three vital signs of a subject including forehead temperature, respiratory rate and heart rate at the same time. We designed the system based on one thermal camera and one RGB camera. Ten healthy subjects with different skin colors participated in several different experiments to verify the feasibility of the proposed system for real-world applications. By comparing with the reference data collected at the same time from other devices, such as a thermometer, a respiration chest-belt and a PPG sensor, it was shown that our methods have acceptable accuracy in estimating BT, RR and HR.
Our methods can detect multiple subjects in a very short time, and can handle the problem of faces being occluded by masks very well. Our methods perform well in estimating vital signs at near distances (60 cm–120 cm) for subjects with and without face masks, which are great improvements compared to previous research because we enhance the detection range and allow for the wearing of masks and slight movement of the subjects during the measurements. As a result, our system is capable of contactless monitoring of vital signs and can be potentially applied for the entrance screening of people’s health condition. For application scenarios, such as detecting the vital signs in an emergency situation, HR measurement with an error less than 5 beats/min is likely to be acceptable [7].
For future work, we will implement our methods on the embedded devices or mobile devices to make the whole system portable and capable of running in real time, and a larger study population will be included to validate the stability and repeatability of the proposed approach. In addition, we will use our methods on other thermal cameras and RGB cameras to test the robustness and portability of the methods.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s22020627/s1, Video S1: Respiration monitored by a thermal camera with different face masks.

Author Contributions

Conceptualization, F.Y., S.H. and M.B.; methodology, F.Y., S.H. and M.B.; programming, F.Y. and S.H.; validation, F.Y. and S.H.; data collection, F.Y. and S.H.; writing–original draft preparation, F.Y., S.H., S.S., A.Y. and M.B.; writing-review and edit, M.B.; supervision and project administration, M.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Sciences and Engineering Research Council of Canada (NSERC) through Dr. Bolic’s NSERC Alliance COVID-19 grant.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Also, the consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to data sharing is not included in initial project proposal.

Acknowledgments

We acknowledge the support of the Natural Sciences and Engineering Research Council of Canada (NSERC). We would like to thanks J & M Group Inc., Toronto, Canada for their support. We had a number of productive discussions and meetings with Magesh Raghavan and Harappriyan Ulaganathan and we are thankful for their support and guidance. We would like to thank SOSCIP Consortium, Ontario, Canada for providing computational resources used in this research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. WHO. Coronavirus Disease (COVID-19). [Online]. 10 November 2020. Available online: https://www.who.int/health-topics/coronavirus (accessed on 6 April 2021).
  2. Li, X.; Chen, J.; Zhao, G.; Pietikäinen, M. Remote Heart Rate Measurement from Face Videos under Realistic Situations. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 4264–4271. [Google Scholar] [CrossRef]
  3. Cho, Y.; Julier, S.J.; Marquardt, N. Robust tracking of respiratory rate in high-dynamic range scenes using mobile thermal imaging. Biomed. Opt. Express 2017, 8, 4480–4503. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Mutlu, K.; Rabell, J.E.; del Olmo, P.M.; Haesler, S. IR thermography-based monitoring of respiration phase without image segmentation. J. Neurosci. Methods 2018, 301, 1–8. [Google Scholar] [CrossRef] [PubMed]
  5. Weenk, M.; van Goor, H.; Frietman, B.; Engelen, L.; van Laarhoven, C.J.; Smit, J.; Bredie, S.J.; van de Belt, T. Continuous monitoring of vital signs using wearable devices on the general ward: Pilot study. JMIRmHealth uHealth 2017, 5, 1–12. [Google Scholar] [CrossRef]
  6. Verkruysse, W.; Svaasand, L.O.; Nelson, J.S. Remote plethysmographic imaging using ambient light. Opt. Express 2008, 16, 21434–21445. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Poh, M.Z.; McDuff, D.J.; Picard, R.W. Non-contact, automated cardiac pulse measurements using video imaging and blind source separation. Opt. Express 2010, 18, 10762–10774. [Google Scholar] [CrossRef]
  8. Zhao, F.; Li, M.; Qian, Y.; Tsien, J.Z. Remote measurements of heart and respiration rates for telemedicine. PLoS ONE 2013, 8, e71384. [Google Scholar] [CrossRef] [Green Version]
  9. Shao, D.; Yang, Y.; Liu, C.; Tsow, F.; Yu, H. Noncontact monitoring breathing pattern, exhalation flow rate and pulse transit time. IEEE Trans. Biomed. Eng. 2014, 61, 2760–2767. [Google Scholar] [CrossRef] [PubMed]
  10. Reyes, B.A.; Reljin, N.; Kong, Y.; Nam, Y. Tidal volume and instantaneous respiration rate estimation using a volumetric surrogate signal acquired via a smartphone camera. IEEE J. Biomed. Health Inform. 2017, 21, 764–777. [Google Scholar] [CrossRef] [PubMed]
  11. Muhammad, U.; Evans, R.; Saatchi, R.; Kingshott, R. P016 Using non-invasive thermal imaging for apnoea detection. BMJ Open Respir. Res. 2019, 6. [Google Scholar] [CrossRef]
  12. Scebba, G.; da Poian, G.; Karlen, W. Multispectral Video Fusion for Non-Contact Monitoring of Respiratory Rate and Apnea. IEEE Trans. Biomed. Eng. 2021, 68, 350–359. [Google Scholar] [CrossRef] [PubMed]
  13. Viola, P.; Jones, M. Rapid object detection using a boosted cascade of simple features. In Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2001, Kauai, HI, USA, 8–14 December 2001; p. I. [Google Scholar] [CrossRef]
  14. Asthana, A.; Zafeiriou, S.; Cheng, S.; Pantic, M. Robust Discriminative Response Map Fitting with Constrained Local Models. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 3444–3451. [Google Scholar]
  15. Zhang, K.; Zhang, Z.; Li, Z.; Qiao, Y. Joint Face Detection and Alignment Using Multitask Cascaded Convolutional Networks. IEEE Signal Process. Lett. 2016, 23, 1499–1503. [Google Scholar] [CrossRef] [Green Version]
  16. Zhang, S.; Zhu, X.; Lei, Z.; Shi, H.; Wang, X.; Li, S.Z. S3FD: Single Shot Scale-Invariant Face Detector. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 192–201. [Google Scholar]
  17. Tang, X.; Du, D.K.; He, Z.; Liu, J. PyramidBox: A Context-assisted Single Shot Face Detector. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  18. Jiang, Z.; Hu, M.; Gao, Z.; Fan, L.; Dai, R.; Pan, Y.; Tang, W.; Zhai, G.; Lu, Y. Detection of Respiratory Infections Using RGB-Infrared Sensors on Portable Device. IEEE Sens. J. 2020, 20, 13674–13681. [Google Scholar] [CrossRef]
  19. Kazemi, V.; Sullivan, J. One millisecond face alignment with an ensemble of regression trees. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 1867–1874. [Google Scholar]
  20. Deng, J.; Guo, J.; Ververas, E.; Kotsia, I.; Zafeiriou, S. RetinaFace: Single-Shot Multi-Level Face Localisation in the Wild. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 5202–5211. [Google Scholar] [CrossRef]
  21. Yang, S.; Luo, P.; Loy, C.C.; Tang, X. WIDER FACE: A Face Detection Benchmark. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 5525–5533. [Google Scholar]
  22. Rouast, P.V.; Adam, M.T.; Chiong, R.; Cornforth, D.; Lux, E. Remote heart rate measurement using low-cost RGB face video: A technical literature review. Front. Comput. Sci. 2018, 12, 858–872. [Google Scholar] [CrossRef]
  23. Lewandowska, M.; Rumiński, J.; Kocejko, T.; Nowak, J. Measuring pulse rate with a webcam—A non-contact method for evaluating cardiac activity. In Proceedings of the 2011 IEEE Federated Conference on Computer Science and Information Systems (FedCSIS), Szczecin, Poland, 18–21 September 2011; pp. 405–410. [Google Scholar]
  24. Kwon, S.; Kim, H.; Park, K.S. Validation of heart rate extraction using video imaging on a built-in camera system of a smartphone. In Proceedings of the 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, San Diego, CA, USA, 28 August–1 September 2012; pp. 2174–2177. [Google Scholar]
  25. Hsu, Y.; Lin, Y.L.; Hsu, W. Learning-based heart rate detection from remote photoplethysmography features. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 4433–4437. [Google Scholar]
  26. Holton, B.D.; Mannapperuma, K.; Lesniewski, P.J.; Thomas, J.C. Signal recovery in imaging photoplethysmography. Physiol. Meas. 2013, 34, 1499. [Google Scholar] [CrossRef] [PubMed]
  27. Feng, L.; Po, L.M.; Xu, X.; Li, Y. Motion artifacts suppression for remote imaging photoplethysmography. In Proceedings of the 2014 19th International Conference on Digital Signal Processing, Hong Kong, China, 20–23 August 2014; pp. 18–23. [Google Scholar]
  28. Cardoso, J.F. High-order contrasts for independent component analysis. Neural Comput. 1999, 11, 157–192. [Google Scholar] [CrossRef] [PubMed]
  29. Ali, M. Reza. Realization of the Contrast Limited Adaptive Histogram Equalization (CLAHE) for Real-Time Image Enhancement. VLSI Signal Process. 2004, 38, 35–44. [Google Scholar]
Figure 1. The illustration of vital signs measurement experiment setup.
Figure 1. The illustration of vital signs measurement experiment setup.
Sensors 22 00627 g001
Figure 2. The workflow of vital signs estimation using RGB and thermal cameras.
Figure 2. The workflow of vital signs estimation using RGB and thermal cameras.
Sensors 22 00627 g002
Figure 3. An example of face detection using different detectors (a). No face detected by MTCNN face detector, (b) face (blue box) detected by S3FD face detector, (c) face (blue box) detected by PyramidBox face detector, (d) face (blue box) and facial landmarks (colored dots) detected by RetinaFace detector.
Figure 3. An example of face detection using different detectors (a). No face detected by MTCNN face detector, (b) face (blue box) detected by S3FD face detector, (c) face (blue box) detected by PyramidBox face detector, (d) face (blue box) and facial landmarks (colored dots) detected by RetinaFace detector.
Sensors 22 00627 g003
Figure 4. An example of multiple subjects’ faces (red bounding boxes) and facial landmarks (colored dots) detected by RetinaFace detector when they were wearing face masks.
Figure 4. An example of multiple subjects’ faces (red bounding boxes) and facial landmarks (colored dots) detected by RetinaFace detector when they were wearing face masks.
Sensors 22 00627 g004
Figure 5. ROIs employed for vital signs estimation, (a). the location of R O I H R on the RGB image, (b). The locations of R O I R R and R O I B T on the simultaneous thermal image.
Figure 5. ROIs employed for vital signs estimation, (a). the location of R O I H R on the RGB image, (b). The locations of R O I R R and R O I B T on the simultaneous thermal image.
Sensors 22 00627 g005
Figure 6. A demonstration of ROIs localization using facial landmarks, (a) the definitions of R O I B T and R O I R R , (b) the definition of R O I H R .
Figure 6. A demonstration of ROIs localization using facial landmarks, (a) the definitions of R O I B T and R O I R R , (b) the definition of R O I H R .
Sensors 22 00627 g006
Figure 7. (a) Facial landmarks and partitions when the subject is looking at the camera. (b) Scenario when the subject is not looking at the camera and the ratio F a c e d e f l e c t i o n reaches 0.17.
Figure 7. (a) Facial landmarks and partitions when the subject is looking at the camera. (b) Scenario when the subject is not looking at the camera and the ratio F a c e d e f l e c t i o n reaches 0.17.
Sensors 22 00627 g007
Figure 8. Demonstration of RGB and thermal image alignment, (a) original RGB image (640 × 480), (b) cropped RGB image (256 × 192), (c) aligned thermal image (256 × 192).
Figure 8. Demonstration of RGB and thermal image alignment, (a) original RGB image (640 × 480), (b) cropped RGB image (256 × 192), (c) aligned thermal image (256 × 192).
Sensors 22 00627 g008
Figure 9. Methods developed for remote HR estimation.
Figure 9. Methods developed for remote HR estimation.
Sensors 22 00627 g009
Figure 10. (a) Respiration waveform extracted from the thermal camera (red) and the reference measurement using a breath belt (blue), (b) BVP waveform extracted the RGB camera (red) and reference PPG waveform (blue).
Figure 10. (a) Respiration waveform extracted from the thermal camera (red) and the reference measurement using a breath belt (blue), (b) BVP waveform extracted the RGB camera (red) and reference PPG waveform (blue).
Sensors 22 00627 g010
Figure 11. (a) Original thermal image with R O I R R detected (red bounding box), (b) the contrast of the thermal image enhanced by CLAHE.
Figure 11. (a) Original thermal image with R O I R R detected (red bounding box), (b) the contrast of the thermal image enhanced by CLAHE.
Sensors 22 00627 g011
Figure 12. Demonstration of signal quality assessment, (a) A 15s rBVP waveform and detected peaks using proposed approaches, (b) rBVP pulses (gray curves) and the averaged rBVP template (red curves), SQI = 0.94.
Figure 12. Demonstration of signal quality assessment, (a) A 15s rBVP waveform and detected peaks using proposed approaches, (b) rBVP pulses (gray curves) and the averaged rBVP template (red curves), SQI = 0.94.
Sensors 22 00627 g012
Figure 13. Respiratory signals extracted from thermal video (red) and from a reference breath belt (blue) of a subject at 75 cm with different face masks, (a) No face mask, (b) medical face mask, (c) N90 face mask, and (d) N95 face mask.
Figure 13. Respiratory signals extracted from thermal video (red) and from a reference breath belt (blue) of a subject at 75 cm with different face masks, (a) No face mask, (b) medical face mask, (c) N90 face mask, and (d) N95 face mask.
Sensors 22 00627 g013
Figure 14. (a) Correlation plot of the estimated RR and reference RR for experiments when subjects were not wearing face masks, (b) Bland–Altman plot of the estimated RR and reference RR for experiments when subjects were not wearing face masks, (c) Correlation plot of the estimated RR and reference RR for experiments when subjects were wearing face masks, (d) Bland–Altman plot of the estimated RR and reference RR for experiments when subjects were wearing face masks.
Figure 14. (a) Correlation plot of the estimated RR and reference RR for experiments when subjects were not wearing face masks, (b) Bland–Altman plot of the estimated RR and reference RR for experiments when subjects were not wearing face masks, (c) Correlation plot of the estimated RR and reference RR for experiments when subjects were wearing face masks, (d) Bland–Altman plot of the estimated RR and reference RR for experiments when subjects were wearing face masks.
Sensors 22 00627 g014
Figure 15. (a) Correlation plot of the estimated HR and reference HR for subjects who have a pale skin tone, (b) Bland–Altman plot of the estimated HR and reference HR for subjects who have a pale skin tone, (c) Correlation plot of the estimated HR and reference HR for subjects who have medium or dark skin tones, (d) Bland–Altman plot of the estimated HR and reference HR for subjects who have medium or dark skin tones.
Figure 15. (a) Correlation plot of the estimated HR and reference HR for subjects who have a pale skin tone, (b) Bland–Altman plot of the estimated HR and reference HR for subjects who have a pale skin tone, (c) Correlation plot of the estimated HR and reference HR for subjects who have medium or dark skin tones, (d) Bland–Altman plot of the estimated HR and reference HR for subjects who have medium or dark skin tones.
Sensors 22 00627 g015
Table 1. The mean absolute error between the estimated RR and the reference RR of a subject with different face masks at different detection ranges (unit: breaths/min).
Table 1. The mean absolute error between the estimated RR and the reference RR of a subject with different face masks at different detection ranges (unit: breaths/min).
75 cm120 cm200 cm
No mask0.410.280.80
Mask1 (medical/cloth)0.520.450.59
Mask2 (N90)0.060.010.17
Mask3 (N95)0.060.420.88
Table 2. Mean absolute error between the estimated RR and the reference RR (unit: breaths/min). Boldface character denotes the best result.
Table 2. Mean absolute error between the estimated RR and the reference RR (unit: breaths/min). Boldface character denotes the best result.
60 cm80 cm100 cm120 cm
No MaskMaskNo MaskMaskNo MaskMaskNo MaskMask
Subject11.830.502.041.602.730.742.080.74
Subject22.680.402.240.362.490.231.250.52
Subject31.840.461.720.561.690.572.010.18
Subject41.702.302.061.652.311.692.091.54
Subject51.041.651.991.752.100.881.261.15
Subject61.942.222.581.402.451.043.102.16
Subject71.820.592.180.490.650.601.140.78
Subject81.651.241.350.642.171.011.381.14
Subject91.181.670.990.961.121.071.241.42
Subject101.712.472.301.802.852.461.561.70
Table 3. Mean and standard deviation of the absolute error between the estimated HR and the reference HR (unit: beats/min). Boldface character denotes the best result.
Table 3. Mean and standard deviation of the absolute error between the estimated HR and the reference HR (unit: beats/min). Boldface character denotes the best result.
60 cm80 cm100 cm120 cm
Skin Tone M eHR SD eHR M eHR SD eHR M eHR SD eHR M eHR SD eHR
Subject1pale2.921.213.021.114.311.572.931.31
Subject2pale3.582.594.083.263.823.054.834.00
Subject3pale2.612.424.282.863.893.014.502.88
Subject4pale3.352.312.171.822.881.822.501.23
Subject5pale1.733.212.621.441.811.762.031.65
Subject6pale1.501.781.051.431.862.291.763.02
Subject7medium1.681.552.612.903.052.553.092.69
Subject8medium2.531.533.073.052.352.593.042.89
Subject9dark2.721.753.831.883.312.442.211.56
Subject10dark3.382.152.201.862.201.502.091.88
Table 4. Mean and standard deviation of the absolute error of two-subjects RR and HR estimations (RR unit: breaths/min, HR unit: beats/min).
Table 4. Mean and standard deviation of the absolute error of two-subjects RR and HR estimations (RR unit: breaths/min, HR unit: beats/min).
Subject1Subject2
RRHRRRHR
Experiment 10.89 ± 0.473.60 ± 2.101.31 ± 0.861.72 ± 1.40
Experiment 21.25 ± 0.632.17 ± 2.200.80 ± 0.572.41 ± 1.13
Experiment 30.49 ± 0.331.79 ± 1.061.07 ± 0.581.35 ± 1.12
Experiment 4 (mask)1.57 ± 0.962.03 ± 1.180.74 ± 0.522.39 ± 1.67
Experiment 5 (mask)1.60 ± 0.582.15 ± 2.070.49 ± 0.322.31 ± 1.09
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, F.; He, S.; Sadanand, S.; Yusuf, A.; Bolic, M. Contactless Measurement of Vital Signs Using Thermal and RGB Cameras: A Study of COVID 19-Related Health Monitoring. Sensors 2022, 22, 627. https://doi.org/10.3390/s22020627

AMA Style

Yang F, He S, Sadanand S, Yusuf A, Bolic M. Contactless Measurement of Vital Signs Using Thermal and RGB Cameras: A Study of COVID 19-Related Health Monitoring. Sensors. 2022; 22(2):627. https://doi.org/10.3390/s22020627

Chicago/Turabian Style

Yang, Fan, Shan He, Siddharth Sadanand, Aroon Yusuf, and Miodrag Bolic. 2022. "Contactless Measurement of Vital Signs Using Thermal and RGB Cameras: A Study of COVID 19-Related Health Monitoring" Sensors 22, no. 2: 627. https://doi.org/10.3390/s22020627

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop