Abstract
Remote photoplethysmography (rPPG) is a technique for non-contact estimation of human vital signs in video. It enriches knowledge about human state and makes interpretations of actions in human-computer interaction more accurate. An approach to distributed collection of rPPG dataset is proposed along with a central hub where data is accumulated as links to local storages hosted by participating organizations. An instrument for rPPG data collection is developed and described. It is an Android application, which captures dual camera video from front and rear cameras simultaneously. The front camera captures facial video while the rear camera with flash turned on captures a contact finger video. Facial videos constitute a dataset, while ground truth blood volume pulse (BVP) characteristics can be obtained by the analysis of correspondent finger videos. Such approach allows overcoming organizational and technical limitations of biometric data collection and hosting.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
American National Standards Institute and Association for the Advancement of Medical Instrumentation. Cardiac monitors, heart rate meters, and alarms. Association for the Advancement of Medical Instrumentation, Arlington, Va (2002)
Antink, C.H., Gao, H., Br, C., Leonhardt, S.: Beat-to-beat heart rate estimation fusing multimodal video and sensor data. Biomed. Opt. Express 6(8), 2895–2907 (2015)
Antink, C.H., Lyra, S., Paul, M., Yu, X., Leonhardt, S.: A broader look: camera-based vital sign estimation across the spectrum. Yearb. Med. Inform. 28(1), 102–114 (2019)
Artemyev, M., Churikova, M., Grinenko, M., Perepelkina, O.: Neurodata lab’s approach to the challenge on computer vision for physiological measurement. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (2020)
Blackford, E.B., Estepp, J.R.: Measurements of pulse rate using long-range imaging photoplethysmography and sunlight illumination outdoors. In: Coté, G.L. (ed.) Optical Diagnostics and Sensing XVII: Toward Point-of-Care Diagnostics, vol. 10072, pp. 122–134. SPIE, San Francisco, California, United States (2017)
Blackford, E.B., Estepp, J.R., Piasecki, A.M., Bowers, M.A., Samantha, L.: Long-range non-contact imaging photoplethysmography: cardiac pulse wave sensing at a distance. In: Optical Diagnostics and Sensing XVI: Toward Point-of-Care Diagnostics, vol. 9715, pp. 176–192 (2016)
Blöcher, T., Krause, S., Zhou, K., Zeilfelder, J., Stork, W.: VitalCamSet - a dataset for Photoplethysmography Imaging. In: 2019 IEEE Sensors Applications Symposium (SAS), pp. 1–6 (2019)
Bobbia, S., Macwan, R., Benezeth, Y., Mansouri, A., Dubois, J.: Unsupervised skin tissue segmentation for remote photoplethysmography. Pattern Recogn. Lett. 124, 82–90 (2019)
Chen, W., Hernandez, J., Picard, R.W.: Estimating carotid pulse and breathing rate from near-infrared video of the neck. Physiol. Meas. 39(10), 10NT01 (2018)
Chen, W., Picard, R.W.: Eliminating physiological information from facial videos. In: 2017 12th IEEE International Conference on Automatic Face Gesture Recognition (FG 2017), pp. 48–55 (2017)
Coppetti, T.: Accuracy of smartphone apps for heart rate measurement. Eur. J. Prev. Cardiol. 24(12), 1287–1293 (2017)
Estepp, J.R., Blackford, E.B., Meier, C.M.: Recovering pulse rate during motion artifact with a multi-imager array for non-contact imaging photoplethysmography. In: 2014 IEEE International Conference on Systems, Man, and Cybernetics (SMC), pp. 1462–1469 (2014)
Ghanadian, H., Al Osman, H.: Non-contact heart rate monitoring using multiple RGB cameras. In: Vento, M., Percannella, G. (eds.) Computer Analysis of Images and Patterns CAIP 2019. Lecture Notes in Computer Science, vol. 11679, pp. 85–95. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-29891-3_8
De Haan, G., Jeanne, V.: Robust pulse-rate from chrominance-based rPPG. IEEE Trans. Biomed. Eng. 6600(10), 1–9 (2013)
Han, B., Ivanov, K., Wang, L., Yan, Y.: Exploration of the optimal skin-camera distance for facial photoplethysmographic imaging measurement using cameras of different types. In: Proceedings of the 5th EAI International Conference on Wireless Mobile Communication and Healthcare, pp. 186–189. ICST: Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering, London, Great Britain (2015)
Heusch, G., Anjos, A., Marcel, S.: A Reproducible Study on Remote Heart Rate Measurement. arXiv preprint, arXiv:1709 (2017)
Hoffman, W.F.C., Lakens, D.: Public Benchmark Dataset for Testing rPPG Algorithm Performance. Technical report (2019)
Hsu, G.-S., Chen, M.-S.: Deep learning with time-frequency representation for pulse estimation from facial videos. In: 2017 IEEE International Joint Conference on Biometrics (IJCB), pp. 383–389 (2017)
Jang, D.-G., Park, S., Hahn, M., Park, S.-H.: A real-time pulse peak detection algorithm for the photoplethysmogram. Int. J. Electron. Electr. Eng. 2(1), 45–49 (2014)
Kamshilin, A.A., et al.: A new look at the essence of the imaging photoplethysmography. Sci. Rep. 10494:1–10494:9 (2015)
Karray, F., Alemzadeh, M., Saleh, J., Arab, M.N.: Human-computer interaction: overview on state of the art. Int. J. Smart Sens. Intell. Syst. 1, 137–159 (2008)
Khanam, F.-T.-Z., Al-naji, A.A., Chahl, J.: Remote monitoring of vital signs in diverse non-clinical and clinical scenarios using computer vision systems: a review. Appl. Sci. 9(20), 4474 (2019)
Koelstra, S., et al.: DEAP: a database for emotion analysis; using physiological signals. IEEE Trans. Affect. Comput. 3(1), 18–31 (2012)
Kopeliovich, M., Kalinin, K., Mironenko, Y., Petrushan, M.: On indirect assessment of heart rate in video. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, pp. 1260–1264 (2020)
Kopeliovich, M., Petrushan, M.: Color signal processing methods for webcam-based heart rate evaluation. In: Bi, Y., Bhatia, R., Kapoor, S. (eds.) Intelligent Systems and Applications. IntelliSys 2019. Advances in Intelligent Systems and Computing, pp. 1038:703–723 (2019)
Kranjec, J., Begus, S., Gersak, G., Drnovsek, J.: Non-contact heart rate and heart rate variability measurements: a review. Biomed. Signal Process. Control 13(July), 102–112 (2014)
Lasinger, K., Ranftl, R., Schindler, K., Koltun, V.: Towards robust monocular depth estimation: mixing datasets for zero-shot cross-dataset transfer. CoRR, ArXiv:1907.01341 (2019)
Lee, D., Kim, J., Kwon, S., Park, K.: Heart rate estimation from facial photoplethysmography during dynamic illuminance changes. IEEE Eng. Med. Biol. Soc. 2758–2761 (2015). https://doi.org/10.1109/EMBC.2015.7318963
Li, P., et al.: Video-based pulse rate variability measurement using periodic variance maximization and adaptive two-window peak detection. Sensors (Switzerland) 20(10) (2020)
Li, X., et al.: The OBF database: a large face video database for remote physiological signal measurement and atrial fibrillation detection. In: 2018 13th IEEE International Conference on Automatic Face Gesture Recognition (FG 2018), pp. 242–249 (2018)
Luguern, D., et al.: An assessment of algorithms to estimate respiratory rate from the electrocardiogram and photoplethysmogram. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (2020)
Matsumura, K., Rolfe, P., Yamakoshi, T.: iPhysioMeter: a smartphone photoplethysmograph for measuring various physiological indices. In: Rasooly, A., Herold, K.E. (eds.) Mobile Health Technologies. Methods and Protocols, vol. 1256, pp. 305–326. Humana Press, New York, New York, NY (2015). https://doi.org/10.1007/978-1-4939-2172-0_21
Mironenko, Y., Kalinin, K., Kopeliovich, M., Petrushan, M.: Remote photoplethysmography: rarely considered factors. In: 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, pp. 1197–1206 (2020)
Moço, A., Verkruysse, W.: Pulse oximetry based on photoplethysmography imaging with red and green light: calibratability and challenges. J. Clin. Monit. Comput. 35(1), 123–133 (2020). https://doi.org/10.1007/s10877-019-00449-y
Monfort, A., et al.: Moments in Time Dataset: one million videos for event understanding. IEEE Trans. Pattern Anal. Mach. Intell. 1–8 (2019)
Niu, X., Han, H., Shan, S., Chen, X.: VIPL-HR: a multi-modal database for pulse estimation from less-constrained face video. Jawahar, C., Li, H., Mori, G., Schindler, K. (eds.) Computer Vision – ACCV 2018. ACCV 2018. Lecture Notes in Computer Science, vol. 11365, pp. 1–16 (2018). https://doi.org/10.1007/978-3-030-20873-8_36
Niu, X., Shan, S., Han, H., Chen, H.: RhythmNet: end-to-end heart rate estimation from face via spatial-temporal representation. IEEE Trans. Image Process. 29, 2409–2423 (2020)
Nowara, E.M., Mcduff, D., Veeraraghavan, A.: A Meta-Analysis of the Impact of Skin Type and Gender on Non-contact Photoplethysmography Measurements. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (2020)
Nowara, E.M., Marks, T.K., Mansour, H., Veeraraghavan, A.: SparsePPG: towards driver monitoring using camera-based vital signs estimation in near-infrared. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1353–135309 (2018)
Perepelkina, O., Artemyev, M., Churikova, M., Grinenko, M.: HeartTrack: convolutional neural network for remote video-based heart rate monitoring. In: The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (2020)
Pilz, C.S., Zaunseder, S., Blazek, V.: Local group invariance for heart rate estimation from face videos in the wild. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1335–13358 (2018)
Rouast, P.V., Adam, M.T.P., Chiong, R., et al.: Remote heart rate measurement using low-cost RGB face video: a technical literature review. Front. Comput. Sci. 12, 858–872 (2018). https://doi.org/10.1007/s11704-016-6243-6
Russakovsky, O., et al.: ImageNet large scale visual recognition challenge. Int. J. Comput. Vis. (IJCV) 115(3), 211–252 (2015). https://doi.org/10.1007/s11263-015-0816-y
Sabokrou, M., Pourreza, M., Li, X., Fathy, M., Zhao, G.: Deep-HR: Fast Heart Rate Estimation from Face Video Under Realistic Conditions. ArXiv:abs/2002.04821 (2020)
Sinhal, R., Singh, K., Raghuwanshi, M.M.: An overview of remote photoplethysmography methods for vital sign monitoring. In: Gupta, M., Konar, D., Bhattacharyya, S., Biswas, S. (eds.) Computer Vision and Machine Intelligence in Medical Image Analysis. Advances in Intelligent Systems and Computing, vol. 992, pp. 21–31. Springer, Singapore (2020). https://doi.org/10.1007/978-981-13-8798-2_3
Soleymani, M., Lichtenauer, J., Pun, T., Pantic, M.: A multimodal database for affect recognition and implicit tagging. IEEE Trans. Affect. Comput. 3(1), 42–55 (2012)
Song, R., Chen, H., Cheng, J., Li, Y., Liu, C., Chen, X.: PulseGAN: learning to generate realistic pulse waveforms in remote photoplethysmography. Image Video Process. 25(5), 1373–1384 (2020)
Song, R., Zhang, S., Cheng, J., Li, C., Chen, X.: New insights on super-high resolution for video-based heart rate estimation with a semi-blind source separation method. Comput. Biol. Med. 116, 103535 (2020). https://doi.org/10.1016/j.compbiomed.2019.103535
Špetlík, R., Cech, J.: Visual heart rate estimation with convolutional neural network. In: British Machine Vision Conference (2018)
Stricker, R., Steffen, M., Gross, H.: Non-contact video-based pulse rate measurement on a mobile service robot. In: The 23rd IEEE International Symposium on Robot and Human Interactive Communication, pp. 1056–1062 (2014)
Sun, Y., Papin, C., Azorin-Peris, V., Kalawsky, R., Greenwald, S., Sijung, H.: Use of ambient light in remote photoplethysmographic systems: comparison between a high-performance camera and a low-cost webcam. J. Biomed. Opt. 17(3), 037005 (2012)
Tang, C., Lu, J., Liu, J.: Non-contact heart rate monitoring by combining convolutional neural network skin detection and remote photoplethysmography via a low-cost camera. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, pp. 1390–13906 (2018)
Tayfur, I., Afacan, M.A.: Reliability of smartphone measurements of vital parameters: a prospective study using a reference method. Am. J. Emerg. Med. 37(8), 1527–1530 (2019)
Tsai, Y.C., Lai, P.W., Huang, P.W., Lin, T.M., Wu, B.F.: Vision-based instant measurement system for driver fatigue monitoring. IEEE Access 8, 67342–67353 (2020)
Wang, W., Den Brinker, B., Stuijk, S., De Haan, G.: Algorithmic principles of remote-PPG. IEEE Trans. Biomed. Eng. 64(7), 1479–1491 (2017)
Wang, W., Shan, C.: Impact of makeup on remote-PPG monitoring. Biomed. Phys. Eng. Express 6, 035004 (2020)
Wang, W., Stuijk, S., De Haan, G.: A novel algorithm for remote photoplethysmography : spatial subspace rotation. IEEE Trans. Biomed. Eng. 63(9), 1974–1984 (2016)
Wang, Z., Yang, X., Cheng, K.-T.: Accurate face alignment and adaptive patch selection for heart rate estimation from videos under realistic scenarios. PLOS ONE 13, 1–25 (2018)
Wedekind, D., et al.: Assessment of blind source separation techniques for video-based cardiac pulse extraction. J. Biomed. Opt. 22(3), 035002 (2017). https://doi.org/10.1117/1.JBO.22.3.035002
Wei, B., He, X., Zhang, C., Wu, X.: Non-contact, synchronous dynamic measurement of respiratory rate and heart rate based on dual sensitive regions. BioMed. Eng. Online 16(17), 1–21 (2017). https://doi.org/10.1186/s12938-016-0300-0
Woyczyk, A., Rasche, S., Zaunseder, S.: Impact of sympathetic activation in imaging photoplethysmography. In: 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 1697–1705 (2019)
Woyczyk, S., Fleischhauer, A., Zaunseder, V.: Skin segmentation using active contours and gaussian mixture models for heart rate detection in videos. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops (2020)
Zaunseder, S., Trumpp, A., Wedekind, D., Malberg, H.: Cardiovascular assessment by imaging photoplethysmography - a review. Biomed. Eng. (Biomed. Tech.) 63(06), 617–634 (2018)
Zhan, Q., Wang, W., de Haan, G.: Analysis of CNN-based remote-PPG to understand limitations and sensitivities. Biomed. Opt. Express 11(3), 1268–1283 (2020)
Zhang, Z., et al.: Multimodal spontaneous emotion corpus for human behavior analysis. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3438–3446 (2016)
Zhao, C., Lin, C.-L., Chen, W., Li, Z.: A novel framework for remote photoplethysmography pulse extraction on compressed videos. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1380–138009 (2018)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendices
Appendix
1. Non-public rPPG Datasets
Table 2 lists rPPG-suitable datasets declared as public but have not been accessed in part or in full due to various reasons. The estimated total video duration in a dataset is based on its description in the corresponding paper.
In addition to the above, there are dozens of private datasets mentioned in one or several papers on rPPG: HNU [64], HR-D [44], BSIPL-RPPG [47], and untitled ones [2, 6, 9, 10, 12,13,14,15, 48, 51, 52, 55,56,57, 59,60,61]. The both training and testing datasets of the RePSS competition [29] are also considered private because they have been publicly available only during the competition.
2. Dual-Camera Collector
Dual-Camera Collector or DCC-Client is an open source (https://github.com/Assargadon/dcc-client) Android mobile application that records video from both cameras of a mobile device simultaneously (Fig. 2). It has, though, several key differences from a generic video recorder, even if it is able to capture the video from two cameras simultaneously.
1.1 App Features
First, data records are designed to be anonymous, keeping track of a metadata related to experiment conditions. Users have the ability to provide a desired metadata. Other metadata is determined automatically.
Second, there is no need to have a high-resolution image for a camera capturing a finger, but a flash constantly turned on would be of help. Despite of rear camera usually offers higher capture resolution than a front one, it was considered to use front camera to capture face video and rear camera to capture finger due to the following reasons:
-
Usage of a face-oriented (front) camera to record a face video makes the user able to operate the UI of the application during the recording. Therefore, a user is able to capture a record of themselves without an assistant, on their own.
-
A front camera usually has enough resolution to capture the user face at a regular distance from the device—and this is its purpose.
-
A rear camera almost always has a flashlight, which is useful to record contact BVP signals. Front cameras sometimes have flash too—but less often.
Third, the application provides two video data streams obtained simultaneously from front and rear cameras. While heart rate value is only defined in a sliding window and there is no strict requirement for per-frame synchronization, some methods and studies could be very sensitive to a time offset, for example:
-
Training of deep learning models to extract BVP signal in facial video based on asynchronous ground truth contact BVP signal obtained by rear camera.
-
Studies based on measurement of the phase shift between BVP signals from finger and face.
-
Methods of Heart Rate Variability estimation.
Fourth, the application collects the standardized data to the permanent dataset storage.
1.2 User Experience
The application determines if a device has the capability to capture videos from two cameras simultaneously—unfortunately, not all Android devices are able to do it. If no, an alert is presented to the user, which means that the device cannot be used to collect the dataset.
The application has a simple form containing fields of “User ID”, “Year of Birth” and “Gender” (Fig. 4(a)). “User ID” refers to a surrogate identifier that has the purpose of tracking the same persons, while keeping the anonymity of their personalities. For example, it may be “1” for the first subject, “2” for the second subject and so on. If another record of the first subject is performed (maybe even on a different device), User ID “1” should be used again.
After the user enters the required information, a live video preview is shown in Fig. 4(b). User can start recording by tapping the respective button in the UI. When a recording is to be finished, the user taps the button again.
The dataset entry is represented in a form of three files that are named with a unique prefix <User ID>-<Start timestamp>. This ensures that the file group of the single recording is easily recognized. Two video files contain the streams for the subject’s face and finger. Third one, “.json”, contains metadata about both subject (gender and year of birth, as explained above), device (model, anonymous device unique identifier), and record itself (start and final timestamps of the recording, resolutions of face and finger videos).
After performing one or several recordings, one can use a data cable or another file transfer mechanism to extract recordings from the device and put it to the permanent dataset storage. There is no need to avoid naming collisions because file names are generated to be unique.
Rights and permissions
Copyright information
© 2021 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Kalinin, K., Mironenko, Y., Kopeliovich, M., Petrushan, M. (2021). Towards Collecting Big Data for Remote Photoplethysmography. In: Arai, K. (eds) Intelligent Computing. Lecture Notes in Networks and Systems, vol 285. Springer, Cham. https://doi.org/10.1007/978-3-030-80129-8_6
Download citation
DOI: https://doi.org/10.1007/978-3-030-80129-8_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-80128-1
Online ISBN: 978-3-030-80129-8
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)