A UW-led team has developed a method that uses the camera on a person’s smartphone or computer to take their pulse and breathing rate from a real-time video of their face.
Cristina Zaragoza/Unsplash
A University of Washington-led team has developed a method that uses the camera on a person’s smartphone or computer to take their pulse and respiration signal from a real-time video of their face.
The researchers presented this state-of-the-art system in December at the Neural Information Processing Systems conference.
Now the team is proposing a better system to measure these physiological signals. This system is less likely to be tripped up by different cameras, lighting conditions or facial features, such as skin color. The researchers will present these findings April 8 at the ACM Conference on Health, Interference, and Learning.
“Machine learning is pretty good at classifying images. If you give it a series of photos of cats and then tell it to find cats in other images, it can do it. But for machine learning to be helpful in remote health sensing, we need a system that can identify the region of interest in a video that holds the strongest source of physiological information — pulse, for example — and then measure that over time,” said lead author Xin Liu, a UW doctoral student in the Paul G. Allen School of Computer Science & Engineering.
“Every person is different,” Liu said. “So this system needs to be able to quickly adapt to each person’s unique physiological signature, and separate this from other variations, such as what they look like and what environment they are in.”
Try the researchers’ demo version that can detect a user’s heartbeat over time, which doctors can use to calculate heart rate.
The team’s system is privacy preserving — it runs on the device instead of in the cloud — and uses machine learning to capture subtle changes in how light reflects off a person’s face, which is correlated with changing blood flow. Then it converts these changes into both pulse and respiration rate.
The first version of this system was trained with a dataset that contained both videos of people’s faces and “ground truth” information: each person’s pulse and respiration rate measured by standard instruments in the field. The system then used spatial and temporal information from the videos to calculate both vital signs. It outperformed similar machine learning systems on videos where subjects were moving and talking.
But while the system worked well on some datasets, it still struggled with others that contained different people, backgrounds and lighting. This is a common problem known as “overfitting,” the team said.
The researchers improved the system by having it produce a personalized machine learning model for each individual. Specifically, it helps look for important areas in a video frame that likely contain physiological features correlated with changing blood flow in a face under different contexts, such as different skin tones, lighting conditions and environments. From there, it can focus on that area and measure the pulse and respiration rate.
While this new system outperforms its predecessor when given more challenging datasets, especially for people with darker skin tones, there’s still more work to do, the team said.
“We acknowledge that there is still a trend toward inferior performance when the subject’s skin type is darker,” Liu said. “This is in part because light reflects differently off of darker skin, resulting in a weaker signal for the camera to pick up. Our team is actively developing new methods to solve this limitation.”
The researchers are also working on a variety of collaborations with doctors to see how this system performs in the clinic.
“Any ability to sense pulse or respiration rate remotely provides new opportunities for remote patient care and telemedicine. This could include self-care, follow-up care or triage, especially when someone doesn’t have convenient access to a clinic,” said senior author Shwetak Patel, a professor in both the Allen School and the electrical and computer engineering department. “It’s exciting to see academic communities working on new algorithmic approaches to address this with devices that people have in their homes.”
Original Article: New system that uses smartphone or computer cameras to measure pulse, respiration rate could help future personalized telehealth appointments
More from: University of Washington | Microsoft Research
The Latest Updates from Bing News & Google News
Go deeper with Bing News on:
Smartphone health check
- 'Heat Risk' feature added to MCHD’s smartphone app
As the summer season approaches, those within our community must take steps to protect themselves from high heat exposure.Heat-related illnesses can occur when ...
- 10 Huge Mistakes That Could Be Damaging Your Smartphone
Smartphones are ubiquitous in today's world, and they require a bit of care to keep healthy. Here are 10 mistakes that could be damaging your device.
- How to check battery health of your smartphone?
Battery health v/s battery life. A phone's battery health starts at a maximum of 100% when new and unused, but this percentage decreases over time due to continuous usage, charge ...
- Two-second test to see if you have 'smartphone finger' - and how to fix it
Where would we be these days without our smartphones? While they are undoubtedly useful, did you know that using them to excess can actually cause a telltale physical sign? Here's all you need to know ...
- The best phone 2024: top smartphones in the US right now
The best phone you can buy might be an Apple iPhone or an Android phone from Samsung, Google, or OnePlus, but you can't go wrong with any phone on this list. We've spent quality time reviewing ...
Go deeper with Google Headlines on:
Smartphone health check
[google_news title=”” keyword=”smartphone health check” num_posts=”5″ blurb_length=”0″ show_thumb=”left”]
Go deeper with Bing News on:
Remote health sensing
- Applications of Remote Sensing Over Plateau Mountainous Areas
Plateau mountainous areas occupy about one fifth of the Earth’s surface, they are home to approximately one tenth of the global population, and provide goods and services to about half of humanity.
- Aston University scientist to help make crop monitoring easier and cheaper
An Aston University photonics expert has received a Royal Society Industry Fellowship grant to help make crop monitoring easier and cheaper with remote sensing technology. Dr Sergey ...
- Remote Sensing and Satellite Oceanography
The group focuses on observing and modeling coastal processes including beach evolution, cliff erosion, and nearshore waves. Provide water cycle science, technology and outreach to support effective ...
- Python for Modern GIS and Remote Sensing
remote sensing, and spatial data analysis) to solve the age-old, yet complex problems of natural resource management, land use/cover, invasive species, and forest health. He is not only teaching this ...
- Thermal Infrared Remote Sensing At Yellowstone 301: Thermal Metrics
Calculating important thermal metrics at Yellowstone National Park that can be used for monitoring change over time.
Go deeper with Google Headlines on:
Remote health sensing
[google_news title=”” keyword=”remote health sensing” num_posts=”5″ blurb_length=”0″ show_thumb=”left”]