By: 23 July 2019
Technology allows researchers to see patients’ real-time pain while in the clinic

Many patients, especially those who are anaesthetised or emotionally challenged, cannot communicate precisely about their pain.

For this reason, University of Michigan researchers have developed a technology to help clinicians “see” and map patient pain in real-time, through special augmented reality glasses. Their small feasibility study appears in the Journal of Medical Internet Research.

The technology was tested on 21 volunteer dental patients, and researchers hope to one day include other types of pain and conditions. It’s years away from widespread use in a clinical setting, but the feasibility study is a good first step for dental patients, said Alex DaSilva, associate professor at the U-M School of Dentistry and director of the Headache and Orofacial Pain Effort Lab.

The portable CLARAi (clinical augmented reality and artificial intelligence) platform combines visualisation with brain data using neuroimaging to navigate through a patient’s brain while they’re in the chair.

“It’s very hard for us to measure and express our pain, including its expectation and associated anxiety,” DaSilva said. “Right now, we have a one to 10 rating system, but that’s far from a reliable and objective pain measurement.”

In the study, researchers triggered pain by administering cold to the teeth. Researchers used brain pain data to develop algorithms that, when coupled with new software and neuroimaging hardware, predicted pain or the absence of it about 70 per cent of the time.

Participants wore a sensor-outfitted cap that detected changes to blood flow and oxygenation, thus measuring brain activity and responses to pain. That information was transmitted to a computer and interpreted.

Wearing special augmented reality glasses (in this case, the Microsoft HoloLens), researchers viewed the subject’s brain activity in real time on a reconstructed brain template, while the subjects sat in the clinical chair. The red and blue dots on the image denote location and level of brain activity, and this “pain signature” was mirror-displayed on the augmented reality screen. The more pain signatures the algorithm learns to read, the more accurate the pain assessment.

Co-authors include: Xiao-Su Hu, Andrew Racek, Thiago Nascimento, Mary Bender, Theodore Hall, Sean Petty, Stephanie O’Malley and Niko Kaciroti of U-M; Roger Ellwood of Colgate Palmolive; and Eric Maslowski of Ann Arbor-based Moxytech Inc.

Source: University of Michigan

Image: Hassan Jassar (seated) wears a sensor-outfitted cap that detects changes in blood flow and oxygenation, thus sensing brain activity. That information is transmitted to a computer and interpreted. Thiago Nascimento, left, views this brain activity in real-time while wearing augmented reality glasses, and the computer image shows that particular pain signature in the brain. From left to right, also pictured, Dr. Alex DaSilva, Dajung Kim, Manyoel Lim, Xiao-su Hu. Credit: University of Michigan