Paulus was part of a Kellogg team that developed a device that turns a smartphone into a functioning retinal camera.
In 2016 the project, CellScope Retina, was one of 12 funded by U-M's Translational Research and Commercialization for Life Sciences Hub, which accelerates ideas with a high potential for positively impacting human health.
The new study utilizes the latest generation of the device, now called RetinaScope.
"Traditional retinal cameras are expensive, large, immovable and require special training to operate, whereas RetinaScope is a smartphone-based platform that is cheap, hand-held, and easy to use with no required training" says Paulus, an assistant Prof. of ophthalmology and visual sciences and an assistant Prof. of biomedical engineering.
While smartphone platforms like RetinaScope can be used to deliver high-definition retinal images virtually anywhere, that's only part of the challenge.
"It can take two to seven days for an ophthalmologist to interpret the images," Paulus explains. "To make screening truly accessible, we need to provide on-the-spot feedback, taking the photo and interpreting it while the patient is there to schedule an eye appointment if necessary."
That's where another emerging technology called deep neural network software comes in.
"Deep neural network is an AI software platform that can enhance and review images and provide automated grading of lesions present in DR, indicating which lesions require referral to an ophthalmologist for follow-up," he says.
Paulus' team utilizes a proprietary software platform called EyeArt™ developed by the California-based company Eyenuk. "This is the first study to combine the imaging technology with automated real-time interpretation and compare it to gold standard dilated eye examination," Paulus says, "and the results are very encouraging."
Data was collected from 69 adult patients with diabetes seen in the Kellogg Eye Center Retina Clinic, including previously-recorded results of dilated slit-lamp fundus examinations by their treating clinicians.
After pupillary dilation, RetinaScope was used to image patient retinas and the images were analyzed with EyeArt™ software, which graded them as referral-warranted diabetic retinopathy (RWDR) or non-referral-warranted DR.
The same images were independently evaluated by two expert readers trained to recognize signs of diabetic retinopathy.
"We took the extra step of comparing both automated interpretation and human expert graders with slit lamp evaluation to overcome what we felt was a shortcoming of similar studies conducted elsewhere," says Michael Aaberg, a research assistant in the Paulus lab and study co-author.
The study compared two measurements: if the screening was sensitive enough to find disease, and if it is specific enough to confirm when an individual does not have diabetic retinopathy.
Slit-lamp evaluation confirmed RWDR in 53 subjects (76.8 percent). Automated interpretation had a sensitivity of 86.8 percent (above the 80 percent recommended for an ophthalmic screening device) and specificity of 73.3 percent. One of the human graders achieved a level of sensitivity that was higher by a statistically-significant factor (96.2 percent) and both had lower specificity (40 percent and 46.7 percent).
"This is the first time AI used on a smartphone-based platform has been shown to be effective when compared to the gold standard of clinical evaluation," says Paulus.
Encouraged by both these finding and the soon-to-be-published results of a usability study in a primary care clinic, Paulus' lab continues to pursue hardware and software improvements, as well as FDA clearance.
MEDICA-tradefair.com; Source: Michigan Medicine - University of Michigan