A quick and versatile approach to assist doctors annotate medical scans

-

To the untrained eye, a medical image like an MRI or X-ray appears to be a murky collection of black-and-white blobs. It could be a struggle to decipher where one structure (like a tumor) ends and one other begins. 

When trained to grasp the boundaries of biological structures, AI systems can segment (or delineate) regions of interest that doctors and biomedical employees want to watch for diseases and other abnormalities. As a substitute of losing precious time tracing anatomy by hand across many images, a man-made assistant could do this for them.

The catch? Researchers and clinicians must label countless images to coach their AI system before it might accurately segment. For instance, you’d must annotate the cerebral cortex in quite a few MRI scans to coach a supervised model to grasp how the cortex’s shape can vary in several brains.

Sidestepping such tedious data collection, researchers from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL), Massachusetts General Hospital (MGH), and Harvard Medical School have developed the interactive “ScribblePrompt” framework: a versatile tool that can assist rapidly segment any medical image, even types it hasn’t seen before. 

As a substitute of getting humans mark up each picture manually, the team simulated how users would annotate over 50,000 scans, including MRIs, ultrasounds, and images, across structures within the eyes, cells, brains, bones, skin, and more. To label all those scans, the team used algorithms to simulate how humans would scribble and click on on different regions in medical images. Along with commonly labeled regions, the team also used superpixel algorithms, which find parts of the image with similar values, to discover potential latest regions of interest to medical researchers and train ScribblePrompt to segment them. This synthetic data prepared ScribblePrompt to handle real-world segmentation requests from users.

“AI has significant potential in analyzing images and other high-dimensional data to assist humans do things more productively,” says MIT PhD student Hallee Wong SM ’22, the lead creator on a latest paper about ScribblePrompt and a CSAIL affiliate. “We wish to enhance, not replace, the efforts of medical employees through an interactive system. ScribblePrompt is an easy model with the efficiency to assist doctors concentrate on the more interesting parts of their evaluation. It’s faster and more accurate than comparable interactive segmentation methods, reducing annotation time by 28 percent in comparison with Meta’s Segment Anything Model (SAM) framework, for instance.”

ScribblePrompt’s interface is easy: Users can scribble across the rough area they’d like segmented, or click on it, and the tool will highlight the complete structure or background as requested. For instance, you possibly can click on individual veins inside a retinal (eye) scan. ScribblePrompt may also mark up a structure given a bounding box.

Then, the tool could make corrections based on the user’s feedback. Should you wanted to focus on a kidney in an ultrasound, you might use a bounding box, after which scribble in additional parts of the structure if ScribblePrompt missed any edges. Should you desired to edit your segment, you might use a “negative scribble” to exclude certain regions.

These self-correcting, interactive capabilities made ScribblePrompt the popular tool amongst neuroimaging researchers at MGH in a user study. 93.8 percent of those users favored the MIT approach over the SAM baseline in improving its segments in response to scribble corrections. As for click-based edits, 87.5 percent of the medical researchers preferred ScribblePrompt.

ScribblePrompt was trained on simulated scribbles and clicks on 54,000 images across 65 datasets, featuring scans of the eyes, thorax, spine, cells, skin, abdominal muscles, neck, brain, bones, teeth, and lesions. The model familiarized itself with 16 kinds of medical images, including microscopies, CT scans, X-rays, MRIs, ultrasounds, and images.

“Many existing methods don’t respond well when users scribble across images since it’s hard to simulate such interactions in training. For ScribblePrompt, we were capable of force our model to concentrate to different inputs using our synthetic segmentation tasks,” says Wong. “We desired to train what’s essentially a foundation model on a variety of diverse data so it could generalize to latest kinds of images and tasks.”

After taking in a lot data, the team evaluated ScribblePrompt across 12 latest datasets. Even though it hadn’t seen these images before, it outperformed 4 existing methods by segmenting more efficiently and giving more accurate predictions in regards to the exact regions users wanted highlighted.

“​​Segmentation is essentially the most prevalent biomedical image evaluation task, performed widely each in routine clinical practice and in research — which results in it being each very diverse and a vital, impactful step,” says senior creator Adrian Dalca SM ’12, PhD ’16, CSAIL research scientist and assistant professor at MGH and Harvard Medical School. “ScribblePrompt was fastidiously designed to be practically useful to clinicians and researchers, and hence to substantially make this step much, much faster.”

“The vast majority of segmentation algorithms which were developed in image evaluation and machine learning are not less than to some extent based on our ability to manually annotate images,” says Harvard Medical School professor in radiology and MGH neuroscientist Bruce Fischl, who was not involved within the paper. “The issue is dramatically worse in medical imaging during which our ‘images’ are typically 3D volumes, as human beings haven’t any evolutionary or phenomenological reason to have any competency in annotating 3D images. ScribblePrompt enables manual annotation to be carried out much, much faster and more accurately, by training a network on precisely the kinds of interactions a human would typically have with a picture while manually annotating. The result’s an intuitive interface that permits annotators to naturally interact with imaging data with far greater productivity than was previously possible.”

Wong and Dalca wrote the paper with two other CSAIL affiliates: John Guttag, the Dugald C. Jackson Professor of EECS at MIT and CSAIL principal investigator; and MIT PhD student Marianne Rakic SM ’22. Their work was supported, partially, by Quanta Computer Inc., the Eric and Wendy Schmidt Center on the Broad Institute, the Wistron Corp., and the National Institute of Biomedical Imaging and Bioengineering of the National Institutes of Health, with hardware support from the Massachusetts Life Sciences Center.

Wong and her colleagues’ work shall be presented on the 2024 European Conference on Computer Vision and was presented as an oral talk on the DCAMI workshop on the Computer Vision and Pattern Recognition Conference earlier this 12 months. They were awarded the Bench-to-Bedside Paper Award on the workshop for ScribblePrompt’s potential clinical impact.

ASK ANA

What are your thoughts on this topic?
Let us know in the comments below.

0 0 votes
Article Rating
guest
0 Comments
Oldest
Newest Most Voted
Inline Feedbacks
View all comments

Share this article

Recent posts

0
Would love your thoughts, please comment.x
()
x