Annotating regions of interest in medical images, a process known as segmentation, is often one of the first steps clinical researchers take when running a new study involving biomedical images. Manual segmentation can be extremely time-consuming, particularly when delineating challenging areas such as the hippocampus in brain scans.
To address this inefficiency, researchers at MIT have developed an innovative artificial intelligence-based system designed to streamline this process. The new tool allows researchers to rapidly segment biomedical imaging datasets by engaging in simple interactions such as clicking, scribbling, and drawing boxes. The AI model then leverages these interactions to predict segmentations, improving efficiency to the extent that it can eventually segment new images accurately without further user input.
This advancement is primarily due to the model’s architecture, which utilizes information from previously segmented images to enhance predictions, thereby reducing the repetitive workload typically associated with image segmentation tasks. Unlike traditional segmentation models, which may require extensive pre-trained datasets, this system is designed to operate effectively without such prerequisites, making it accessible for use by clinical researchers who may lack machine-learning expertise or robust computational resources.
As Hallee Wong, the lead author of the paper on this tool, articulates, the hope is to eliminate the barriers that prevent scientists from engaging in meaningful studies due to time constraints. By automating much of the segmentation process, the system allows clinical researchers to focus more on analysis and interpretation rather than manual data preparation, potentially leading to breakthroughs that were previously deemed unattainable.
The researchers highlight two conventional methods for segmenting medical images: interactive segmentation, which requires marking areas of interest for each image, and developing a task-specific AI model that necessitates extensive manual work to train. The new system, named MultiverSeg, merges the benefits of these approaches by enabling the user to reference a context set of images. This allows for a more efficient segmentation process, improving accuracy while demanding significantly less user intervention over time.
Testing has shown that MultiverSeg demands fewer interactions as users segment subsequent images. For example, by the ninth image processed, researchers found that only two interactions were required to achieve segmentation accuracy superior to traditional models. With image types such as X-rays, the learning curve can be even shorter, potentially requiring only one or two manual segmentations.
Importantly, the tool’s interactivity allows users to refine the AI model’s predictions efficiently, contributing to a more accurate outcome with less total effort. Compared to previous systems developed by the researchers, MultiverSeg has demonstrated a 90 percent accuracy rate with significantly fewer user interactions necessary.
Looking ahead, researchers plan to collaborate with clinical partners to evaluate the system in real-world settings and gather feedback for ongoing enhancements. Additionally, they are considering expanding MultiverSeg’s capabilities to include the segmentation of 3D biomedical images, further broadening the tool’s applicability.
This groundbreaking work is supported in part by Quanta Computer, Inc. and the National Institutes of Health, further underscoring the potential for AI to transform clinical research through more effective and scalable image analysis.