VentureBeat/Ideogram
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More
MRI images are understandably complex and data-heavy.
Because of this, developers training large language models (LLMs) for MRI analysis have had to slice captured images into 2D. But this results in just an approximation of the original image, thus limiting the model’s ability to analyze intricate anatomical structures. This creates challenges in complex cases involving brain tumors, skeletal disorders or cardiovascular diseases.
But GE Healthcare appears to have overcome this massive hurdle, introducing the industry’s first full-body 3D MRI research foundation model (FM) at this year’s AWS re:Invent. For the first time, models can use full 3D images of the entire body.
GE Healthcare’s FM was built on AWS from the ground up — there are very few models specifically designed for medical imaging like MRIs — and is based on more than 173,000 images from over 19,000 studies. Developers say they have been able to train the model with five times less compute than previously required.
GE Healthcare has not yet commercialized the foundation model; it is still in an evolutionary research phase. An early evaluator, Mass General Brigham, is set to begin experimenting with it soon.
“Our vision is to put these models into the hands of technical teams working in healthcare systems, giving them powerful tools for developing research and clinical applications faster, and also more cost-effectively,” GE HealthCare chief AI officer Parry Bhatia told VentureBeat.
Enabling real-time analysis of complex 3D MRI data
While this is a groundbreaking development, generative AI and LLMs are not new territory for the company. The team has been working with advanced technologies for more than 10 years, Bhatia explained.
One of its flagship products is AIR Recon DL, a deep learning-based reconstruction algorithm that allows radiologists to more quickly achieve crisp images. The algorithm removes noise from raw images and improves signal-to-noise ratio, cutting scan times by up to 50%. Since 2020, 34 million patients have been scanned with AIR Recon DL.
GE Healthcare began working on its MRI FM at the beginning of 2024. Because the model is multimodal, it can support image-to-text searching, link images and words, and segment and classify diseases. The goal is to give healthcare professionals more details in one scan than ever before, said Bhatia, leading to faster, more accurate diagnosis and treatment.
“The model has significant potential to enable real-time analysis of 3D MRI data, which can improve medical procedures like biopsies, radiation therapy and robotic surgery,” Dan Sheeran, GM for health care and life sciences at AWS, told VentureBeat.
Already, it has outperformed other publicly-available research models in tasks including classification of prostate cancer and Alzheimer’s disease. It has exhibited accuracy up to 30% in matching MRI scans with text descriptions in image retrieval — which might not sound all that impressive, but it’s a big improvement over the 3% capability exhibited by similar models.
“It has come to a stage where it’s giving some really robust results,” said Bhatia. “The implications are huge.”
Doing more with (much less) data
The MRI process requires a few different types of datasets to support various techniques that map the human body, Bhatia explained.
What’s known as a T1-weighted imaging technique, for instance, highlights fatty tissue and decreases the signal of water, while T2-weighted imaging enhances water signals. The two methods are complementary and create a full picture of the brain to help clinicians detect abnormalities like tumors, trauma or cancer.
“MRI images come in all different shapes and sizes, similar to how you would have books in different formats and sizes, right?” said Bhatia.
To overcome challenges presented by diverse datasets, developers introduced a “resize and adapt” strategy so that the model could process and react to different variations. Also, data may be missing in some areas — an image may be incomplete, for instance — so they taught the model simply to ignore those instances.
“Instead of getting stuck, we taught the model to skip over the gaps and focus on what was available,” said Bhatia. “Think of this as solving a puzzle with some missing pieces.”
The developers also employed semi-supervised student-teacher learning, which is particularly helpful when there is limited data. With this method, two different neural networks are trained on both labeled and unlabeled data, with the teacher creating labels that help the student learn and predict future labels.
“We’re now using a lot of these self-supervised technologies, which don’t require huge amounts of data or labels to train large models,” said Bhatia. “It reduces the dependencies, where you can learn more from these raw images than in the past.”
This helps to ensure that the model performs well in hospitals with fewer resources, older machines and different kinds of datasets, Bhatia explained.
He also underscored the importance of the models’ multimodality. “A lot of technology in the past was unimodal,” said Bhatia. “It would look only into the image, into the text. But now they’re becoming multi-modal, they can go from image to text, text to image, so that you can bring in a lot of things that were done with separate models in the past and really unify the workflow.”
He emphasized that researchers only use datasets that they have rights to; GE Healthcare has partners who license de-identified data sets, and they’re careful to adhere to compliance standards and policies.
Using AWS SageMaker to tackle computation, data challenges
Undoubtedly, there are many challenges when building such sophisticated models — such as limited computational power for 3D images that are gigabytes in size.
“It’s a massive 3D volume of data,” said Bhatia. “You need to bring it into the memory of the model, which is a really complex problem.”
To help overcome this, GE Healthcare built on Amazon SageMaker, which provides high-speed networking and distributed training capabilities across multiple GPUs, and leveraged Nvidia A100 and tensor core GPUs for large-scale training.
“Because of the size of the data and the size of the models, they cannot send it into a single GPU,” Bhatia explained. SageMaker allowed them to customize and scale operations across multiple GPUs that could interact with one another.
Developers also used Amazon FSx in Amazon S3 object storage, which allowed for faster reading and writing for datasets.
Bhatia pointed out that another challenge is cost optimization; with Amazon’s elastic compute cloud (EC2), developers were able to move unused or infrequently used data to lower-cost storage tiers.
“Leveraging Sagemaker for training these large models — mainly for efficient, distributed training across multiple high-performance GPU clusters — was one of the critical components that really helped us to move faster,” said Bhatia.
He emphasized that all components were built from a data integrity and compliance perspective that took into account HIPAA and other regulatory regulations and frameworks.
Ultimately, “these technologies can really streamline, help us innovate faster, as well as improve overall operational efficiencies by reducing the administrative load, and eventually drive better patient care — because now you’re providing more personalized care.”
Serving as a basis for other specialized fine-tuned models
While the model for now is specific to the MRI domain, researchers see great opportunities to expand into other areas of medicine.
Sheeran pointed out that, historically, AI in medical imaging has been constrained by the need to develop custom models for specific conditions in specific organs, requiring expert annotation for each image used in training.
But that approach is “inherently limited” due to the different ways diseases manifest across individuals, and introduces generalizability challenges.
“What we truly need is thousands of such models and the ability to rapidly create new ones as we encounter novel information,” he said. High-quality labeled datasets for each model are also essential.
Now with generative AI, instead of training discrete models for each disease/organ combination, developers can pre-train a single foundation model that can serve as a basis for other specialized fine-tuned models downstream.
For instance, GE Healthcare’s model could be expanded into areas such as radiation therapy, where radiologists spend significant time manually marking organs that might be at risk. It could also help reduce scan time during x-rays and other procedures that currently require patients to sit still in a machine for extended periods, said Bhatia.
Sheeran marveled that “we’re not just expanding access to medical imaging data through cloud-based tools; we’re changing how that data can be utilized to drive AI advancements in healthcare.”
Daily insights on business use cases with VB Daily
If you want to impress your boss, VB Daily has you covered. We give you the inside scoop on what companies are doing with generative AI, from regulatory shifts to practical deployments, so you can share insights for maximum ROI.
Read our Privacy Policy
Thanks for subscribing. Check out more VB newsletters here.
An error occured.