Using a model trained on three data sets could lead to automated categorization of “likely normal” and “likely abnormal” brain MRI results.
MRI has grown in popularity as a tool to image the brain for various diseases or damage, leading to a heavy volume of scans for radiologists to interpret. Using an artificial intelligence (AI)-based system could help automate those interpretations to quickly identify the patients who need the most emergent care.
“MRI is increasingly used clinically for the detection and diagnosis of abnormalities. The resulting image overload presents an urgent need for improved radiologic workflow,” said a team of investigators from Massachusetts General Hospital (MGH) and Brigham and Women’s Hospital Center for Clinical Data Science. “Automatically identifying abnormal findings in medical images could improve the way worklists are prioritized, enabling improved patient care and accelerated patient discharge.”
The team, co-led by Romane Gauriau, Ph.D., former machine learning scientist from MGH, and Bernardo C. Bizzo, M.D., Ph.D., an MGH radiologist, published their findings Wednesday in Radiology: Artificial Intelligence. Their efforts resulted in a triaging system for brain MRIs that can detect multiple abnormalities from 3D volumetric imaging using a convolutional neural network and categorize the findings as “likely normal” and “likely abnormal.”
This is the first investigation to use a large, clinically relevant dataset and full-volume MRI data to identify overall brain abnormalities, they said.
For their retrospective study, the team used three large datasets that included more than 9,000 brain MRI scans culled from various institutions across two continents. They randomly split these datasets into training, validation, and test sets. All images were annotated by four radiologists who collated the studies into 10 categories: likely normal, almost normal, neoplasm, hemorrhage, infarct, infection or inflammatory disease, demyelination disease, vascular lesions, congenital, and other. In addition, FLAIR sequences were reviewed and categorized as “likely normal” and “likely abnormal.”
Based on their preliminary results, the team said, the model demonstrated relatively good performance in its ability to distinguish between “likely normal” and “likely abnormal” exams. Specifically, model A, trained on a dataset of axial FLAIR sequence exams conducted on adults and tested on a similar dataset that also included children, reached an F1-score of 0.72 and an area under the curve of 0.78 when compared with radiologic reports.
“Our experiments reveal reasonable performance for triaging purposes of a neural network for classification of likely normal or abnormal brain MRIs based on axial FLAIR sequences,” the team said. “Such a system could enable the creation of a ‘likely abnormal’ worklist which could be used to improve radiology workflow.”
With such a tool, they said, experienced radiologists and neuroradiologists could concentrate on the scans that were more urgent and time-consuming, and trainees and junior attendings could use it as a learning opportunity.
Their results also pointed to the generalizability of the model. They tested it on a validation set that was acquired during a different time period from a different institution than the data used the train the algorithm.
“The problem we are trying to tackle is very, very complex because there are a huge variety of abnormalities on MRI,” Gauriau said. “We showed that this model is promising enough to start evaluating if it can be used in a clinical environment.”
Like similar models that have been used to pinpoint abnormalities with CT scans and chest X-rays, this model has the potential to significantly improve turnaround time, as well as identifying incidental findings in outpatient settings.
“Say you fell and hit your head, then went to the hospital and they ordered a brain MRI,” she said. “This algorithm could detect if you have brain injury from the fall, but it may also detect an unexpected finding, such as a brain tumor. Having the ability could really help improve patient care.”
These findings point to the need for continued work, the team said. The next steps involve evaluating the model’s clinical utility and potential value for radiologist. Ultimately, they said, they wanted to be able further develop the model to provide the binary outputs “likely normal” and “likely abnormal.”
For more coverage based on industry expert insights and research, subscribe to the Diagnostic Imaging e-Newsletter here.
New Study Examines Short-Term Consistency of Large Language Models in Radiology
November 22nd 2024While GPT-4 demonstrated higher overall accuracy than other large language models in answering ACR Diagnostic in Training Exam multiple-choice questions, researchers noted an eight percent decrease in GPT-4’s accuracy rate from the first month to the third month of the study.
FDA Grants Expanded 510(k) Clearance for Xenoview 3T MRI Chest Coil in GE HealthCare MRI Platforms
November 21st 2024Utilized in conjunction with hyperpolarized Xenon-129 for the assessment of lung ventilation, the chest coil can now be employed in the Signa Premier and Discovery MR750 3T MRI systems.
FDA Clears AI-Powered Ultrasound Software for Cardiac Amyloidosis Detection
November 20th 2024The AI-enabled EchoGo® Amyloidosis software for echocardiography has reportedly demonstrated an 84.5 percent sensitivity rate for diagnosing cardiac amyloidosis in heart failure patients 65 years of age and older.
New Study Examines Agreement Between Radiologists and Referring Clinicians on Follow-Up Imaging
November 18th 2024Agreement on follow-up imaging was 41 percent more likely with recommendations by thoracic radiologists and 36 percent less likely on recommendations for follow-up nuclear imaging, according to new research.