A novel machine-learning algorithm used with MRI can harmonize brain volumetric data of patients undergoing Alzheimer's disease assessment gathered from different scanners, researchers have reported.
"When evaluated on brain MRI marker data from participants along the Alzheimer's disease spectrum, our new model outperformed the other approaches we tested on both seen and unseen [i.e., scanners not included in the initial training of the model] scanners," wrote a team led by Damian Archetti, PhD, of IRCCS Istituto Centro San Giovanni di Dio Fatebenefratelli in Brescia, Italy. The findings were published December 18 in Radiology: Artificial Intelligence.
Structural MR imaging, such as T1-weighted exams, are commonly used in memory clinics for diagnosing Alzheimer's disease and differentiating Alzheimer's from other types of dementias, the group explained. Radiologists visually interpret these exams, but these interpretations are subjective and "prone to intrarater and interrater variability," it noted. Using quantitative imaging markers culled from brain volumetric data shows promise for improving readers' diagnostic confidence, and there are AI algorithms that track these. But differences in MRI acquisition protocols and scanners can affect the consistency of brain volumetry assessment.
Neuroharmony is a relatively new tool for harmonizing volumetric data from images taken using new or unseen MRI scanners. It uses image quality metrics (IQM) as predictors to remove scanner-related effects in brain-volumetric data using random forest regression and has been tested on data from cognitively healthy people; however, it has not been tested on data from patients with neurodegenerative diseases, the researchers explained.
To this end, Archetti and colleagues developed an extended version of Neuroharmony to track interactions between Alzheimer's disease pathology and image quality metrics in 20,864 individuals with and without cognitive impairment. These data spanned 11 prospective and retrospective study cohorts and 43 scanners. The group then evaluated the expanded algorithm's ability to remove scanner-related variations in brain volumes (that is, to track marker concordance between scanners) while continuing to delineate disease-related signals.
The investigators found that marker concordances between scanners were significantly better (p < 0.001) compared with data previously processed without the algorithm. They also found that the expanded model showed higher concordance than the original model (0.75 compared with 0.7, with 1 as reference) and better-preserved disease signals (area under the receiver operating curve, or AUC, -0.006 compared with -0.091). Finally, the group reported that marker concordance was better in scanners used to train the algorithm (> 0.97) than in those not used for this purpose (< 0.79) -- a result that emerged independently of study participants' cognitive status.
The study results show promise, but further research is needed, according to the authors.
"Further validation using different processing pipelines and evaluation criteria would be essential for clinical use of the model in applications related to cognitive decline, such as memory clinics and clinical trials of new interventions for neurodegenerative diseases," they concluded.
The complete study can be found here.