The pros and cons of using standardised protocols
The evaluation of imaging biomarker efficiency is a complicated task when it is based on real-world data derived from a single centre [1]. Over the last decade, there have been more than 125 registered trials investigating treatment effects and monitoring disease progression in amyotrophic lateral sclerosis [2]. Despite that, fixed protocols aimed at measuring imaging biomarkers’ effect size, sensitivity, and specificity remain out of reach. The problem with imaging biomarker discovery and validation is the tremendous variability in the clinical course of this disease [3]. Since clinical course variability is primarily due to multiple mechanisms and progression-driving factors, it may be more productive to look at a syndrome affecting a pathway rather than a more specific gene-protein interaction [4]. The disease severity and duration variability mean that the imaging biomarkers might contribute to the diagnosis. However, image-derived measures alone will not be sufficient to detect the disease. Nonetheless, an abnormal finding in an imaging study is not necessarily associated with clinical progression, and might not be present in some patients at the beginning of the disease course. The main problems are the small sample size, the low power to detect the difference, and restricted funding for the reference dataset. The small sample size is a particular challenge because amyotrophic lateral sclerosis is rare. However, open reference datasets are available [5] and enable the comparison of clinical patient populations to extensive standardised, high-quality imaging data [6].
Automatic vs. manual extraction of regions of interest (ROI)
Manual identification of critical landmarks is feasible in small clinical datasets, but is more difficult when encompassing extensive population studies. Research presented in this issue has manually extracted anatomical landmarks and average values [1]. Achieving reproducibility of these measurements using automatic registration-based methods would represent a significant advance [7]. Software solutions for extracting quantitative markers and landmarks of the motor system are progressing at pace. The clinical community needs to be aware of the methods, already widely available, for analysing large datasets of brain images for research purposes and the achievement of tangible improvements using high-performance computing platforms (HPC) and federated machine learning [8].
A recent achievement in brain tumour research highlights the potential. It is critical to use quantitative parameters that include cortical thickness, cortico-spinal tract coherence measures, and brain and lateral ventricle volumes obtained from reference populations of healthy controls matched for age and gender. Such measures enable clinicians to track the evolution of the brain injury in question well before the damage becomes apparent and irreversible [7, 9]. Efforts to harness the global community of researchers interested in neurodegenerative diseases to share protocols, image processing pipelines, and DNA-based biomarkers, are already underway in what is known as the ‘United Consortium’ [10]. Aggregating imaging resources from multiple sites using centralised high-performance computing (HPC) enables the discovery of new structural features and validation of existing structural biomarkers on large datasets [11]. A Canadian project has already started to combine and aggregate data from patients diagnosed with amyotrophic lateral sclerosis and other neurodegenerative diseases and controls. MRI datasets have been collected and analysed using a distributed database [12].
Biological precision vs. technical feasibility
The source of the abnormal signal captured in the recent study by Maj et al. came from an altered white matter tract signal, and possibly myelination [1]. Identifying and measuring demyelination of the cortico-spinal pathway has been attempted using various magnetic resonance imaging methods. Currently, a precise imaging-histological correlation is hard to reach within a reasonable timeframe applicable to patient-focused clinical protocols. However, hope lies in synthetic MRI methods based on multi-dynamic multi-echo (MDME) sequences. It is important to determine how various pathologies will affect the measurements based on T2-relaxation before comprehensive clinical implementation begins [13]. Diffusion-weighted imaging remains the most commonly available sequence that measures radial diffusivity and apparent diffusion coefficient feasible on most modern clinical MRI scanners with an acceptable acquisition duration time.
In summary, it is essential to understand that the goal of standardising MRI data acquisition in this field should unite all neurologists and all clinicians. Ontario Neurodegenerative Disease Research Initiative (ONDRI), a Canadian research community, has identified all neurodegenerative diseases as a global challenge and is using all available technology to facilitate the search for a solution in the form of a unified platform for the acquisition and analysis of structural brain biomarkers [14].
MRI acquisitions using widely available standardised protocols will enable analysis using validated software processing pipelines [15, 16]. The results will allow the evaluation of single-subject results against a large and growing database of morphological biomarkers for tracking the course of the clinical disease. This will lead to a better prognosis and improve the detection of progression in every patient, leading to the faster discovery of therapeutic agents.