Skip to main content
Advertisement

Main menu

  • Home
  • Content
    • Current Issue
    • Accepted Manuscripts
    • Article Preview
    • Past Issue Archive
    • Video Articles
    • AJNR Case Collection
    • Case of the Week Archive
    • Case of the Month Archive
    • Classic Case Archive
  • Special Collections
    • AJNR Awards
    • Low-Field MRI
    • Alzheimer Disease
    • ASNR Foundation Special Collection
    • Photon-Counting CT
    • View All
  • Multimedia
    • AJNR Podcasts
    • AJNR SCANtastic
    • Trainee Corner
    • MRI Safety Corner
    • Imaging Protocols
  • For Authors
    • Submit a Manuscript
    • Submit a Video Article
    • Submit an eLetter to the Editor/Response
    • Manuscript Submission Guidelines
    • Statistical Tips
    • Fast Publishing of Accepted Manuscripts
    • Graphical Abstract Preparation
    • Imaging Protocol Submission
    • Author Policies
  • About Us
    • About AJNR
    • Editorial Board
    • Editorial Board Alumni
  • More
    • Become a Reviewer/Academy of Reviewers
    • Subscribers
    • Permissions
    • Alerts
    • Feedback
    • Advertisers
    • ASNR Home

User menu

  • Alerts
  • Log in

Search

  • Advanced search
American Journal of Neuroradiology
American Journal of Neuroradiology

American Journal of Neuroradiology

ASHNR American Society of Functional Neuroradiology ASHNR American Society of Pediatric Neuroradiology ASSR
  • Alerts
  • Log in

Advanced Search

  • Home
  • Content
    • Current Issue
    • Accepted Manuscripts
    • Article Preview
    • Past Issue Archive
    • Video Articles
    • AJNR Case Collection
    • Case of the Week Archive
    • Case of the Month Archive
    • Classic Case Archive
  • Special Collections
    • AJNR Awards
    • Low-Field MRI
    • Alzheimer Disease
    • ASNR Foundation Special Collection
    • Photon-Counting CT
    • View All
  • Multimedia
    • AJNR Podcasts
    • AJNR SCANtastic
    • Trainee Corner
    • MRI Safety Corner
    • Imaging Protocols
  • For Authors
    • Submit a Manuscript
    • Submit a Video Article
    • Submit an eLetter to the Editor/Response
    • Manuscript Submission Guidelines
    • Statistical Tips
    • Fast Publishing of Accepted Manuscripts
    • Graphical Abstract Preparation
    • Imaging Protocol Submission
    • Author Policies
  • About Us
    • About AJNR
    • Editorial Board
    • Editorial Board Alumni
  • More
    • Become a Reviewer/Academy of Reviewers
    • Subscribers
    • Permissions
    • Alerts
    • Feedback
    • Advertisers
    • ASNR Home
  • Follow AJNR on Twitter
  • Visit AJNR on Facebook
  • Follow AJNR on Instagram
  • Join AJNR on LinkedIn
  • RSS Feeds

AJNR Awards, New Junior Editors, and more. Read the latest AJNR updates

Getting new auth cookie, if you see this message a lot, tell someone!
Research ArticlePediatric Neuroimaging
Open Access

Automatic Machine Learning to Differentiate Pediatric Posterior Fossa Tumors on Routine MR Imaging

H. Zhou, R. Hu, O. Tang, C. Hu, L. Tang, K. Chang, Q. Shen, J. Wu, B. Zou, B. Xiao, J. Boxerman, W. Chen, R.Y. Huang, L. Yang, H.X. Bai and C. Zhu
American Journal of Neuroradiology July 2020, 41 (7) 1279-1285; DOI: https://doi.org/10.3174/ajnr.A6621
H. Zhou
gDepartment of Neurology (H.Z., L.T., B.X.), Xiangya Hospital of Central South University, Changsha, Hunan, China
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for H. Zhou
R. Hu
aFrom the School of Computer Science and Engineering (R.H., B.Z., C.Z.)
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for R. Hu
O. Tang
fWarren Alpert Medical School, Brown University (O.T.), Providence, Rhode Island
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for O. Tang
C. Hu
jDepartment of Neurology (C.H.), Hunan Provincial People’s Hospital, Changsha, Hunan, China
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for C. Hu
L. Tang
gDepartment of Neurology (H.Z., L.T., B.X.), Xiangya Hospital of Central South University, Changsha, Hunan, China
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for L. Tang
K. Chang
iDepartment of Radiology (K.C.), Athinoula A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Harvard Medical School, Boston, Massachusetts
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for K. Chang
Q. Shen
dRadiology (Q.S., J.W.), Second Xiangya Hospital of Central South University, Changsha, Hunan, China
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Q. Shen
J. Wu
dRadiology (Q.S., J.W.), Second Xiangya Hospital of Central South University, Changsha, Hunan, China
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for J. Wu
B. Zou
aFrom the School of Computer Science and Engineering (R.H., B.Z., C.Z.)
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for B. Zou
B. Xiao
gDepartment of Neurology (H.Z., L.T., B.X.), Xiangya Hospital of Central South University, Changsha, Hunan, China
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for B. Xiao
J. Boxerman
eDepartment of Diagnostic Imaging (J.B., H.X.B.), Rhode Island Hospital
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for J. Boxerman
W. Chen
kDepartment of Pathology (W.C.), Hunan Children’s Hospital, Changsha, Hunan, China
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for W. Chen
R.Y. Huang
hDepartment of Radiology (R.Y.H.), Brigham and Women’s Hospital, Boston, Massachusetts
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for R.Y. Huang
L. Yang
cDepartments of Neurology (L.Y.)
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for L. Yang
H.X. Bai
eDepartment of Diagnostic Imaging (J.B., H.X.B.), Rhode Island Hospital
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for H.X. Bai
C. Zhu
aFrom the School of Computer Science and Engineering (R.H., B.Z., C.Z.)
bCollege of Literature and Journalism (C.Z.), Central South University, Changsha, Hunan, China
lMobile Health Ministry of Education-China Mobile Joint Laboratory (C.Z.), China.
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for C. Zhu
  • Article
  • Figures & Data
  • Supplemental
  • Info & Metrics
  • Responses
  • References
  • PDF
Loading

Abstract

BACKGROUND AND PURPOSE: Differentiating the types of pediatric posterior fossa tumors on routine imaging may help in preoperative evaluation and guide surgical resection planning. However, qualitative radiologic MR imaging review has limited performance. This study aimed to compare different machine learning approaches to classify pediatric posterior fossa tumors on routine MR imaging.

MATERIALS AND METHODS: This retrospective study included preoperative MR imaging of 288 patients with pediatric posterior fossa tumors, including medulloblastoma (n = 111), ependymoma (n = 70), and pilocytic astrocytoma (n = 107). Radiomics features were extracted from T2-weighted images, contrast-enhanced T1-weighted images, and ADC maps. Models generated by standard manual optimization by a machine learning expert were compared with automatic machine learning via the Tree-Based Pipeline Optimization Tool for performance evaluation.

RESULTS: For 3-way classification, the radiomics model by automatic machine learning with the Tree-Based Pipeline Optimization Tool achieved a test micro-averaged area under the curve of 0.91 with an accuracy of 0.83, while the most optimized model based on the feature-selection method χ2 score and the Generalized Linear Model classifier achieved a test micro-averaged area under the curve of 0.92 with an accuracy of 0.74. Tree-Based Pipeline Optimization Tool models achieved significantly higher accuracy than average qualitative expert MR imaging review (0.83 versus 0.54, P < .001). For binary classification, Tree-Based Pipeline Optimization Tool models achieved an area under the curve of 0.94 with an accuracy of 0.85 for medulloblastoma versus nonmedulloblastoma, an area under the curve of 0.84 with an accuracy of 0.80 for ependymoma versus nonependymoma, and an area under the curve of 0.94 with an accuracy of 0.88 for pilocytic astrocytoma versus non-pilocytic astrocytoma.

CONCLUSIONS: Automatic machine learning based on routine MR imaging classified pediatric posterior fossa tumors with high accuracy compared with manual expert pipeline optimization and qualitative expert MR imaging review.

ABBREVIATIONS:

AUC
area under the curve
AutoML
automatic machine learning
CHSQ
χ2 score
EP
ependymoma
MB
medulloblastoma
ML
machine learning
PA
pilocytic astrocytoma
TPOT
Tree-Based Pipeline Optimization Tool

Among childhood malignancies, pediatric brain tumors are the second most common and the leading cause of death from solid tumors.1,2 Posterior fossa tumors make up a disproportionate portion of brain tumors in the pediatric population, accounting for 54%–70% of tumors compared with <20% in the adult population.3 The most common subtypes of posterior fossa tumors among children are medulloblastoma (MB), pilocytic astrocytoma (PA), and ependymoma (EP).4,5 Discrimination of these 3 malignancies is important due to the differing natural histories and prognoses for each. Accurate preoperative diagnosis could also help in preoperative evaluation and guide surgical planning for patients with different types of tumors.5,6 MR imaging review is essential for tumor diagnosis and evaluation. However, it is usually far from being a criterion standard, which is diagnosis by histology.

The rapid growth in the field of medical image analysis in the past decade has facilitated the development of radiomics, which converts digital images to mineable data via high-throughput extraction of a large number of quantitative features.7 The radiomics model has the potential power to improve predictive performance8 and has been used to improve the differentiation among different brain tumors such as glioblastoma and anaplastic oligodendroglioma,9 and metastatic brain tumors.10 However, there have been few studies exploring the classification of pediatric posterior fossa tumors using a radiomics analysis on clinical routine MR imaging.

Machine learning (ML) is a method of data analysis that automates analytic model building. It is a branch of artificial intelligence in which systems are designed to learn from data, identify patterns, and make decisions with minimal human intervention.11 In recent years, machine learning approaches for radiomics feature selection have developed rapidly. However, few studies have compared these recently developed feature-selection methods and predictive modeling methods.12 Therefore, selection of the most accurate ML pipeline often requires extensive manual testing by an expert in machine learning.

The Tree-Based Pipeline Optimization Tool (TPOT; https://pypi.org/project/TPOT/) is a specific open-source form of automated machine learning (autoML) that automatically chooses the most optimal machine learning pipeline without the need for human intervention.13,14 TPOT has been benchmarked on many simulated and real-world supervised classification tasks.14 However, performance of automated TPOT still needs to be proved in different types of data by comparison with manually optimized feature selection and classification.

The primary goal of our study was to investigate the value of autoML (TPOT) to differentiate the main types of pediatric posterior fossa tumors on routine MR imaging and to compare its performance with manual expert optimization and qualitative expert MR imaging review.

MATERIALS AND METHODS

Patient Cohort

All patients included were newly diagnosed with histologically confirmed MB, EP, or PA from 4 large academic hospitals in the Hunan Province in China from January 2013 to December 2018. The study was approved by the institutional review boards of all participating institutions. The inclusion criteria were the following: 1) pathologically confirmed MB, EP, or PA; 2) available preoperative MR imaging examination including a T1-weighted contrast-enhanced sequence, a T2-weighted sequence, and ADC maps; and 3) quality of the images adequate for analysis, without motion or artifacts. All patients included in the analysis did not undergo any treatment before MR imaging. Patients’ age and sex were collected and incorporated into the radiomics analysis as clinical variables.

MR Imaging Acquisition

The MR imaging was performed using 3T scanners (Magnetom Trio, Siemens, n = 168 patients; Achieva, Philips Healthcare, n = 87) or 1.5T scanners (Avanto, Siemens, n = 33). The main MR imaging protocols across hospitals used in our study are listed in On-line Table 1. ADC images were calculated from acquired DWI with b-values of 0- and 1000-s/mm2.

Image Segmentation

MR images of all patients were loaded into 3D Slicer software (Version 4.10; http://www.slicer.org), which was used as a user-driven manual active contour-segmentation tool to segment tumor volumes. ROIs were manually drawn section by section on the T2WI, T1WI contrast-enhanced sequence, and ADC images by 2 authors (H.Z., with 7 years of clinical experience in neuroradiology, and L.T., with 5 years of clinical experience in neuroradiology). The ROIs defined in our study include both the enhancing and nonenhancing components of tumor as well as peritumoral edema. On-line Figure 1 shows an example of ROI delineation on a T1WI contrast-enhanced sequence, T2WI, and ADC maps. The intraclass correlation coefficient value was used to assess interobserver reproducibility of generated masks.

Radiomics Feature Extraction

Radiomics features were extracted from each patient’s MR imaging for the T1WI contrast-enhanced sequence, T2WI, and ADC maps. For each image space, 79 nontexture (morphology and intensity-based) and 94 texture features were extracted according to the guidelines defined by the Image Biomarker Standardization Initiative.15 Each of the 94 texture features was computed 32 times using all possible combinations of the following extraction parameters, a process known as “texture optimization”:16 1) isotropic voxels of 1, 2, 3 , and 4 mm; 2) a fixed bin number discretization algorithm, with and without equalization; and 3) the number of gray levels of 8, 16, 32, and 64 for a fixed bin number. A total of 79 + 32 × 94 or 3087 radiomics features were thus computed in this study. All the features were normalized using unity-based normalization, and features from the T1WI contrast-enhanced sequence, T2WI, and ADC maps were combined into 1 dataset.

Radiomics Model Building and Analysis

Before radiomics analysis, all included tumor cases were randomly divided into a training set (70% of cases) and testing set (30% of cases). All the radiomics features extracted above with clinical variables (age and sex) were incorporated to train multiclass models for MB, EP, and PA. Different binary models were trained for MBs versus non-MBs, EPs versus non-EPs, and PAs versus non-PAs. An automated optimized pipeline was computed on the dataset by TPOT, which chooses the most optimal machine learning pipeline for an inputted dataset through genetic programming.17 The following settings were used in the optimization process: number of generations, 5; population size, 5; and 5-fold cross-validation on the training set. In this study, the TPOT pipeline was replicated 10 times to yield 10 different models for comparison. The model with the best performance was finally selected and tested on the testing set.

For the manual expert optimized pipeline, radiomics features were selected for training using 13 different feature-selection methods to reduce the dimensionality of the datasets. Ten machine learning classifiers were trained to yield diagnostic models. Details of the feature-selection methods and classifiers used are shown in On-line Table 2. Each classifier was trained on the training set 13 times using 13 different feature-selection methods and was validated through 5-fold cross-validation. Classifiers were trained on 10, 20, 50, and 100 selected features, and performances on the training set were recorded. The best-performing models on the training set were then tested on the final testing set. When there were multiple combinations with the same cross-validation results on the training set, the combination with the best performance on the testing set was selected. In addition to performance, the stability of classifiers was recorded. The relative SD for classifier stability was calculated. More details and the results of the stability calculations can be found in On-line Table 3. Figure 1 provides the workflow of the machine learning algorithms used in this study.

FIG 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
FIG 1.

Machine learning workflow.

Expert Evaluation

The final performance of the model on the test set was compared with 2 expert radiologists’ interpretations: Q.S. and J.W. with 7 and 5 years of experience reading brain MR imaging, respectively. The experts were blinded to all clinical data and histopathologic information.

Statistical Analysis

The following performance metrics were calculated for binary classification: area under receiver operating characteristic curve (AUC), accuracy, sensitivity, and specificity. For multiclass (3-way) classification, micro-averaged AUC, accuracy, sensitivity, and specificity were calculated. The micro-average aggregated the contributions of all classes to compute the average metric, which is more preferable when dealing with multiclass classification.18 The 95% confidence intervals on accuracy, sensitivity, and specificity were calculated using the adjusted Wald method.19 The P values were calculated using the binomial test. A P value < .05 was considered statistically significant.

Code Availability

The implementation of the radiomics feature extraction was based on the “radiomics-develop” package from the Naqa Lab at McGill University.16 This code is available for public use on GitHub at https://github.com/mvallieres/radiomics-develop. The implementation of the machine learning models was based on the scikit-learn package of Python at https://github.com/scikit-learn/scikit-learn. The Auto-ML script used the TPOT package from the Epistasis Lab.17 This code is publicly available at https://github.com/subhanik1999/Radiomics-ML.

RESULTS

Patient Characteristics

A total of 288 patients with pediatric posterior fossa tumors (111 MBs, 70 EPs, and 107 PAs) were included in the study cohort. The mean age of all patients was 8.6 years, ranging from 0.25 to 18 years. The mean ages of patients with MB, EP, and PA were 9.0, 9.1, and 7.9 years, respectively (P = .291). There were 161 males and 127 females for the whole cohort. The sex ratio for each type of tumor (male/female) was 65:46, 36:34, and 60:47, respectively (P = .640).

Agreement of Segmentation

For each case in the study, the average required time for segmentation was about 10 minutes. The average intraclass correlation coefficient between the 2 segmenters was 0.91 for T2WI, 0.92 for the T1WI contrast-enhanced sequence, and 0.86 for ADC.

Multiclass Classification for 3 Tumors (MB versus EP versus PA)

For the TPOT pipeline, 10 separate models were generated for differentiation of the 3 tumor types. Overall, all 10 models showed a favorable micro-averaged AUC on the training set (On-line Table 4). Model 8 achieved the best performance with the highest micro-averaged AUC (0.93) among the 10 models. The parameters of all the TPOT models are described in detail in On-line Table 4. The top 20 important radiomics features are listed in On-line Table 5. In the testing cohort, the best model (model 8) achieved a micro-averaged AUC of 0.91 and an accuracy of 0.83 (95% CI, 0.72–0.90). The sensitivity and specificity of the diagnostic model for each type of tumor are shown in Table 1.

View this table:
  • View inline
  • View popup
Table 1:

Performance of multiclass classification across all models and experts in test set

For manual expert optimization, micro-averaged AUC scores on the training set from each combination of feature-selection method and classifier are shown in On-line Fig 2. The combination of χ2 score (CHSQ) as the feature-selection method and the FSL General Linear Model (GLM; http://fsl.fmrib.ox.ac.uk/fsl/fslwiki/GLM) as the classifier achieved the highest micro-averaged AUC of 0.93 on the training set. As the top-performing model, CHSQ + GLM was then tested on the final test set and achieved a test micro-averaged AUC of 0.92 and accuracy of 0.74 (95% CI, 0.62–0.83). The sensitivity and specificity of the diagnostic model for each type of tumor are shown in Table 1. The top 20 important radiomics features used by the CHSQ + GLM model are shown in On-line Table 6.

Binary Classification

For the TPOT pipeline, 30 separate models were generated for the binary classification of MB versus non-MB, EP versus non-EP, and PA versus non-PA (10 models in each classification).

The performances of all the models on the training set are shown in On-line Table 7. Among all models, model 9 for MB versus non-MB, model 8 for EP versus non-EP, and model 7 for PA versus non-PA achieved the highest AUCs (0.97, 0.84, and 0.96 respectively). The parameters of all the TPOT models are described in detail in On-line Table 7. Then the best models were tested on the test set, and performances were as follows: For all 3 binary classifications, the AUC ranged from 0.84 to 0.94, accuracy ranged from 0.80 to 0.88, sensitivity ranged from 0.52 to 0.95, and specificity ranged from 0.81 to 0.93. The test performances for all 3 models are reported in detail in Table 2.

View this table:
  • View inline
  • View popup
Table 2:

Performance of binary classification across all models and experts in the test set

For manual expert optimization, the AUC scores from each combination of feature-selection method and classifier are shown in On-line Figs 3–5. The combinations of feature-selection method and classifier that achieved the highest AUCs on the training set for the 3 different binary classifications are as follows: Relief + support vector machine (AUC = 0.97) for MB versus non-MB, joint mutual information + neural network (AUC = 0.79) for EP versus non-EP, and CHSQ + GLM (AUC = 0.96) for PA versus non-PA. The models Relief + support vector machine, joint mutual information + neural network, and CHSQ + GLM with the best performance on the training set above were then tested on the final testing set. For all 3 binary classifications and their optimal feature-selection method-classifier combination, the AUC ranged from 0.70 to 0.98, accuracy ranged from 0.71 to 0.91, sensitivity ranged from 0.19 to 0.96, and specificity ranged from 0.88 to 0.95. The test performances for all 3 models are reported in detail in Table 2.

Expert Evaluation

For multiclass classification, expert 1 had an overall test accuracy of 0.58 (95% CI, 0.46–0.69) with sensitivities of 0.50–0.65 and specificities of 0.67–0.86 across the 3 tumor types. Expert 2 achieved an overall test accuracy of 0.50 (95% CI, 0.38–0.62) along with sensitivities of 0.43–0.57 and specificities of 0.66–0.80 across the 3 tumor types (Table 1). For binary classification, expert 1 achieved a test accuracy of 0.67–0.74 and expert 2 achieved a test accuracy of 0.64–0.68 across the 3 binary classifications. The sensitivity and specificity of each group are shown in Table 2.

Comparison of Radiomics Model Performance with Expert Evaluation

For multiclass classification, the TPOT model achieved significantly higher test accuracy compared with average expert MR imaging review (accuracy, 0.83 versus 0.54; P < .001). The TPOT model also had higher accuracy than the CHSQ + GLM model by the manual expert optimized pipeline (accuracy, 0.83 versus 0.74), but this difference was not statistically significant (P = .160). For binary classification, the TPOT models shared similar performance with the models by manual expert optimized pipeline in the classification of MB versus non-MB (accuracy, 0.85 versus 0.91; P = .247), EP versus non-EP (accuracy, 0.80 versus 0.71; P = .333), and PA versus non-PA (accuracy, 0.88 versus 0.86; P = .385). The TPOT models outperformed average expert MR imaging review in the classification of MB versus non-MB (accuracy, 0.85 versus 0.66; P < .001), EP versus non-EP (accuracy, 0.80 versus 0.71; P = .177), and PA versus non-PA (accuracy, 0.88 versus 0.71; P = .002). The receiver operating characteristic curves of all radiomics models compared with expert evaluations are shown in Figs 2 and 3). On-line Figure 6 depicts examples of agreement or disagreement between the TPOT model and expert review in multiclass classification of pediatric posterior fossa tumors.

FIG 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
FIG 2.

The receiver operating characteristic curves of the radiomic models compared with expert MR imaging evaluations in multiclass classification. Acc indicates accuracy.

FIG 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
FIG 3.

The receiver operating characteristic curves of the radiomic models compared with expert MR imaging evaluations in the binary classification for medulloblastoma versus nonmedulloblastoma (A), ependymoma versus nonependymoma (B), and pilocytic astrocytoma versus non-pilocytic astrocytoma (C). RELF indicates Relief; acc, accuracy; SVM, support vector machine; Nnet, neural network; JMI, joint mutual information.

DISCUSSION

Qualitative MR imaging review is critical for tumor diagnosis and evaluation. However, it often offers limited information regarding tumor type because pediatric posterior fossa tumors share similar appearances across conventional modalities.20,21 Many studies have suggested that ADC maps may be accurate in the differentiation of these tumors. Parameters such as mean ADC, minimal ADC, and tumor/normal ADC ratio were studied and used as thresholds for discrimination.22⇓⇓⇓-26 However, studies have shown overlap of ADC values among the different posterior fossa tumor types.27,28 Several studies have also explored the integration of other advanced MR technologies, including MR spectroscopy and MR perfusion. However, these advanced MR imaging modalities are not used on a routine basis and vary in performance when discriminating the pediatric posterior fossa tumors.2,21

In a previous retrospective study of 40 pediatric posterior fossa tumors (17 MBs, 16 PAs, and 7 EPs) by Rodriguez et al,29 models using support vector machine–based classifiers and ADC histogram features were trained and achieved average correct classifications of 95.8%, 96.9%, and 94.3% on MB, PA, and EP, respectively. The subsequent joint classifier for 3 tumors achieved correct classification of 91.4%. However, a major limitation of this study is the small data size used for analysis. Despite the good performance in the training set, the final models were not tested on a separate dataset. Furthermore, the model performance can vary when using different classifiers during the process. In this study, the authors tested only 1 classifier (support vector machine), which may result in bias and the possibility of missing a model with better performance.

In our study, 288 patients with the most common pediatric posterior fossa tumor subtypes were included, and autoML via TPOT was used to train the diagnostic models based on the T1WI contrast-enhanced sequence, T2WI, and ADC maps. We also compared the models by the automated TPOT pipeline with models by the manual expert optimized pipeline selected by the training models with different combinations of 13 different feature-selection methods and 10 classifiers. Both pipeline-generation methods were trained and then tested on separate data. Overall, radiomics models based on the TPOT pipeline outperformed MR imaging review by qualitative experts. Although the difference is not statistically significant, the TPOT model achieved better performance than models built by the manual expert optimized pipeline. TPOT has delivered a promising predictive performance in many other studies, including genomics applications17 and clinical metabolic profiling,30 and was found to be quite accurate in predicting H3 K27M mutation in gliomas, with accuracy ranging from 0.60 to 0.84 in the testing cohort.31 In contrast to standard manual ML analysis, TPOT evaluates the complete pipelines of feature selection and classification on the basis of their cross-validated score (mean squared error or balanced accuracy) in addition to model selection and hyperparameter optimization.14 The automatic TPOT is a valuable tool for construction of an optimal radiomics model without the need for extensive manual testing by experts in ML. Given no prior knowledge about a problem, TPOT frequently outperforms standard manual ML analyses.17,32

During model construction, a well-known issue with regard to AutoML is “overfitting,” in which improper tuning during model selection occurs.14 AutoML is easily affected by the data size; therefore, a small data size could result in overfitting. In our study, 288 samples were included to avoid the issue of high variance. Another issue is data imbalance, which can lead to a biased analysis and results for machine learning. Our study included relatively balanced data from 3 groups (111 versus 70 versus 107) and thus reduced the effect of this issue.

Despite the robustness of these results, the classification scheme in the present study does not obviate tumor diagnosis by histopathology, the criterion standard. Histopathology is needed for a truly confirmatory diagnosis, offers the opportunity to profile nontumor cells in the tumor mass that play an important role in the pathogenesis of these malignancies,33 and classifies tumors into molecular subgroups that are not appreciated by imaging. In an era of personalized medicine and therapeutic approaches like immunotherapy, these factors are especially important. However, we anticipate that the MR imaging–based classification scheme of this study may improve clinical care for pediatric posterior fossa tumors in several potential ways. First, an initial imaging-based classification may function as a noninvasive method to plan a surgical approach and resection extent, even before the invasive biopsy required for histopathology has occurred.29 Second, an earlier indication of a pediatric patient’s potential tumor subtype may expedite the planning of any potential neoadjuvant therapy.34 Third, the imaging features identified herein may complement, rather than replace, histopathology in the case of diagnosing pediatric fossa tumors with more complex histopathologic features.

There are several possible improvements to our study. First, there can be selection bias for our study because the patients were identified via a search of the pathology data base. Second, future studies can include advanced imaging modalities such as MR spectroscopy and MR perfusion, which may further improve model performance. However, adding these advanced MR images is also a limitation in that these sequences may not be available at every institution. Third, the criterion standard for comparisons was the diagnosis indicated on the final pathology report. Consensus reads by multiple pathologists were not performed due to challenges, including resource constraints and the multi-institutional nature of the study dataset. In addition, we were unable to analyze less frequent pediatric posterior fossa tumors such as atypical teratoid/rhabdoid tumor, hemangioblastoma, and diffuse midline gliomas due to their low sample size in our study cohort and the aforementioned risk of model overfitting. A similar methodology to the present study and a larger multi-institutional cohort may facilitate the future incorporation of these subtypes into our MR imaging–based classification schema. Finally, although there is a good likelihood of achieving greater performance by experts with Certificates of Added Qualification in pediatric neuroradiology, this comparison was unfortunately unable to be facilitated in the present study due to differences in standard practices around the world. As a point of reference, earlier studies have documented human accuracy of pediatric brain tumor classification as approximately between 70% and 80% by pediatric neuroradiologists with Certificates of Added Qualification.35,36

CONCLUSIONS

AutoML of TPOT based on conventional MR imaging can classify pediatric posterior fossa tumors with high accuracy compared with the manual expert optimized pipeline and expert radiologists.

Footnotes

  • This work was supported by National Natural Science Foundation of China grant under award No. 8181101287 to H. Bai; the National Natural Science Foundation of China (No. 61702559); the 111 project under grant No. B18059; the Planned Science and Technology Project of Hunan Province, China (No. 2017WK2074) to Chengzhang Zhu; the National Natural Science Foundation of China (81301988 to Li Yang); the Natural Science Foundation of Hunan Province for Young Scientists, China (grant No: 2018JJ3709 to Li Yang) and the National Science and Technology Major Project (No. 2018AAA0102102) to Beiji Zou.

  • Disclosures: Ken Chang—RELATED: Grant: National Institutes of Health, Comments: Research reported in this publication was supported by a training grant from the National Institute of Biomedical Imaging and Bioengineering of the National Institutes of Health under award number 5T32EB1680 and by the National Cancer Institute of the National Institutes of Health under award No. F30CA239407 to K. Chang. Raymond Y. Huang—UNRELATED: Grants/Grants Pending: research grant from Agios Pharmaceuticals.* Li Yang—RELATED: Grant: Natural Science Foundation of China/Natural Science Foundation of Hunan Province for Young Scientists, China, Comments: This funding came from a government granting agency.* *Money paid to the institution.

Indicates open access to non-subscribers at www.ajnr.org

References

  1. 1.↵
    1. Curtin SC,
    2. Miniño AM,
    3. Anderson RN
    . National Center for Health Statistics. Declines in cancer death rates among children and adolescents in the United States, 1999–2014. NCHS data brief, no 257, Sept 2016. https://www.cdc.gov/nchs/products/databriefs/db257.htm. Accessed March 21, 2020
  2. 2.↵
    1. Brandão LA,
    2. Poussaint TY
    . Posterior fossa tumors. Neuroimaging Clin N Am 2017;27:1–37 doi:10.1016/j.nic.2016.08.001 pmid:27889018
    CrossRefPubMed
  3. 3.↵
    1. O’Brien D,
    2. Caird J,
    3. Kennedy M, et al
    . Posterior fossa tumours in childhood: evaluation of presenting clinical features. Ir Med J 2001;94:52–53 pmid:11321174
    PubMed
  4. 4.↵
    1. Prasad KS,
    2. Ravi D,
    3. Pallikonda V, et al
    . Clinicopathological study of pediatric posterior fossa tumors. J Pediatr Neurosci 2017;12:245–50 doi:10.4103/jpn.JPN_113_16 pmid:29204199
    CrossRefPubMed
  5. 5.↵
    1. Bidiwala S,
    2. Pittman T
    . Neural network classification of pediatric posterior fossa tumors using clinical and imaging data. Pediatr Neurosurg 2004;40:8–15 doi:10.1159/000076571 pmid:15007223
    CrossRefPubMed
  6. 6.↵
    1. Becker LE
    . Pathology of pediatric brain tumors. Neuroimaging Clin N Am 1999;9:671–90 pmid:10517939
    PubMed
  7. 7.↵
    1. Rizzo S,
    2. Botta F,
    3. Raimondi S, et al
    . Radiomics: the facts and the challenges of image analysis. Eur Radiol Exp 2018;2:36 doi:10.1186/s41747-018-0068-z pmid:30426318
    CrossRefPubMed
  8. 8.↵
    1. Gillies RJ,
    2. Kinahan PE,
    3. Hricak H
    . Radiomics: images are more than pictures, they are data. Radiology 2016;278:563–77 doi:10.1148/radiol.2015151169 pmid:26579733
    CrossRefPubMed
  9. 9.↵
    1. Fan Y,
    2. Chen C,
    3. Zhao F, et al
    . Radiomics-based machine learning technology enables better differentiation between glioblastoma and anaplastic oligodendroglioma. Front Oncol 2019;9:1164 doi:10.3389/fonc.2019.01164 pmid:31750250
    CrossRefPubMed
  10. 10.↵
    1. Chen C,
    2. Ou X,
    3. Wang J, et al
    . Radiomics-based machine learning in differentiation between glioblastoma and metastatic brain tumors. Front Oncol 2019;9: 806 doi:10.3389/fonc.2019.00806 pmid:31508366
    CrossRefPubMed
  11. 11.↵
    1. Sidey-Gibbons JA,
    2. Sidey-Gibbons CJ
    . Machine learning in medicine: a practical introduction. BMC Med Res Methodol 2019;19:64 doi:10.1186/s12874-019-0681-4 pmid:30890124
    CrossRefPubMed
  12. 12.↵
    1. Parmar C,
    2. Grossmann P,
    3. Bussink J, et al
    . Machine learning methods for quantitative radiomic biomarkers. Sci Rep 2015;5:13087 doi:10.1038/srep13087 pmid:26278466
    CrossRefPubMed
  13. 13.↵
    1. Olson R,
    2. Urbanowicz R,
    3. Andrews P, et al
    . Automating biomedical data science through tree-based pipeline optimization. In: Proceedings of the 19th European Conference, EvoApplications 2016, Porto, Portugal; March 30 to April 1, 2016
  14. 14.↵
    1. Le TT,
    2. Fu W,
    3. Moore JH
    . Scaling tree-based automated machine learning to biomedical big data with a feature set selector. Bioinformatics 2020;36:250–56 doi:10.1093/bioinformatics/btz470 pmid:31165141
    CrossRefPubMed
  15. 15.↵
    1. Zwanenburg A,
    2. Vallieres M,
    3. Abdalah MA, et al
    . The image biomarker standardization initiative: standardized quantitative radiomics for high-throughput image-based phenotyping. Radiology 2020;295:328–38 doi:10.1148/radiol.2020191145 pmid:32154773
    CrossRefPubMed
  16. 16.↵
    1. Vallieres M,
    2. Kay-Rivest E,
    3. Perrin LJ, et al
    . Radiomics strategies for risk assessment of tumour failure in head-and-neck cancer. Sci Rep 2017;7:10117 doi:10.1038/s41598-017-10371-5 pmid:28860628
    CrossRefPubMed
  17. 17.↵
    1. Olson RS,
    2. Urbanowicz RJ,
    3. Andrews PC, et al
    . Automating biomedical data science through tree-based pipeline optimization. In: Proceedings of the 19th European Conference, EvoApplications 2016; Porto, Portugal; March 30 to April 1, 2016
  18. 18.↵
    1. Van Asch V
    . Macro-and micro-averaged evaluation measures [BASIC DRAFT]. https://pdfs.semanticscholar.org/1d10/6a2730801b6210a67f7622e4d192bb309303.pdf. Accessed March 21, 2020
  19. 19.↵
    1. Agresti A,
    2. Coull BA
    . Approximate is better than “exact” for interval estimation of binomial proportions. The American Statistician 1998;52:119–26 doi:10.1080/00031305.1998.10480550
    CrossRef
  20. 20.↵
    1. Poretti A,
    2. Meoded A,
    3. Huisman TA
    . Neuroimaging of pediatric posterior fossa tumors including review of the literature. J Magn Reson Imaging 2012;35:32–47 doi:10.1002/jmri.22722 pmid:21989968
    CrossRefPubMed
  21. 21.↵
    1. Kerleroux B,
    2. Cottier JP,
    3. Janot K, et al
    . Posterior fossa tumors in children: radiological tips & tricks in the age of genomic tumor classification and advance MR technology. J Neuroradiol 2020;47:46–53 doi:10.1016/j.neurad.2019.08.002 pmid:31541639
    CrossRefPubMed
  22. 22.↵
    1. Pierce T,
    2. Kranz PG,
    3. Roth C, et al
    . Use of apparent diffusion coefficient values for diagnosis of pediatric posterior fossa tumors. Neuroradiol J 2014;27:233–44 doi:10.15274/NRJ-2014-10027 pmid:24750714
    CrossRefPubMed
  23. 23.↵
    1. Zitouni S,
    2. Koc G,
    3. Doganay S, et al
    . Apparent diffusion coefficient in differentiation of pediatric posterior fossa tumors. Jpn J Radiol 2017;35:448–53 doi:10.1007/s11604-017-0652-9 pmid:28550357
    CrossRefPubMed
  24. 24.↵
    1. Rumboldt Z,
    2. Camacho DL,
    3. Lake D, et al
    . Apparent diffusion coefficients for differentiation of cerebellar tumors in children. AJNR Am J Neuroradiol 2006;27:1362–69 pmid:16775298
    Abstract/FREE Full Text
  25. 25.↵
    1. Gimi B,
    2. Cederberg K,
    3. Derinkuyu B, et al
    . Utility of apparent diffusion coefficient ratios in distinguishing common pediatric cerebellar tumors. Acad Radiol 2012;19:794–800 doi:10.1016/j.acra.2012.03.004 pmid:22513110
    CrossRefPubMed
  26. 26.↵
    1. Koral K,
    2. Alford R,
    3. Choudhury N, et al
    . Applicability of apparent diffusion coefficient ratios in preoperative diagnosis of common pediatric cerebellar tumors across two institutions. Neuroradiology 2014;56:781–88 doi:10.1007/s00234-014-1398-z pmid:24974083
    CrossRefPubMed
  27. 27.↵
    1. Jaremko JL,
    2. Jans LB,
    3. Coleman LT, et al
    . Value and limitations of diffusion-weighted imaging in grading and diagnosis of pediatric posterior fossa tumors. AJNR Am J Neuroradiol 2010;31:1613–16 doi:10.3174/ajnr.A2155 pmid:20538820
    Abstract/FREE Full Text
  28. 28.↵
    1. Schneider JF,
    2. Confort-Gouny S,
    3. Viola A, et al
    . Multiparametric differentiation of posterior fossa tumors in children using diffusion-weighted imaging and short echo-time 1H-MR spectroscopy. J Magn Reson Imaging 2007;26:1390–98 doi:10.1002/jmri.21185 pmid:17968955
    CrossRefPubMed
  29. 29.↵
    1. Rodriguez Gutierrez D,
    2. Awwad A,
    3. Meijer L, et al
    . Metrics and textural features of MRI diffusion to improve classification of pediatric posterior fossa tumors. AJNR Am J Neuroradiol 2014;35:1009–15 doi:10.3174/ajnr.A3784 pmid:24309122
    Abstract/FREE Full Text
  30. 30.↵
    1. Orlenko A,
    2. Moore JH,
    3. Orzechowski P, et al
    . Considerations for automated machine learning in clinical metabolic profiling: altered homocysteine plasma concentration associated with metformin exposure. Pac Symp Biocomput 2018;23: 460–71 pmid:29218905
    PubMed
  31. 31.↵
    1. Su X,
    2. Chen N,
    3. Sun H, et al
    . Automated machine learning based on radiomics features predicts H3 K27M mutation in midline gliomas of the brain. Neuro Oncol 2020;22:393–401 doi:10.1093/neuonc/noz184 pmid:31563963
    CrossRefPubMed
  32. 32.↵
    1. Olson RS,
    2. Cava W,
    3. Mustahsan Z, et al
    . Data-driven advice for applying machine learning to bioinformatics problems. Pac Symp Biocomput 2018;23: 192–203 pmid:29218881
    PubMed
  33. 33.↵
    1. Quail DF,
    2. Joyce JA
    . The microenvironmental landscape of brain tumors. Cancer Cell 2017;31:326–41 doi:10.1016/j.ccell.2017.02.009 pmid:28292436
    CrossRefPubMed
  34. 34.↵
    1. Massimino M,
    2. Cefalo G,
    3. Riva D, et al
    . Long-term results of combined preradiation chemotherapy and age-tailored radiotherapy doses for childhood medulloblastoma. J Neurooncol 2012;108:163–71 doi:10.1007/s11060-012-0822-7 pmid:22350379
    CrossRefPubMed
  35. 35.↵
    1. Manias KA,
    2. Gill SK,
    3. MacPherson L, et al
    . Diagnostic accuracy and added value of qualitative radiological review of 1H-magnetic resonance spectroscopy in evaluation of childhood brain tumors. Neurooncol Pract 2019;6:428–37 doi:10.1093/nop/npz010 pmid:31832213
    CrossRefPubMed
  36. 36.↵
    1. Manias K,
    2. Gill SK,
    3. Zarinabad N, et al
    . Evaluation of the added value of 1H-magnetic resonance spectroscopy for the diagnosis of pediatric brain lesions in clinical practice. Neurooncol Pract 2018;5:18–27 doi:10.1093/nop/npx005 pmid:29692921
    CrossRefPubMed
  • Received November 30, 2019.
  • Accepted after revision April 30, 2020.
  • © 2020 by American Journal of Neuroradiology
View Abstract
PreviousNext
Back to top

In this issue

American Journal of Neuroradiology: 41 (7)
American Journal of Neuroradiology
Vol. 41, Issue 7
1 Jul 2020
  • Table of Contents
  • Index by author
  • Complete Issue (PDF)
Advertisement
Print
Download PDF
Email Article

Thank you for your interest in spreading the word on American Journal of Neuroradiology.

NOTE: We only request your email address so that the person you are recommending the page to knows that you wanted them to see it, and that it is not junk mail. We do not capture any email address.

Enter multiple addresses on separate lines or separate them with commas.
Automatic Machine Learning to Differentiate Pediatric Posterior Fossa Tumors on Routine MR Imaging
(Your Name) has sent you a message from American Journal of Neuroradiology
(Your Name) thought you would like to see the American Journal of Neuroradiology web site.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Cite this article
H. Zhou, R. Hu, O. Tang, C. Hu, L. Tang, K. Chang, Q. Shen, J. Wu, B. Zou, B. Xiao, J. Boxerman, W. Chen, R.Y. Huang, L. Yang, H.X. Bai, C. Zhu
Automatic Machine Learning to Differentiate Pediatric Posterior Fossa Tumors on Routine MR Imaging
American Journal of Neuroradiology Jul 2020, 41 (7) 1279-1285; DOI: 10.3174/ajnr.A6621

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
0 Responses
Respond to this article
Share
Bookmark this article
Automatic Machine Learning to Differentiate Pediatric Posterior Fossa Tumors on Routine MR Imaging
H. Zhou, R. Hu, O. Tang, C. Hu, L. Tang, K. Chang, Q. Shen, J. Wu, B. Zou, B. Xiao, J. Boxerman, W. Chen, R.Y. Huang, L. Yang, H.X. Bai, C. Zhu
American Journal of Neuroradiology Jul 2020, 41 (7) 1279-1285; DOI: 10.3174/ajnr.A6621
del.icio.us logo Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One
Purchase

Jump to section

  • Article
    • Abstract
    • ABBREVIATIONS:
    • MATERIALS AND METHODS
    • RESULTS
    • DISCUSSION
    • CONCLUSIONS
    • Footnotes
    • References
  • Figures & Data
  • Supplemental
  • Info & Metrics
  • Responses
  • References
  • PDF

Related Articles

  • PubMed
  • Google Scholar

Cited By...

  • Evolving Role and Translation of Radiomics and Radiogenomics in Adult and Pediatric Neuro-Oncology
  • Radiomics Can Distinguish Pediatric Supratentorial Embryonal Tumors, High-Grade Gliomas, and Ependymomas
  • Radiomic Phenotypes Distinguish Atypical Teratoid/Rhabdoid Tumors from Medulloblastoma
  • Crossref
  • Google Scholar

This article has not yet been cited by articles in journals that are participating in Crossref Cited-by Linking.

More in this TOC Section

Pediatric Neuroimaging

  • Frontal Paraventricular Cysts
  • Sodium MRI in Pediatric Brain Tumors
  • FRACTURE MR in Congenital Vertebral Anomalies
Show more Pediatric Neuroimaging

Functional

  • Kurtosis and Epileptogenic Tubers: A Pilot Study
  • Glutaric Aciduria Type 1: DK vs. Conventional MRI
  • Multiparametric MRI in PEDS Pontine Glioma
Show more Functional

Similar Articles

Advertisement

Indexed Content

  • Current Issue
  • Accepted Manuscripts
  • Article Preview
  • Past Issues
  • Editorials
  • Editor's Choice
  • Fellows' Journal Club
  • Letters to the Editor
  • Video Articles

Cases

  • Case Collection
  • Archive - Case of the Week
  • Archive - Case of the Month
  • Archive - Classic Case

More from AJNR

  • Trainee Corner
  • Imaging Protocols
  • MRI Safety Corner
  • Book Reviews

Multimedia

  • AJNR Podcasts
  • AJNR Scantastics

Resources

  • Turnaround Time
  • Submit a Manuscript
  • Submit a Video Article
  • Submit an eLetter to the Editor/Response
  • Manuscript Submission Guidelines
  • Statistical Tips
  • Fast Publishing of Accepted Manuscripts
  • Graphical Abstract Preparation
  • Imaging Protocol Submission
  • Evidence-Based Medicine Level Guide
  • Publishing Checklists
  • Author Policies
  • Become a Reviewer/Academy of Reviewers
  • News and Updates

About Us

  • About AJNR
  • Editorial Board
  • Editorial Board Alumni
  • Alerts
  • Permissions
  • Not an AJNR Subscriber? Join Now
  • Advertise with Us
  • Librarian Resources
  • Feedback
  • Terms and Conditions
  • AJNR Editorial Board Alumni

American Society of Neuroradiology

  • Not an ASNR Member? Join Now

© 2025 by the American Society of Neuroradiology All rights, including for text and data mining, AI training, and similar technologies, are reserved.
Print ISSN: 0195-6108 Online ISSN: 1936-959X

Powered by HighWire