Journal Article DKFZ-2024-01006

http://join2-wiki.gsi.de/foswiki/pub/Main/Artwork/join2_logo100x88.png
Reproducible Radiomics Features from Multi-MRI-Scanner Test-Retest-Study: Influence on Performance and Generalizability of Models.

 ;  ;  ;  ;  ;  ;  ;  ;  ;  ;  ;  ;  ;

2025
Wiley-Liss New York, NY

Journal of magnetic resonance imaging 61(2), 676-686 () [10.1002/jmri.29442]
 GO

This record in other databases:

Please use a persistent id in citations: doi:

Abstract: Radiomics models trained on data from one center typically show a decline of performance when applied to data from external centers, hindering their introduction into large-scale clinical practice. Current expert recommendations suggest to use only reproducible radiomics features isolated by multiscanner test-retest experiments, which might help to overcome the problem of limited generalizability to external data.To evaluate the influence of using only a subset of robust radiomics features, defined in a prior in vivo multi-MRI-scanner test-retest-study, on the performance and generalizability of radiomics models.Retrospective.Patients with monoclonal plasma cell disorders. Training set (117 MRIs from center 1); internal test set (42 MRIs from center 1); external test set (143 MRIs from center 2-8).1.5T and 3.0T; T1-weighted turbo spin echo.The task for the radiomics models was to predict plasma cell infiltration, determined by bone marrow biopsy, noninvasively from MRI. Radiomics machine learning models, including linear regressor, support vector regressor (SVR), and random forest regressor (RFR), were trained on data from center 1, using either all radiomics features, or using only reproducible radiomics features. Models were tested on an internal (center 1) and a multicentric external data set (center 2-8).Pearson correlation coefficient r and mean absolute error (MAE) between predicted and actual plasma cell infiltration. Fisher's z-transformation, Wilcoxon signed-rank test, Wilcoxon rank-sum test; significance level P < 0.05.When using only reproducible features compared with all features, the performance of the SVR on the external test set significantly improved (r = 0.43 vs. r = 0.18 and MAE = 22.6 vs. MAE = 28.2). For the RFR, the performance on the external test set deteriorated when using only reproducible instead of all radiomics features (r = 0.33 vs. r = 0.44, P = 0.29 and MAE = 21.9 vs. MAE = 20.5, P = 0.10).Using only reproducible radiomics features improves the external performance of some, but not all machine learning models, and did not automatically lead to an improvement of the external performance of the overall best radiomics model.Stage 2.

Keyword(s): feature selection ; generalizability ; machine learning ; multicenter ; radiomics ; reproducibility

Classification:

Note: EA:E010#LA:E230# / Volume 61, Issue2, February 2025, Pages 676-686

Contributing Institute(s):
  1. E010 Radiologie (E010)
  2. C060 Biostatistik (C060)
  3. E230 Medizinische Bildverarbeitung (E230)
  4. DKTK HD zentral (HD01)
Research Program(s):
  1. 315 - Bildgebung und Radioonkologie (POF4-315) (POF4-315)

Appears in the scientific report 2024
Database coverage:
Medline ; Clarivate Analytics Master Journal List ; Current Contents - Clinical Medicine ; DEAL Wiley ; Essential Science Indicators ; IF < 5 ; JCR ; NationallizenzNationallizenz ; SCOPUS ; Science Citation Index Expanded ; Web of Science Core Collection
Click to display QR Code for this record

The record appears in these collections:
Document types > Articles > Journal Article
Institute Collections > E010
Public records
Publications database

 Record created 2024-05-13, last modified 2025-07-31


Fulltext:
Download fulltext PDF Download fulltext PDF (PDFA)
Rate this document:

Rate this document:
1
2
3
 
(Not yet reviewed)