Clinical validation of deep learning algorithms for radiotherapy targeting of non-small-cell lung cancer: an observational study
Artificial intelligence (AI) and deep learning have shown great potential in streamlining clinical tasks. However, most studies remain confined to in silico validation in small internal cohorts, without external validation or data on real-world clinical utility. We developed a strategy for the clini...
Saved in:
Published in | The Lancet. Digital health Vol. 4; no. 9; pp. e657 - e666 |
---|---|
Main Authors | , , , , , , , , , , , , , , , |
Format | Journal Article |
Language | English |
Published |
England
01.09.2022
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | Artificial intelligence (AI) and deep learning have shown great potential in streamlining clinical tasks. However, most studies remain confined to in silico validation in small internal cohorts, without external validation or data on real-world clinical utility. We developed a strategy for the clinical validation of deep learning models for segmenting primary non-small-cell lung cancer (NSCLC) tumours and involved lymph nodes in CT images, which is a time-intensive step in radiation treatment planning, with large variability among experts.
In this observational study, CT images and segmentations were collected from eight internal and external sources from the USA, the Netherlands, Canada, and China, with patients from the Maastro and Harvard-RT1 datasets used for model discovery (segmented by a single expert). Validation consisted of interobserver and intraobserver benchmarking, primary validation, functional validation, and end-user testing on the following datasets: multi-delineation, Harvard-RT1, Harvard-RT2, RTOG-0617, NSCLC-radiogenomics, Lung-PET-CT-Dx, RIDER, and thorax phantom. Primary validation consisted of stepwise testing on increasingly external datasets using measures of overlap including volumetric dice (VD) and surface dice (SD). Functional validation explored dosimetric effect, model failure modes, test-retest stability, and accuracy. End-user testing with eight experts assessed automated segmentations in a simulated clinical setting.
We included 2208 patients imaged between 2001 and 2015, with 787 patients used for model discovery and 1421 for model validation, including 28 patients for end-user testing. Models showed an improvement over the interobserver benchmark (multi-delineation dataset; VD 0·91 [IQR 0·83-0·92], p=0·0062; SD 0·86 [0·71-0·91], p=0·0005), and were within the intraobserver benchmark. For primary validation, AI performance on internal Harvard-RT1 data (segmented by the same expert who segmented the discovery data) was VD 0·83 (IQR 0·76-0·88) and SD 0·79 (0·68-0·88), within the interobserver benchmark. Performance on internal Harvard-RT2 data segmented by other experts was VD 0·70 (0·56-0·80) and SD 0·50 (0·34-0·71). Performance on RTOG-0617 clinical trial data was VD 0·71 (0·60-0·81) and SD 0·47 (0·35-0·59), with similar results on diagnostic radiology datasets NSCLC-radiogenomics and Lung-PET-CT-Dx. Despite these geometric overlap results, models yielded target volumes with equivalent radiation dose coverage to those of experts. We also found non-significant differences between de novo expert and AI-assisted segmentations. AI assistance led to a 65% reduction in segmentation time (5·4 min; p<0·0001) and a 32% reduction in interobserver variability (SD; p=0·013).
We present a clinical validation strategy for AI models. We found that in silico geometric segmentation metrics might not correlate with clinical utility of the models. Experts' segmentation style and preference might affect model performance.
US National Institutes of Health and EU European Research Council. |
---|---|
Bibliography: | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 23 ObjectType-Undefined-3 Joint last authors AH, DSB, HJWLA, and RHM conceived and designed the study. AH, DSB, CVG, JMQ, HR, SP, AS, LCP, IP, ZY, BHK, DEK, and DC collected and processed data. RHM performed segmentations for the Harvard-RT1 dataset. DSB performed the model failure mode analysis. AH performed deep learning analyses and visualisations. AH, DSB, HJWLA, and RHM wrote the initial draft of the manuscript. AH created the figures. DSB, JMQ, HR, SP, AS, LCP, BHK, and RHM took part in the end-user testing. IP and DEK coordinated the end-user testing. AH and PJC conducted statistical tests. AH, DSB, and RHM directly accessed and verified the underlying data. HJWLA and RHM were responsible for the decision to submit the manuscript. All authors subsequently read and revised the manuscript, as well as read and approved the final version. All authors had access to all the data in the study. Contributors |
ISSN: | 2589-7500 2589-7500 |
DOI: | 10.1016/S2589-7500(22)00129-7 |