Purpose: A balance between preserving urinary continence and achievement of negative margins is of clinical relevance while implementary difficulty. Preoperatively accurate detection of prostate cancer (PCa) extracapsular extension (ECE) is thus crucial for determining appropriate treatment options. We aimed to develop and clinically validate an artificial intelligence (AI)-assisted tool for the detection of ECE in patients with PCa using multiparametric MRI.
Methods: 849 patients with localized PCa underwent multiparametric MRI before radical prostatectomy were retrospectively included from two medical centers. The AI tool was built on a ResNeXt network embedded with a spatial attention map of experts’ prior knowledges (PAGNet) from 596 training data sets. The tool was validated in 150 internal and 103 external data sets, respectively; and its clinical applicability was compared with expert-based interpretation and AI-expert interaction.
Results: An index PAGNet model using a single-slice image yielded the highest areas under the receiver operating characteristic curve (AUC) of 0.857 (95% confidence interval [CI], 0.827-0.884), 0.807 (95% CI, 0.735-0.867) and 0.728 (95% CI, 0.631-0.811) in the training, internal test and external test cohorts, compared to the conventional ResNeXt networks. For experts, the inter-reader agreement was observed in only 437/849 (51.5%) patients with a Kappa value 0.343. And the performance of two experts (AUC, 0.632 to 0.741 vs 0.715 to 0.857) was lower (paired comparison, all p values < 0.05) than that of AI assessment. When expert’ interpretations were adjusted by the AI assessments, the performance of both two experts was improved.
Conclusion: Our AI tool, showing improved accuracy, offers a promising alternative to human experts for imaging staging of PCa ECE using multiparametric MRI.
Figure 1
Figure 2
Figure 3
Figure 4
Figure 5
This is a list of supplementary files associated with this preprint. Click to download.
Fig. S1. Flowchart of the study population. RP, radical prostatectomy; mpMRI, multiparametric magnetic resonance imaging.
Fig. S2. Decision curve analysis of AI-based, expert-based, and expert-AI interaction grading approaches for predicting pathological ECE in combined internal and external validation cohort. The y-axis measures the net benefits, and the x-axis is the risk threshold. All models achieve clinical net benefit against treat-all/none-plan. Better clinical risk prediction was observed when using PAGNet or an expert-AI interaction approach as compared to expert-based grading. ECE, extracapsular extension; AI, artificial intelligence.
Fig. S3 Overview of the model development in this study. a. Data preprocessing: 3D mp-MR images were preprocessed to prepare the training data set with the generated attention map. b. Training stage: Ensemble learning with 5-fold cross-validation was used to develop independent AI models by online augmentation. c. Inference stage: The trained model was used to predict one/each slice of MR images to estimate the prediction of ECE. d. Model architecture: ResNeXt model with CBAM was used as the base model. We modified the ResNeXt by embedding the prior-attention in each bottleneck (PAGNet, red part) and the clinical features in the second FC layer (PAGNet+C, purple part). Block1 #32 denoted in the first bottleneck block, the number of all convolution filters was 32. ECE, extracapsular extension.
The process of the generated attention and the random selected examples. Radiological experts labeled the boundary of the prostate and lesion by referring the mp-MR images manually first. Then a prior-attention map was calculated based on these boundaries. The region with a higher value in the attention map denoted the higher risk region of occurring ECE. (a) denotes the calculation of the attention map and (b) denotes the random selected examples of the annotation (bottom) and the corresponding attention map (bottom). ECE, extracapsular extension.
Supplementary Section 1. Image Annotation: Prostate and PCa lesion segmentation was performed with an in-house software (Oncology Imaging Analysis version 2; Shanghai Key Laboratory. of MRI, ECNU, Shanghai, China) on T2WI, DWI, and ADC by two experienced genitourinary radiologists from the two participating institutions. In patients with prostatectomy, postsurgical ex vivo prostates were processed using a previously described protocol. Key steps included sectioning, digitization, and annotation of cancer regions by highly experienced genitourinary pathologists. The histopathological specimens were then assembled into pseudo-whole-mount sections and co-registered to the MRI images using a previously described registration method. In this way, regions of annotated PCa were mapped onto the images to produce the ground truth maps. A central challenge in image labeling is the presence of ambiguous regions, where the true tumor boundary cannot be deduced precisely from the image, and thus multiple equally plausible interpretations exist. To fill this gap, the ROI of each lesion was drawn twice by each of two independent radiologists. Regional identification overlapping in two instances was identified as the authorized ROI of the targeted lesion. For the patient with multiple PCa lesions, only a leading lesion was annotated, referring to those with the higher Prostate Imaging and Reporting and Data System (PI-RADS) version 2.1 (v2.1) score or larger diameter if the score was the same. 2. Image preprocess: An open-source Elastix software (v. 4.10) referring to the suggested parameter file “par0001bspline16” was used for image registration [21]. Aligned images of T2WI, DWI, and ADC with the ROIs of prostate and PCa lesion were resampled to the inner-resolution of 0.5 × 0.5 mm2 by the Bicubic method or linear method (for ROIs). Then the slice including the largest area of the tumor lesion was extracted and cropped a patch with a size of 200×200 with the center of the prostate to increase computation efficiency and focusing on the targeted lesion. All patches of T2WI, ADC, and DWI were normalized by Z-score to make the scale similar before importing into the model. 3. Architecture of Network: A ResNeXt-based model with a convolutional block attention module (CBAM) was used to analyze the multiparametric images by concatenated use of high-resolution T2WI, high-b value (1500 s2/mm) DWI and ADC. Four bottleneck blocks included 3/4/4/3 bottlenecks with 16/32/64/96 filters. In each bottleneck, CBAM was used to self-learn the valid features from a channel-attention module and a spatial-attention module. At the end of the 4th block, we applied two fully connected (FC) layers with 256 nodes and 32 nodes. In the first FC layer, we applied batch-normalization and parametric rectified linear units (ReLU) to extract features. And in the second FC layer, we implemented a sigmoid function to predict the risk of ECE. The details of the ResNeXt was shown in Fig. S3d. Additionally, to guide the ResNeXt network to emulate the judgments of experts who provided the labels of the targeted lesion, we brought in a prior-attention guide (PAGNet) unit, that guides the model to form a computational map highly embedding the expert’ interpretations. To construct this expert-gated network, we first generated a prior-attention map based on the ROIs of PCa and prostate. The attention value of one voxel was set according to its location: a) if the voxel is localized in the intratumor region and out of the prostate counters, we set an attention value of 100%; b) if the voxel is localized in intratumor regions and in the prostate, we measured the distance (DP) from this voxel to the surface of the prostate. The attention value was given to r/D_P, in which r denoted the inner-resolution weighted a coefficient. Then the attention region extended outside with value attenuating as a ratio of r. Examples of the generated attention map were shown in Fig. S4. Then we embedded the above generated-attention map in the CBAM to guide the model in focusing on the PCa region. In CBAM, spatial attention was self-learned and generated the weighted coefficients to focus on the spatial region. The self-learned attention and our generated attention were implemented on all channels of the corresponding feature layers, respectively. Then we concatenated the scaled results and applied the following process (red part of Fig. S3d). This process can utilize both information of the data-driven self-learned features and the annotated prior-generated attention to guide the model focus on the candidate region. 4. Integration of PAGNet and Clinical Identifications: Last, to evaluate the integrative effects of clinical factors on the deep diagnostic networks for improving diagnostic performance. The PSA, age, biopsy Gleason score, percentage of positive cores, and biopsy perineural invasion were added to the PAGNet model, namely PAGNet + C, in which, clinical information was directly added to the second FC layer (32 nodes) of PAGNet by increasing the number of neurons (purple part of Fig. S3d). 5. Model Training and Inference Approach: In the training stage, we used an ensemble learning approach with 5-fold cross-validation to develop a robust model. The cases of training cohort were split into 5 parts, and we used 4 parts as the train data set and the remained 1 part as the validation. For each fold, We first balanced the data set by up-sampling the positive cases randomly and initialized the parameters of the model by the He method. Adam with an initial step of 0.001 was used as the optimizer and the negative log-likelihood loss was used as the loss function. In the first FC layer, we also used a dropout with a ratio of 0.5 in the training process. In each epoch, we balanced the positive and negative samples by random up-sampling. The batch size was set to 24. During the training process, we monitor the loss of the validation data set. If the loss does not decrease in 10 contiguous epochs, we reduced the step of the optimizer to half. If the loss does not decrease in 50 contiguous epochs, we stopped the training. In each fold of training, we applied an online augmentation strategy on the training cohort, which meant that we generated a random transform on each batch of the data set. The transform included rotation, flip, stretch, zoom, and elastic transformations. We also added random Gaussian noise and bias filed of MR images to simulate scanning images to make the model more robust. At last, we developed 5 independent models. (Fig. S3b) In the inference stage, we used the trained 5 models to predict one image and average the prediction to get the final predicted value. Additionally, since one case included several slices with PCa, we could use the model to predict the slice with the largest area of the PCa which was the same as the preprocessing, termed as PAGNet-OneSlice. Or we could predict all slices including PCa lesion and used the maximum value as the final prediction of ECE for this case, termed as PAGNet-MultSlices (Fig. S3c). 6. Package of Model Implementation All above was implemented on Ubuntu 18.04 by 2 NVIDIA Titan X Graphic Cards. The AI model was developed by Python 3.7 with the PyTorch 1.4.1.
Loading...
Posted 19 Mar, 2021
Received 05 Mar, 2021
Invitations sent on 05 Mar, 2021
On 03 Mar, 2021
On 03 Mar, 2021
Posted 19 Mar, 2021
Received 05 Mar, 2021
Invitations sent on 05 Mar, 2021
On 03 Mar, 2021
On 03 Mar, 2021
Purpose: A balance between preserving urinary continence and achievement of negative margins is of clinical relevance while implementary difficulty. Preoperatively accurate detection of prostate cancer (PCa) extracapsular extension (ECE) is thus crucial for determining appropriate treatment options. We aimed to develop and clinically validate an artificial intelligence (AI)-assisted tool for the detection of ECE in patients with PCa using multiparametric MRI.
Methods: 849 patients with localized PCa underwent multiparametric MRI before radical prostatectomy were retrospectively included from two medical centers. The AI tool was built on a ResNeXt network embedded with a spatial attention map of experts’ prior knowledges (PAGNet) from 596 training data sets. The tool was validated in 150 internal and 103 external data sets, respectively; and its clinical applicability was compared with expert-based interpretation and AI-expert interaction.
Results: An index PAGNet model using a single-slice image yielded the highest areas under the receiver operating characteristic curve (AUC) of 0.857 (95% confidence interval [CI], 0.827-0.884), 0.807 (95% CI, 0.735-0.867) and 0.728 (95% CI, 0.631-0.811) in the training, internal test and external test cohorts, compared to the conventional ResNeXt networks. For experts, the inter-reader agreement was observed in only 437/849 (51.5%) patients with a Kappa value 0.343. And the performance of two experts (AUC, 0.632 to 0.741 vs 0.715 to 0.857) was lower (paired comparison, all p values < 0.05) than that of AI assessment. When expert’ interpretations were adjusted by the AI assessments, the performance of both two experts was improved.
Conclusion: Our AI tool, showing improved accuracy, offers a promising alternative to human experts for imaging staging of PCa ECE using multiparametric MRI.
Figure 1
Figure 2
Figure 3
Figure 4
Figure 5
Loading...