In the present study, we developed a prediction model for N0 status in primary breast cancer using AI on mammographic images and routine preoperative clinicopathological characteristics. Several radiological variables were associated with N0 status, supporting the hypothesis that mammographic features can aid in preoperatively identifying these patients. Transpara variables improved discrimination when added to an exclusively preoperative clinicopathological model. The final model had an AUC of 0.695 (CI: 0.653–0.736). If an FNR of 10% is accepted, reflecting the accepted FNR of SLNB, the proposed model could putatively support the omission of 23.8% SLNB. Moreover, preoperative radiologic tumor size was closely associated with postoperative pathologic tumor size and was the strongest predictor of N0 status.
Pathologic and radiologic measurements on tumor size were strongly associated in this material (Supplementary Table 4, Additional File 2), which is in accordance with previous research. An evaluation between pathology, MRI, ultrasound, and mammography showed that mammographic tumor size was similar to the reported pathological tumor size but tended to be slightly underestimated [11]. In this material we observed that decreasing radiologic tumor size was associated with N0 status. Pathological tumor size is a strong predictor for SLN status and often included in prediction models, although it is a postoperative variable [10, 19, 20]. These and previous results indicate that mammographic tumor size could replace pathological tumor size assessed from the surgical specimen as a predictor for N0 status.
The first model (Figure 2), including variables proposed by Dihge et al. [10], displayed similar discriminatory capacity when including all cases or only complete radiological cases, indicating that the AUC of the first model can be compared with that of the final model. The final model (AUC 0.695) had a lower AUC than the first model (AUC 0.775), which was expected as strong predictors, that could only be determined through the evaluation of the surgical specimen, were excluded from the model to make it clinically useful in a preoperative setting. The AIC was lower for the final model than for the first model. The fourth, reduced preoperative clinicopathological, model had a slightly lower AUC than the final model. The higher discriminatory capacity of the final model for N0 status indicates that it could potentially be used in clinical practice to preoperatively determine the omission of SLNB in selected patients. Importantly, the results indicate a predictive value of radiological variables in a preoperative prediction model.
When elaborating on the radiological variables used in this study, it is important to note that Transpara is not intended to be used as a tool for prediction of SLN status, although this study also indicates a potential predictive value in this setting. Additional development of Transpara in this direction could improve its predictive ability for SLN status. The potential clinical use and definitions of medicolegal regulations regarding this type of diagnostic tool are debated and yet to be determined [35, 36]. Implementation of image analysis software in clinical practice could have a dual application as the software is designed to detect breast cancer in a screening setting and could therefore be integrated both into cancer detection and prediction of SLN status [37]. Implementing a prediction model would entail additional costs associated with procurement, whereas omission of SLNB would likely reduce costs associated with surgery, as has been revealed previously for the ANN model proposed by Dihge et al. [10, 38]. Omission could improve the quality of life and reduce postoperative morbidity, as revealed in the INSEMA trial, where arm morbidity was significantly less frequent in the untreated group than in the SLNB group [3]. Therefore, a preoperative prediction model for N0 status could be of great clinical relevance and high value for individual patients.
The clinicopathological ANN model proposed by Dihge et al. which included some postoperative variables, had an AUC of 0.740. They performed a MLR including the same variables for comparison with an AUC of 0.727, which is slightly lower than that of the proposed ANN model [10]. In comparison with the first model in the present paper, the lower discriminatory capacity of the MLR model could be due to differences in methods where cross-validation and techniques to reduce overfitting were used in the original article. A limitation of previous prediction models is that key variables can only be obtained postoperatively [6, 7, 10]. In other studies, this problem was circumvented by including radiological variables from different imaging modalities exclusively or in addition to clinicopathological variables. Liu et al. proposed an exclusively radiological ANN model using CECT [23]. However, this method is not feasible for clinical implementation as all patients with breast cancer generally do not undergo CECT during the initial routine workup, an argument which also applies to models that include MRI images [24, 25]. Given the wide implementation of mammography screening programs, mammographic images are available for all patients and can be used for preoperative diagnostics. Cen et al. proposed a model that included age, postoperative pathological tumor size and microcalcification density on mammographic images, resulting in a model with an AUC of 0.70 [19]. They revealed that a microcalcification density >20 cm2 was associated with a positive ALN status, similar to our findings. A negative association between the absence of calc clusters and N0 status was discovered in our study but was not included in the model proposed by Cen et al. Correlation with an unknown predictive variable could explain the association with the absence of calc clusters. Transpara’s assessment of calc clusters includes several features, such as cluster area, which can be compared to microcalcification density [39]. Yang et al. [21] created a prediction model for ALN status using a radiomic signature on mammography (n=147) and with an AUC of 0.88 in the validation cohort, indicating a promising predictive capacity of mammographic features for N0 status. A study by Hack et al. [20] evaluated mammographic density and postoperative clinicopathological variables for prediction of N0 status. No association was discovered between mammographic density and SLN status, similar to our findings.
A strength of this study was the relatively large cohort of 770 patients. All eligible cases during a four-year period were consecutively included, and the cohort should therefore be representative of the breast cancer population at Skåne University Hospital (Lund, Sweden) during the study period. This single- center approach makes it likely that all patients were examined identically, thus reducing the bias caused by differing assessment criteria between clinics. In addition, in this study, we captured comprehensive radiological data, used two AI-based systems, Transpara and LIBRA, and explored their association with SLN status. Another strength is the assessment of Transpara’s accuracy through cross-checking for the correct tumor location. All cases were included to resemble a clinical setting, regardless of the variable. Additionally, as missing data was handled by multiple imputation, all available cases were included.
A limitation of our study is that 65% of patients were N0 in this study, which is lower that the prevalence in more recent cohorts. This could be due to the difference in study time, as more breast cancers are being discovered earlier than in past decades. Another limitation is the exclusion of 45 cases due to missing images in the Transpara sub-cohort. No cause was identified despite repeated contact with technical support at the hospital and the PACS provider. The number of missing cases was relatively small; however, there were differences in patient and tumor characteristics between the missing and included cases, which could have affected the results. Other limitations include the data-driven categorization of Transpara variables and the histological type, as well as the lack of internal validation, which could have resulted in a slight overestimation of the AUC. Additionally, external validation is required to assess the actual predictive value of the final model proposed in this study. The performance of Transpara variables could also be biased owing to missing mammograms. The prediction model on which this study is based is an ANN model that captures non-linear associations and interactions, whereas the MLR model captures only linear effects on the log odds scale. This is a limitation and a strength, as the risk of overfitting is much lower with MLR. No interaction variables were included in the MLR models, as this would require a larger cohort. Considering these limitations, the results of our study should be interpreted with caution. Validation in an external cohort should be performed to further evaluate the proposed model.