2.1 Subjective Causal Identification Strategy
Our 37-question survey was conducted automatically through an online link in English between January 10th and February 6th, 2022. The timing of our survey captured roughly two years of pandemic experience. Given that prior pandemics lasted 1-2 years, our study provides a holistic view of the impact of COVID-19 on graduate medical education [28]. We recruited the help of the International Society for Aesthetic Plastic Surgery (ISAPS) to reach plastic surgery residents around the world. ISAPS is the leading professional body for board-certified plastic surgeons with a network of residents in over 100 countries. To our knowledge, no societies in other specialties with comparable network currently exist. The survey was disseminated by 63 associations of plastic surgeons, including ISAPS, to their resident members via email and social media. The survey questions and the dissemination strategy are provided in the Supplementary Information Appendix. This project was approved by the Institutional Review Board (IRB) at Stanford University.
2.2 Study Design
The survey design follows a causal inference framework. Recovering the impact of COVID-19 at the resident level entails comparison of the resident’s outcomes in two alternate states of the world: with the pandemic and without. With standard data on realizations, a given resident is observed in only one state of the world (in our case, COVID-19=1 ). The alternate outcomes are counter- factual and unobserved. A large econometric and statistics literature studies the identification of counterfactual outcomes from realized data [29, 30].
We follow Aucejo et al. in directly asking individuals for their expected outcomes with and
without COVID-19 [31]. From the collected data, we directly calculate the resident-level subjective treatment effect. Our approach builds on a growing literature that uses subjective expectations to understand decision-making under uncertainty [32, 33, 34, 35]. The soundness of our approach relies on the key assumption that residents have well-formed expectations for outcomes in both the realized and the counterfactual state.
Plastic surgery residents who did not work in hospitals that treated COVID-19 patients serve as the control group for residents in hospitals with COVID-19 patients. We consider two groups of COVID-impacted residents: (1) residents in hospitals with COVID-19 patients who did not work in COVID-19 wards and (2) residents in hospitals with COVID-19 patients who worked in COVID-19 wards.
2.3 Outcomes
Our study has two primary sets of outcomes: the training inputs and the expected training outputs. Inputs include surgeries residents participated or scrubbed in, seminar or lectures attended, and independent study. Outputs include surgical skill, scientific knowledge, overall competence, and future professional prospects.
Residents reported the number of surgeries and seminars per week or month before and during the pandemic. We calculate the percentage change before and during the pandemic. Residents also reported the magnitude of change in their study time before and during the pandemic. We create a score variable that takes the values -1, -0.5, 0, 0.5, and 1 for the responses decreased significantly, decreasedslightly, didnotchange, increasedslightly, and increasedsignificantly, respectively. We also create a binary variable that takes the value 1
when the response was decreased slightly or decreased significantly. Another binary variable captures the responses increased slightly and increasedsignificantly.
We asked residents whether the impact of the pandemic on their surgical skill and scientific knowledge has been significantly negative, slightly negative, zero, slightly positive, or significantly positive. We also asked residents whether they expect, in the absence of any remedial measures, the pandemic would make them significantlyless, slightlyless, equally, slightlymore, or significantly more competent compared with residents who did not face the pandemic during their residency.
For the outputs of surgical skill, scientific knowledge, and overall competence, we construct three variables. The first is a score variable that assigns the values -1, -0.5, 0, 0.5, and 1 for the responses significantly less/negative, slightly less/negative, no impact, slightly more/positive, and significantly more/positive impact, respectively. The second variable is binary and takes the value one when the respondent replied slightly or significantly less/negative impact. The third variable is also binary and takes the value one when the respondent replied slightly or significantly more/positive impact. Respondents who reported slightly or significantly lower competence were asked whether they anticipate an impact on their professional future and development. We create a binary variable that takes the value one when the participant replied probably or definitely yes. We also asked residents who reported lower competence than plastic surgeons before them the reasons that contributed to their responses. We create a binary variable for each of the reasons provided for reported lower overall competence. Those variables allow us to causally attribute reported lower competence to pandemic-related and non-pandemic-related justifications.
2.4 Statistical Analysis
Our main analytic strategy is a simple-differences model that compares changes in learning inputs and expected outputs before and during the pandemic of residents in hospitals with COVID- 19 patients and those in hospitals without (specification (1) in the Supplementary Information Appendix). This approach filters out pandemic-related influences on residents across all hospitals. For instance, some plastic surgery procedures may have been postponed across all hospitals during the pandemic, regardless of whether they treated COVID-19 cases [36].
We use multivariate regression models for continuous outcomes. For binary outcomes, we use multivariate linear probability models. All models include adjustments for gender, whether the resident has dependents, and the different levels of training program quality.1 We directly asked respondents whether they believe that their training program prepares competent plastic surgeons.2 All analyses use heteroskedasticity-robust standard errors. Our analysis relies on the assumption that had the pandemic not occurred, the learning inputs and outputs of residents in hospitals that treated COVID-19 patients would have been comparable to those of residents in hospitals that did not treat COVID-19 patients.
The association between changes in learning inputs and expected learning outputs during the pandemic was also investigated. We implemented models with pairwise interaction terms between indicators for whether residents worked in COVID hospitals/wards and changes of learning inputs during the pandemic (specification (2) in the Supplementary Information Appendix). We use linear models for ease of interpretation of interaction terms, as is standard practice in differences analyses [37].
1Table S3 presents estimates when we do not control for reported training program quality.
2This question is intended to infer program quality regardless of the pandemic.
The Supplementary Information Appendix offers several heterogeneity analyses on the differen- tial impact of the pandemic on residents in hospitals with COVID-19 patients by year of training, quality of training program, hospital type, gender, whether the resident has dependents, whether the resident has prior experience in plastic surgery, and number of surgeries and seminars prior to the pandemic.