In our study, we identified a total of 77 retracted articles in these three high-impact journals. Among the 77 retracted articles, 94.8% of first authors were related to one retracted article, only four (5.2%) authors were responsible for two retracted publications (Supplemental Appendix 1).
Hua F et al pointed out that high-impact journals, especially biomedical journals, policed articles they published by adopting the Committee on Publication Ethics (COPE) or similar policies [12]. Fang et al also concluded that high-impact journals may retract fraudulent publications more rapidly with their greater visibility and enhanced scrutiny [5]. Our study showed that the median time to retraction was 651 (range 2-8474) days in the three top journals: 1703 days for retraction because of fraud, and 290 days for error. Gaudino M et al observed a median time of 1.8 years in biomedical literature [13]. It seems that the time for retraction in high-impact journals was not significantly shorter in the present study. The time from publication to retraction varied among reasons. Our results corresponded to earlier studies that concluded those articles retracted for error took a shorter time than fraud to retract [5, 14, 15]. The phenomenon may be partly attributed to the time interval required to identify fraud [5]. Besides, articles are most likely to be retracted in the first year after publication and most retractions occur in the first several years [16]. Likewise, about two-fifths of retractions occurred in the first year after publication in the present study.
The number of citations of retracted articles was correlated with the time-to-retraction and thus with how widely the invalid findings were propagated [17]. Gaudino M et al also identified the median number of citations was 9 [13]. Besides, Chambers LM et al observed a media number of 8 in 2019 [4]. However, our study implied that the median number of citations was 79 (range 0-2677), which is significantly higher than these studies. It means it is easy for compromised scientific conclusions of these top journals to be disseminated and will bring devastating disaster. Hence, these high-impact journals need to take more rigorous methods to reduce the lag-time to retraction.
Seventy-seven articles originated from 19 countries. We found the greatest number of retracted articles within these three journals were in the United States (37.7%), followed by the United Kingdom (11.7%) and China (6.5%). Similarly, a previous study suggested that about 41% of retractions searched from Retraction Watch Database were from the United States and China [4]. Notably, China is second only to the United States in the volume of papers published in journals listed in the Science Citation Index. However, China only accounted for 6.5% of retractions in the three journals. There was a possible relationship between the number of retractions for each country and its research output [18]. Although the United States and China are the two countries with the largest research output, the United States leads in most subject areas in the Scimago Journal Rankings (1996–2020) [19]. It indicates that the bulk of its articles was published in high-impact journals. Hence, the number of retractions in the United States was much higher than that in China in these high-impact journals.
Two studies on Medline pointed error and fraud were the most common reason for retractions in high-impact journals [5, 18]. Likewise, error and fraud were also the lead reason in our study. Previous studies reported that the proportion of retraction attributed to scientific misconduct was about 50%, 62.2%, 67.4% respectively in all journals [11, 13, 5]. However, fraud was only involved in 24.7% (n = 19) of retractions in our study. Firstly, we deduce these top journals may be more proficient in identifying misconduct after authors submitting their manuscripts. Secondly, the definition of scientific misconduct varied among studies. We included fabrication, falsification, and plagiarism, which was defined as scientific misconduct by the United States government. Others may included fake peer reviews, forged authorship, duplication, lack or withdrawal of informed consent, etc additionally which were thought to be dishonest and unethical behaviors by scientists widely. Hence, future studies may attach importance to developing a practical and standardized classification of reasons for retractions that could improve the relevance and uniformity of the research regarding retracted papers. Because better understanding the root causes of retractions can not only help to change scientific culture but also prevent the public from losing trust in science [20, 21]. A fake peer review is also a frequent reason for retractions. A previous retrospective analysis confirmed that 32.8% of articles were retracted because of a fake review in 134 retracted articles [22]. Besides, Q.-H. Vuong et al concluded a fake peer review was the most common reason in a study on the basis of retraction data through February 2019 [17]. Unexpectedly, no retraction was because of a fake peer review in our datasets. It may suggest that these high-impact journals own an excellent system of peer review which is worthy for other publishers to follow.
There are some strengths in our study. It is the first report on the features of retractions in high-impact journals. We used the Retraction Watch Database which is the largest and most comprehensive database worldwide containing a large number of retracted papers in all disciplines. It adopts a unified taxonomy when dealing with retraction notices issued by various editors/authors with different forms, clarity, and level of detail [23]. We believe our analysis of the three top journals could serve as a reference for future investigation. We acknowledge that there are several limitations in the present study. Firstly, The quantity of retracted articles in our study is too small. Our retraction data are only based on the Retraction Watch Database, but not all the retracted articles were included in this database. Secondly, some articles were retracted for multiple reasons, but the classification of reasons we adopted is mutually exclusive. It means the conclusion may change if we assign the several reasons for one retraction into all corresponding categories. Thirdly, we did not identify which citations were post-retraction citations that were defined by Kim et al as “those occurring at least one year after the retraction” [24]. The result may be different if we adopt such a method to extract data. Although there are some defects, we hope the present study can attract enough attention among the biomedical community and bring benefits in improving the quality of biomedical literature.
We would like to give some suggestions to reduce the number of retractions in these high-impact journals. For the policymakers: First, reinforce integrity education, especially in countries with high research output. Second, improve the incentives in the academic community. Abritis A et al confirmed that some institutions worldwide give money rewards when researchers publish a certain number of articles in these high-impact journals [25]. For the publishers, develop new tools to help editors to discover certain types of errors. For the authors: First, it is wise to keep the raw data carefully and conduct a comprehensive check to the error of data, methods, analyses, text, results, and conclusions when they submit their manuscripts. Second, it will be advised to check the status of the referenced articles before publication.