This study explores the potential of using tweets to predict article retractions, by analyzing the Twitter mention data of retracted articles as the treatment group and unretracted articles that were matched as a control group. The results show that tweets could predict article retractions with an accuracy of 57%-60% by machine learning models. Sentiment analysis is not effective in predicting article retractions. The study sheds light on a novel method of detecting scientific misconduct in the early stage.
Can tweets predict article retractions?
Er-Te Zheng*, Hui-Zhen Fu**, Zhichao Fang*,***
* zhengerte@ruc.edu.cn; z.fang@cwts.leidenuniv.nl
0000-0001-8759-3643; 0000-0002-3802-2227
School of Information Resource Management, Renmin University of China, China
** fuhuizhen@zju.edu.cn
0000-0002-1534-9374
Department of Information Resources Management, Zhejiang University, China
*** z.fang@cwts.leidenuniv.nl
0000-0002-3802-2227
Centre for Science and Technology Studies (CWTS), Leiden University, The Netherlands
This study explores the potential of using tweets to predict article retractions, by analyzing the Twitter mention data of retracted articles as the treatment group and unretracted articles that were matched as a control group. The results show that tweets could predict article retractions with an accuracy of 57%-60% by machine learning models. Sentiment analysis is not effective in predicting article retractions. The study sheds light on a novel method of detecting scientific misconduct in the early stage.
Scientific misconduct and questionable research practices have become more prevalent in recent years, undermining the credibility of scientific research. Traditional methods for identifying problematic articles have focused primarily on text-based plagiarism (Eysenbach, 2000; Wager, 2011) and image manipulation (Koppers et al., 2017; Parker et al., 2022; Pflugfelder, 2022), but they are limited to detect more sophisticated forms of misconduct, such as data falsification, and authorship issues. To address this challenge, alternative sources of information can be explored, such as reader comments on social media. Twitter has emerged as a source of discussion on scientific articles, accounting for over 80% of comments across all platforms (Peng et al., 2022). Despite some research studying Twitter mentions of retracted articles (Bornmann & Haunschild, 2018; Haunschild & Bornmann, 2021), there is a lack of large-scale studies examining the potential of Twitter as a tool for detecting misconduct in scientific research. In this study, we aim to fill this gap by addressing the following research questions:
1. Are there sentimental differences between tweets about retracted and unretracted articles?
2. Can machine learning models help to predict article retractions through tweets?
We collected a total of 9,364 retracted articles from the Web of Science (WoS) and Retraction Watch databases, which were published between 2012 and 2021. Out of these, we identified 3,628 articles (38.7%) that had been mentioned at least once on Twitter, as the treatment group for our analysis.
To establish a control group, we employed the coarsened exact matching (CEM) technique to match 3,505 unretracted articles to 3,505 retracted articles based on the same issue of the same journal, with similar number of authors and tweets. We then assigned the retraction time lag of the retracted articles to their matched unretracted articles, thereby screening out pre-retraction tweets to predict whether the articles would be retracted. As a result, we obtained 15,383 tweets related to the retracted articles and 11,031 tweets related to the unretracted articles (hereinafter retracted tweets and unretracted tweets).
We calculated the Levenshtein distance between the tweet texts and the article title, to exclude the tweets simply echoing the title of the mentioned articles. Figure 1 shows the distribution of the Levenshtein distance for both retracted tweets and unretracted tweets. We excluded tweets with a similarity distance greater than 90. Finally, we obtained 10,932 retracted tweets and 6,962 unretracted tweets for further analysis.
Figure 1: Histogram of Levenshtein distance between tweet text and article title.
Differences in sentiment between retracted tweets and unretracted tweets may potentially help to predict article retractions. To explore this, the study used the Textblob package in Python to calculate the sentiment polarity score (ranging from -1 to 1) and the sentiment subjectivity score (ranging from 0 to 1) of the tweets.
Machine learning is a method that enable computers to learn from data and make predictions or decisions without explicit programming. This study aims to explore whether machine learning models can help to predict article retractions through tweets, and if possible, which model has the highest prediction accuracy. Specifically, we employed four classical machine learning models in our study, including Naive Bayes (NB), Random Forest (RF), Support Vector Machine (SVM), and Logistic Regression (LR).
The study calculated sentiment polarity and subjectivity scores of retracted and unretracted tweets (Figure 2). However, we found that there were no significant differences in sentiment scores between the two groups. The research suggests that it is challenging to use sentiment scores of tweets to predict article retractions.
Figure 2: Sentiment polarity and subjectivity scores of retracted and unretracted tweets.
Table 1 lists the accuracy of each machine learning model in predicting article retractions. The accuracy results of the various machine learning models are similar, ranging between 57% and 60%, with Logistic Regression being the best. Overall, the results indicate that machine learning models can predict article retractions to some extent.
Table 1. Prediction accuracy of the models for tweet prediction of article retractions.
Model | Prediction accuracy (%) |
Retracted recall (%) |
Retracted precision (%) |
Unretracted recall (%) |
Unretracted precision (%) |
---|---|---|---|---|---|
LR | 60.28 | 55.83 | 56.88 | 64.06 | 63.08 |
RF | 58.31 | 43.56 | 55.91 | 70.83 | 59.65 |
NB | 58.03 | 69.33 | 53.30 | 48.44 | 65.03 |
SVM | 57.18 | 17.18 | 62.22 | 91.15 | 56.45 |
Although we found that tweets could predict article retractions to some extent, the prediction accuracy still needs to be improved. Due to the complexity of users on social media and the entertainment nature of social media sharing behavior, people need to hold critical perspectives to scientific articles shared on social media, as they may have potential issues that need to be carefully considered.
Competing interests
The authors have no conflicts of interest to declare.
Funding information
This study was supported by the Soft Science Research Program of the Zhejiang Provincial Department of Science of Technology (No. 2021C35040).
Bornmann, L., & Haunschild, R. (2018). Allegation of scientific misconduct increases Twitter attention. Scientometrics, 115, 1097-1100.
Eysenbach, G. (2000). Report of a case of cyberplagiarism-and reflections on detecting and preventing academic misconduct using the Internet. Journal of medical internet Research, 2(1), e793.
Haunschild, R., & Bornmann, L. (2021). Can tweets be used to detect problems early with scientific papers? A case study of three retracted COVID-19/SARS-CoV-2 papers. Scientometrics, 126(6), 5181-5199.
Koppers, L., Wormer, H., & Ickstadt, K. (2017). Towards a Systematic Screening Tool for Quality Assurance and Semiautomatic Fraud Detection for Images in the Life Sciences. Sci Eng Ethics, 23(4), 1113-1128.
Parker, L., Boughton, S., Lawrence, R., et al. (2022). Experts identified warning signs of fraudulent research: a qualitative study to inform a screening tool. J Clin Epidemiol, 151, 1-17.
Peng, H., Romero, D. M., & Horvát, E.-Á. (2022). Dynamics of cross-platform attention to retracted papers. Proceedings of the National Academy of Sciences, 119(25), e2119086119.
Pflugfelder, E. H. (2022). Evidence Engines: Common Rhetorical Features of Fraudulent Academic Articles. Written Communication, 39(2), 303-331.
Wager, E. (2011). How journals can prevent, detect and respond to misconduct. Notfall+ Rettungsmedizin, 14(8), 613-615.
Zheng, E., Fu, H. & Fang, Z. (2023). Can tweets predict article retractions? [preprint]. 27th International Conference on Science, Technology and Innovation Indicators (STI 2023). https://doi.org/10.55835/644126a8763e8d2091a0cfdc
No reviews to show. Please remember to LOG IN as some reviews may be only visible to specific users.