Platform logo
Explore Communities
27th International Conference on Science, Technology and Innovation Indicators (STI 2023) logo
27th International Conference on Science, Technology and Innovation Indicators (STI 2023)Community hosting publication
There is an updated version of this publication, open Version 2.
poster

Can tweets predict article retractions?

20/04/2023| By
Er-Te Er-Te Zheng,
+ 1
Zhichao Zhichao Fang
663 Views
0 Comments
Disciplines
Keywords
Abstract

This study explores the potential of using tweets to predict article retractions, by analyzing the Twitter mention data of retracted articles as the treatment group and unretracted articles that were matched as a control group. The results show that tweets could predict article retractions with an accuracy of 57%-60% by machine learning models. Sentiment analysis is not effective in predicting article retractions. The study sheds light on a novel method of detecting scientific misconduct in the early stage.

Preview automatically generated form the publication file.

Can tweets predict article retractions?

Er-Te Zheng*, Hui-Zhen Fu**, Zhichao Fang*,***

* zhengerte@ruc.edu.cn; z.fang@cwts.leidenuniv.nl

0000-0001-8759-3643; 0000-0002-3802-2227

School of Information Resource Management, Renmin University of China, China

** fuhuizhen@zju.edu.cn

0000-0002-1534-9374

Department of Information Resources Management, Zhejiang University, China

*** z.fang@cwts.leidenuniv.nl

0000-0002-3802-2227

Centre for Science and Technology Studies (CWTS), Leiden University, The Netherlands

This study explores the potential of using tweets to predict article retractions, by analyzing the Twitter mention data of retracted articles as the treatment group and unretracted articles that were matched as a control group. The results show that tweets could predict article retractions with an accuracy of 57%-60% by machine learning models. Sentiment analysis is not effective in predicting article retractions. The study sheds light on a novel method of detecting scientific misconduct in the early stage.

1. Introduction

Scientific misconduct and questionable research practices have become more prevalent in recent years, undermining the credibility of scientific research. Traditional methods for identifying problematic articles have focused primarily on text-based plagiarism (Eysenbach, 2000; Wager, 2011) and image manipulation (Koppers et al., 2017; Parker et al., 2022; Pflugfelder, 2022), but they are limited to detect more sophisticated forms of misconduct, such as data falsification, and authorship issues. To address this challenge, alternative sources of information can be explored, such as reader comments on social media. Twitter has emerged as a source of discussion on scientific articles, accounting for over 80% of comments across all platforms (Peng et al., 2022). Despite some research studying Twitter mentions of retracted articles (Bornmann & Haunschild, 2018; Haunschild & Bornmann, 2021), there is a lack of large-scale studies examining the potential of Twitter as a tool for detecting misconduct in scientific research. In this study, we aim to fill this gap by addressing the following research questions:

1. Are there sentimental differences between tweets about retracted and unretracted articles?

2. Can machine learning models help to predict article retractions through tweets?

2. Data and methods

2.1. Data

We collected a total of 9,364 retracted articles from the Web of Science (WoS) and Retraction Watch databases, which were published between 2012 and 2021. Out of these, we identified 3,628 articles (38.7%) that had been mentioned at least once on Twitter, as the treatment group for our analysis.

2.2. Methods

2.2.1. Coarsened Exact Matching (CEM)

To establish a control group, we employed the coarsened exact matching (CEM) technique to match 3,505 unretracted articles to 3,505 retracted articles based on the same issue of the same journal, with similar number of authors and tweets. We then assigned the retraction time lag of the retracted articles to their matched unretracted articles, thereby screening out pre-retraction tweets to predict whether the articles would be retracted. As a result, we obtained 15,383 tweets related to the retracted articles and 11,031 tweets related to the unretracted articles (hereinafter retracted tweets and unretracted tweets).

2.2.2. Similarity calculation

We calculated the Levenshtein distance between the tweet texts and the article title, to exclude the tweets simply echoing the title of the mentioned articles. Figure 1 shows the distribution of the Levenshtein distance for both retracted tweets and unretracted tweets. We excluded tweets with a similarity distance greater than 90. Finally, we obtained 10,932 retracted tweets and 6,962 unretracted tweets for further analysis.

Figure 1: Histogram of Levenshtein distance between tweet text and article title.

2.2.3. Sentiment analysis

Differences in sentiment between retracted tweets and unretracted tweets may potentially help to predict article retractions. To explore this, the study used the Textblob package in Python to calculate the sentiment polarity score (ranging from -1 to 1) and the sentiment subjectivity score (ranging from 0 to 1) of the tweets.

2.2.4. Machine learning prediction

Machine learning is a method that enable computers to learn from data and make predictions or decisions without explicit programming. This study aims to explore whether machine learning models can help to predict article retractions through tweets, and if possible, which model has the highest prediction accuracy. Specifically, we employed four classical machine learning models in our study, including Naive Bayes (NB), Random Forest (RF), Support Vector Machine (SVM), and Logistic Regression (LR).

3. Results and discussion

3.1. Sentiment analysis

The study calculated sentiment polarity and subjectivity scores of retracted and unretracted tweets (Figure 2). However, we found that there were no significant differences in sentiment scores between the two groups. The research suggests that it is challenging to use sentiment scores of tweets to predict article retractions.

Figure 2: Sentiment polarity and subjectivity scores of retracted and unretracted tweets.

3.2. Machine learning prediction

Table 1 lists the accuracy of each machine learning model in predicting article retractions. The accuracy results of the various machine learning models are similar, ranging between 57% and 60%, with Logistic Regression being the best. Overall, the results indicate that machine learning models can predict article retractions to some extent.

Table 1. Prediction accuracy of the models for tweet prediction of article retractions.

Model

Prediction accuracy

(%)

Retracted recall

(%)

Retracted precision

(%)

Unretracted recall

(%)

Unretracted precision

(%)

LR 60.28 55.83 56.88 64.06 63.08
RF 58.31 43.56 55.91 70.83 59.65
NB 58.03 69.33 53.30 48.44 65.03
SVM 57.18 17.18 62.22 91.15 56.45

4. Conclusions

Although we found that tweets could predict article retractions to some extent, the prediction accuracy still needs to be improved. Due to the complexity of users on social media and the entertainment nature of social media sharing behavior, people need to hold critical perspectives to scientific articles shared on social media, as they may have potential issues that need to be carefully considered.

Competing interests

The authors have no conflicts of interest to declare.

Funding information

This study was supported by the Soft Science Research Program of the Zhejiang Provincial Department of Science of Technology (No. 2021C35040).

References

Bornmann, L., & Haunschild, R. (2018). Allegation of scientific misconduct increases Twitter attention. Scientometrics, 115, 1097-1100.

Eysenbach, G. (2000). Report of a case of cyberplagiarism-and reflections on detecting and preventing academic misconduct using the Internet. Journal of medical internet Research, 2(1), e793.

Haunschild, R., & Bornmann, L. (2021). Can tweets be used to detect problems early with scientific papers? A case study of three retracted COVID-19/SARS-CoV-2 papers. Scientometrics, 126(6), 5181-5199.

Koppers, L., Wormer, H., & Ickstadt, K. (2017). Towards a Systematic Screening Tool for Quality Assurance and Semiautomatic Fraud Detection for Images in the Life Sciences. Sci Eng Ethics, 23(4), 1113-1128.

Parker, L., Boughton, S., Lawrence, R., et al. (2022). Experts identified warning signs of fraudulent research: a qualitative study to inform a screening tool. J Clin Epidemiol, 151, 1-17.

Peng, H., Romero, D. M., & Horvát, E.-Á. (2022). Dynamics of cross-platform attention to retracted papers. Proceedings of the National Academy of Sciences, 119(25), e2119086119.

Pflugfelder, E. H. (2022). Evidence Engines: Common Rhetorical Features of Fraudulent Academic Articles. Written Communication, 39(2), 303-331.

Wager, E. (2011). How journals can prevent, detect and respond to misconduct. Notfall+ Rettungsmedizin, 14(8), 613-615.

Figures (2)

Publication ImagePublication Image
Submitted by20 Apr 2023
Download Publication

No reviews to show. Please remember to LOG IN as some reviews may be only visible to specific users.

ReviewerDecisionType
User Avatar
Hidden Identity
Accepted
Peer Review
User Avatar
Hidden Identity
Major Revision
Peer Review