TY - GEN
T1 - Explainability in NLP model
T2 - 2023 Conference on Human Centered Artificial Intelligence - Education and Practice, HCAIep 2023
AU - Yong, Wan Yit
AU - Jaiswal, Rajesh
AU - Perez Tellez, Fernando
N1 - Publisher Copyright:
© 2023 ACM.
PY - 2023/12/14
Y1 - 2023/12/14
N2 - Fake news has found fertile ground on social media. A global health crisis such as COVID-19 further helps propagate fake news on social media. Much research has been done to develop AI systems that classify news as real or fake. However, there is a growing concern about trust in these AI systems. To this end, we attempt to improve the trustworthiness of AI text classification systems. We use tools to explore data, explain feature extraction techniques, interpret the ML models implemented, and explain the decision-making progress of AI systems. In this study, we compared five ML classifiers for our experiments: Naive Bayes, Support Vector Machines (SVMs), Logistic Regression, Decision Tree, and Random Forest. The models were trained on 10700 tweets containing 5,600 real and 5,100 fake tweets related to COVID-19. In comparison, the SVMs model performance was the best, with a detection accuracy of 0.93 and F1 scores of 0.94 and 0.93 for real and fake news, respectively. Global and local explanations are included to understand the overall model behavior, ensuring transparency and fostering confidence in AI users. We have chosen the SVMs model for the explanation section as it was the best model in this study.
AB - Fake news has found fertile ground on social media. A global health crisis such as COVID-19 further helps propagate fake news on social media. Much research has been done to develop AI systems that classify news as real or fake. However, there is a growing concern about trust in these AI systems. To this end, we attempt to improve the trustworthiness of AI text classification systems. We use tools to explore data, explain feature extraction techniques, interpret the ML models implemented, and explain the decision-making progress of AI systems. In this study, we compared five ML classifiers for our experiments: Naive Bayes, Support Vector Machines (SVMs), Logistic Regression, Decision Tree, and Random Forest. The models were trained on 10700 tweets containing 5,600 real and 5,100 fake tweets related to COVID-19. In comparison, the SVMs model performance was the best, with a detection accuracy of 0.93 and F1 scores of 0.94 and 0.93 for real and fake news, respectively. Global and local explanations are included to understand the overall model behavior, ensuring transparency and fostering confidence in AI users. We have chosen the SVMs model for the explanation section as it was the best model in this study.
KW - Covid-19 News
KW - Explainable AI(XAI)
KW - Fake News Detection
KW - Machine Learning Classifier
KW - Natural Language Processing(NLP)
UR - http://www.scopus.com/inward/record.url?scp=85183315833&partnerID=8YFLogxK
U2 - 10.1145/3633083.3633212
DO - 10.1145/3633083.3633212
M3 - Conference contribution
AN - SCOPUS:85183315833
T3 - ACM International Conference Proceeding Series
SP - 29
EP - 35
BT - HCAIep 2023 - Proceedings of the 2023 Conference on Human Centered Artificial Intelligence - Education and Practice
PB - Association for Computing Machinery
Y2 - 15 December 2023
ER -