Fine-Tuning BERT Models to Classify Misinformation on Garlic and COVID-19 on Twitter
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Kim, Myeong Gyu | - |
dc.contributor.author | Kim, Minjung | - |
dc.contributor.author | Kim, Jae Hyun | - |
dc.contributor.author | Kim, Kyungim | - |
dc.date.accessioned | 2022-06-09T21:40:17Z | - |
dc.date.available | 2022-06-09T21:40:17Z | - |
dc.date.created | 2022-06-09 | - |
dc.date.issued | 2022-05 | - |
dc.identifier.issn | 1661-7827 | - |
dc.identifier.uri | https://scholar.korea.ac.kr/handle/2021.sw.korea/141775 | - |
dc.description.abstract | Garlic-related misinformation is prevalent whenever a virus outbreak occurs. With the outbreak of COVID-19, garlic-related misinformation is spreading through social media, including Twitter. Bidirectional Encoder Representations from Transformers (BERT) can be used to classify misinformation from a vast number of tweets. This study aimed to apply the BERT model for classifying misinformation on garlic and COVID-19 on Twitter, using 5929 original tweets mentioning garlic and COVID-19 (4151 for fine-tuning, 1778 for test). Tweets were manually labeled as 'misinformation' and 'other.' We fine-tuned five BERT models (BERTBASE, BERTLARGE, BERTweet-base, BERTweet-COVID-19, and BERTweet-large) using a general COVID-19 rumor dataset or a garlic-specific dataset. Accuracy and F1 score were calculated to evaluate the performance of the models. The BERT models fine-tuned with the COVID-19 rumor dataset showed poor performance, with maximum accuracy of 0.647. BERT models fine-tuned with the garlic-specific dataset showed better performance. BERTweet models achieved accuracy of 0.897-0.911, while BERTBASE and BERTLARGE achieved accuracy of 0.887-0.897. BERTweet-large showed the best performance with maximum accuracy of 0.911 and an F1 score of 0.894. Thus, BERT models showed good performance in classifying misinformation. The results of our study will help detect misinformation related to garlic and COVID-19 on Twitter. | - |
dc.language | English | - |
dc.language.iso | en | - |
dc.publisher | MDPI | - |
dc.title | Fine-Tuning BERT Models to Classify Misinformation on Garlic and COVID-19 on Twitter | - |
dc.type | Article | - |
dc.contributor.affiliatedAuthor | Kim, Kyungim | - |
dc.identifier.doi | 10.3390/ijerph19095126 | - |
dc.identifier.scopusid | 2-s2.0-85128560926 | - |
dc.identifier.wosid | 000795237300001 | - |
dc.identifier.bibliographicCitation | INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH, v.19, no.9 | - |
dc.relation.isPartOf | INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH | - |
dc.citation.title | INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH | - |
dc.citation.volume | 19 | - |
dc.citation.number | 9 | - |
dc.type.rims | ART | - |
dc.type.docType | Article | - |
dc.description.journalClass | 1 | - |
dc.description.isOpenAccess | Y | - |
dc.description.journalRegisteredClass | scie | - |
dc.description.journalRegisteredClass | ssci | - |
dc.description.journalRegisteredClass | scopus | - |
dc.relation.journalResearchArea | Environmental Sciences & Ecology | - |
dc.relation.journalResearchArea | Public, Environmental & Occupational Health | - |
dc.relation.journalWebOfScienceCategory | Environmental Sciences | - |
dc.relation.journalWebOfScienceCategory | Public, Environmental & Occupational Health | - |
dc.subject.keywordAuthor | bidirectional encoder representations from transformers (BERT) | - |
dc.subject.keywordAuthor | COVID-19 | - |
dc.subject.keywordAuthor | garlic | - |
dc.subject.keywordAuthor | misinformation | - |
dc.subject.keywordAuthor | - |
Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.
(02841) 서울특별시 성북구 안암로 14502-3290-1114
COPYRIGHT © 2021 Korea University. All Rights Reserved.
Certain data included herein are derived from the © Web of Science of Clarivate Analytics. All rights reserved.
You may not copy or re-distribute this material in whole or in part without the prior written consent of Clarivate Analytics.