Back to Search Start Over

Expansive data, extensive model: Investigating discussion topics around LLM through unsupervised machine learning in academic papers and news.

Authors :
Jung HS
Lee H
Woo YS
Baek SY
Kim JH
Source :
PloS one [PLoS One] 2024 May 31; Vol. 19 (5), pp. e0304680. Date of Electronic Publication: 2024 May 31 (Print Publication: 2024).
Publication Year :
2024

Abstract

This study presents a comprehensive exploration of topic modeling methods tailored for large language model (LLM) using data obtained from Web of Science and LexisNexis from June 1, 2020, to December 31, 2023. The data collection process involved queries focusing on LLMs, including "Large language model," "LLM," and "ChatGPT." Various topic modeling approaches were evaluated based on performance metrics, including diversity and coherence. latent Dirichlet allocation (LDA), nonnegative matrix factorization (NMF), combined topic models (CTM), and bidirectional encoder representations from Transformers topic (BERTopic) were employed for performance evaluation. Evaluation metrics were computed across platforms, with BERTopic demonstrating superior performance in diversity and coherence across both LexisNexis and Web of Science. The experiment result reveals that news articles maintain a balanced coverage across various topics and mainly focus on efforts to utilize LLM in specialized domains. Conversely, research papers are more concise and concentrated on the technology itself, emphasizing technical aspects. Through the insights gained in this study, it becomes possible to investigate the future path and the challenges that LLMs should tackle. Additionally, they could offer considerable value to enterprises that utilize LLMs to deliver services.<br />Competing Interests: The authors have declared that no competing interests exist.<br /> (Copyright: © 2024 Jung et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.)

Details

Language :
English
ISSN :
1932-6203
Volume :
19
Issue :
5
Database :
MEDLINE
Journal :
PloS one
Publication Type :
Academic Journal
Accession number :
38820285
Full Text :
https://doi.org/10.1371/journal.pone.0304680