Open access
Date
2023-12Type
- Conference Paper
ETH Bibliography
yes
Altmetrics
Abstract
Topic models help make sense of large text collections. Automatically evaluating their output and determining the optimal number of topics are both longstanding challenges, with no effective automated solutions to date. This paper evaluates the effectiveness of large language models (LLMs) for these tasks. We find that LLMs appropriately assess the resulting topics, correlating more strongly with human judgments than existing automated metrics. However, the type of evaluation task matters — LLMs correlate better with coherence ratings of word sets than on a word intrusion task. We find that LLMs can also guide users toward a reasonable number of topics. In actual applications, topic models are typically used to answer a research question related to a collection of texts. We can incorporate this research question in the prompt to the LLM, which helps estimate the optimal number of topics. Show more
Permanent link
https://doi.org/10.3929/ethz-b-000639383Publication status
publishedExternal links
Book title
Proceedings of the 2023 Conference on Empirical Methods in Natural Language ProcessingPages / Article No.
Publisher
Association for Computational LinguisticsEvent
Subject
Topic models; LLMsOrganisational unit
09627 - Ash, Elliott / Ash, Elliott
More
Show all metadata
ETH Bibliography
yes
Altmetrics