Evaluating LLM-Generated Topic Names via Text Reconstruction

Authors

  • Andriy Kosar Textgain / University of Antwerp
  • Mariia Butynets Textgain
  • Guy De Pauw Textgain
  • Walter Daelemans University of Antwerp

Abstract

Automatically generating topic names for texts using large language models (LLMs) has become an innovative approach to topic detection. However, evaluating the quality of these LLM-generated topic names remains challenging, particularly in assessing their semantic relevance to the texts and the correctness of the information they convey. To address this gap, we propose a novel evaluation method that leverages LLMs to reconstruct original texts from generated topic names, then compares the reconstructed texts to the original by measuring their similarity. Topic names that produce reconstructed texts more similar to original ones better convey the original text’s information. This method favors topic names that maintain essential information, minimizing issues like incorrectness and irrelevance. Our experiments show that the reconstruction-based evaluation aligns with human topic name evaluation. This novel method demonstrates its versatility for evaluating other LLM-generated semantic compressions, such as summaries, headlines, and keywords.

Downloads

Published

2025-07-15

How to Cite

Kosar, A., Butynets, M., De Pauw, G., & Daelemans, W. (2025). Evaluating LLM-Generated Topic Names via Text Reconstruction. Computational Linguistics in the Netherlands Journal, 14, 43–65. Retrieved from https://clinjournal.org/clinj/article/view/186

Issue

Section

Articles

Most read articles by the same author(s)

1 2 > >>