The dataset is currently empty. Upload or create new data files. Then, you will be able to explore them in the Dataset Viewer.
ProxAnn Topic Models
ProxAnn Topic Models provides the trained topic models used in
PROXANN: Use-Oriented Evaluations of Topic Models and Document Clustering
(Hoyle et al., ACL 2025).
This collection includes 50-topic models for both the Bills (Adler & Wilkerson, 2008) and Wiki (Merity et al., 2017) corpora.
All source datasets are available at lcalvobartolome/proxann_data.
Overview
| Split | Path | Description |
|---|---|---|
bills_bertopic |
model-runs/bills/bertopic/ |
50-topic BERTopic model trained on Bills using proxann.topic_models.train.BERTopicTrainer (MiniLM-L6-v2 embeddings). |
bills_ctm |
model-runs/bills/ctm/ |
50-topic Contextualized Topic Model (CTM) trained on Bills, following Hoyle et al., 2022. |
bills_mallet |
model-runs/bills/mallet/ |
50-topic LDA–MALLET model trained on Bills, from Hoyle et al., 2022. |
wiki_bertopic |
model-runs/wiki/bertopic/ |
50-topic BERTopic model trained on Wiki using proxann.topic_models.train.BERTopicTrainer. |
wiki_ctm |
model-runs/wiki/ctm/ |
50-topic CTM model trained on Wiki, from Hoyle et al., 2022. |
wiki_mallet |
model-runs/wiki/mallet/ |
50-topic LDA–MALLET model trained on Wiki, from Hoyle et al., 2022. |
Repository Layout
model-runs/
├── bills/
│ ├── bertopic/
│ ├── ctm/
│ └── mallet/
└── wiki/
├── bertopic/
├── ctm/
└── mallet/
Each folder contains model-specific artifacts (see below).
Artifacts Summary
| Model Type | Core Files | Notes |
|---|---|---|
| BERTopic | betas.npy, thetas.npz |
Topic–word and doc–topic matrices; trained with proxann.topic_models.train.BERTopicTrainer. Optional: config.yaml, vocab.txt, document_topic_info.csv. |
| CTM | beta.npy, train.theta.npy |
Topic–word and doc–topic matrices from Hoyle et al., 2022. Optional: config.yml, test.theta.npy, topics.txt. |
| LDA–MALLET | beta.npy, doctopics.npz.npy |
Topic–word and doc–topic matrices from Hoyle et al., 2022. Optional: state.mallet.gz, topickeys.txt, inferencer.mallet. |
Notes
- Embeddings: BERTopic trained with
sentence-transformers/all-MiniLM-L6-v2and CTM withsentence-transformers/multi-qa-mpnet-base-dot-v1. - Topics: All models use 50 topics.
- CTM & MALLET: Directly adapted from Hoyle et al., 2022 experimental setup.
- Data: Tokens and embeddings from ProxAnn Data.
Related Resources
- ProxAnn GitHub Repository
- ProxAnn Data on Hugging Face
- Are Neural Topic Models Broken? (Hoyle et al., 2022)
- Bills Dataset — Adler & Wilkerson (2008)
- WikiText Dataset — Merity et al. (2017)
License & Attribution
Released under the MIT License. Text content derives from Wikipedia (Merity et al., 2017) and the Congressional Bills Project (Adler & Wilkerson, 2008). Please provide attribution when reusing these materials.
Citation
If you use this dataset, please cite:
@inproceedings{hoyle-etal-2025-proxann,
title = "{P}rox{A}nn: Use-Oriented Evaluations of Topic Models and Document Clustering",
author = "Hoyle, Alexander Miserlis and
Calvo-Bartolom{\'e}, Lorena and
Boyd-Graber, Jordan Lee and
Resnik, Philip",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://aclanthology.org/2025.acl-long.772/",
doi = "10.18653/v1/2025.acl-long.772",
pages = "15872--15897",
ISBN = "979-8-89176-251-0",
abstract = "Topic models and document-clustering evaluations either use automated metrics that align poorly with human preferences, or require expert labels that are intractable to scale. We design a scalable human evaluation protocol and a corresponding automated approximation that reflect practitioners' real-world usage of models. Annotators{---}or an LLM-based proxy{---}review text items assigned to a topic or cluster, infer a category for the group, then apply that category to other documents. Using this protocol, we collect extensive crowdworker annotations of outputs from a diverse set of topic models on two datasets. We then use these annotations to validate automated proxies, finding that the best LLM proxy is statistically indistinguishable from a human annotator and can therefore serve as a reasonable substitute in automated evaluations."
}
- Downloads last month
- 36