Skip to content

Commit f6b8f4c

Browse files
Fixed citations
1 parent c2c1b2b commit f6b8f4c

1 file changed

Lines changed: 31 additions & 28 deletions

File tree

paper.bib

Lines changed: 31 additions & 28 deletions
Original file line numberDiff line numberDiff line change
@@ -51,6 +51,8 @@ @inproceedings{topmost
5151
abstract = "Topic models have a rich history with various applications and have recently been reinvigorated by neural topic modeling. However, these numerous topic models adopt totally distinct datasets, implementations, and evaluations. This impedes quick utilization and fair comparisons, and thereby hinders their research progress and applications. To tackle this challenge, we in this paper propose a Topic Modeling System Toolkit (TopMost). Compared to existing toolkits, TopMost stands out by supporting more extensive features. It covers a broader spectrum of topic modeling scenarios with their complete lifecycles, including datasets, preprocessing, models, training, and evaluations. Thanks to its highly cohesive and decoupled modular design, TopMost enables rapid utilization, fair comparisons, and flexible extensions of diverse cutting-edge topic models. Our code, tutorials, and documentation are available at https://github.com/bobxwu/topmost."
5252
}
5353

54+
@article{quantitative_text_analysis, title={Quantitative text analysis}, volume={4}, url={https://www.nature.com/articles/s43586-024-00302-w#citeas}, DOI={10.1038/s43586-024-00302-w}, number={1}, journal={Nature Reviews Methods Primers}, author={Nielbo, Kristoffer L. and Karsdorp, Folgert and Wevers, Melvin and Lassche, Alie and Baglini, Rebekah B. and Kestemont, Mike and Tahmasebi, Nina}, year={2024}, month=apr }
55+
5456
@inproceedings{stream,
5557
title = "{STREAM}: Simplified Topic Retrieval, Exploration, and Analysis Module",
5658
author = {Thielmann, Anton and
@@ -125,12 +127,12 @@ @misc{top2vec
125127
primaryClass={cs.CL}
126128
}
127129

128-
@inproceedings{prodlda,
129-
title={Autoencoding Variational Inference For Topic Models},
130-
author={Akash Srivastava and Charles Sutton},
131-
booktitle={International Conference on Learning Representations},
132-
year={2017},
133-
url={https://api.semanticscholar.org/CorpusID:29842525}
130+
@inproceedings{prodla,
131+
title={Autoencoding Variational Inference For Topic Models},
132+
author={Akash Srivastava and Charles Sutton},
133+
booktitle={International Conference on Learning Representations},
134+
year={2017},
135+
url={https://openreview.net/forum?id=BybtVK9lg}
134136
}
135137

136138
@article{scikit-learn,
@@ -180,31 +182,32 @@ @misc{ctm_docs
180182
author={Bianchi, Federico and Terragni, Silvia and Hovy, Dirk},
181183
title={Contextualized Topic Models — Contextualized Topic Models 2.5.0 documentation}, url={https://contextualized-topic-models.readthedocs.io/en/latest/introduction.html}, year={2020} }
182184

183-
@misc{fastopic,
184-
title={FASTopic: A Fast, Adaptive, Stable, and Transferable Topic Modeling Paradigm},
185-
author={Xiaobao Wu and Thong Nguyen and Delvin Ce Zhang and William Yang Wang and Anh Tuan Luu},
186-
year={2024},
187-
eprint={2405.17978},
188-
archivePrefix={arXiv},
189-
primaryClass={cs.CL},
190-
url={https://arxiv.org/abs/2405.17978},
185+
@inproceedings{fastopic,
186+
title={FASTopic: Pretrained Transformer is a Fast, Adaptive, Stable, and Transferable Topic Model},
187+
author={Wu, Xiaobao and Nguyen, Thong Thanh and Zhang, Delvin Ce and Wang, William Yang and Luu, Anh Tuan},
188+
booktitle={The Thirty-eighth Annual Conference on Neural Information Processing Systems},
189+
year={2024}
191190
}
192191

193-
@article{sentence_transformers,
194-
author = {Nils Reimers and
195-
Iryna Gurevych},
196-
title = {Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks},
197-
journal = {CoRR},
198-
volume = {abs/1908.10084},
199-
year = {2019},
200-
url = {http://arxiv.org/abs/1908.10084},
201-
eprinttype = {arXiv},
202-
eprint = {1908.10084},
203-
timestamp = {Thu, 26 Nov 2020 12:13:54 +0100},
204-
biburl = {https://dblp.org/rec/journals/corr/abs-1908-10084.bib},
205-
bibsource = {dblp computer science bibliography, https://dblp.org}
192+
@inproceedings{sentence_transformers,
193+
title = "Sentence-{BERT}: Sentence Embeddings using {S}iamese {BERT}-Networks",
194+
author = "Reimers, Nils and
195+
Gurevych, Iryna",
196+
editor = "Inui, Kentaro and
197+
Jiang, Jing and
198+
Ng, Vincent and
199+
Wan, Xiaojun",
200+
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)",
201+
month = nov,
202+
year = "2019",
203+
address = "Hong Kong, China",
204+
publisher = "Association for Computational Linguistics",
205+
url = "https://aclanthology.org/D19-1410/",
206+
doi = "10.18653/v1/D19-1410",
207+
pages = "3982--3992",
208+
abstract = "BERT (Devlin et al., 2018) and RoBERTa (Liu et al., 2019) has set a new state-of-the-art performance on sentence-pair regression tasks like semantic textual similarity (STS). However, it requires that both sentences are fed into the network, which causes a massive computational overhead: Finding the most similar pair in a collection of 10,000 sentences requires about 50 million inference computations ({\textasciitilde}65 hours) with BERT. The construction of BERT makes it unsuitable for semantic similarity search as well as for unsupervised tasks like clustering. In this publication, we present Sentence-BERT (SBERT), a modification of the pretrained BERT network that use siamese and triplet network structures to derive semantically meaningful sentence embeddings that can be compared using cosine-similarity. This reduces the effort for finding the most similar pair from 65 hours with BERT / RoBERTa to about 5 seconds with SBERT, while maintaining the accuracy from BERT. We evaluate SBERT and SRoBERTa on common STS tasks and transfer learning tasks, where it outperforms other state-of-the-art sentence embeddings methods."
206209
}
207-
210+
208211
@software{topicwizard,
209212
author = {Kardos, Márton},
210213
month = nov,

0 commit comments

Comments
 (0)