Full Text:  <3015>

Summary:  <2051>

CLC number: TP391.1

On-line Access: 2024-08-27

Received: 2023-10-17

Revision Accepted: 2024-05-08

Crosschecked: 2015-05-07

Cited: 3

Clicked: 8172

Citations:  Bibtex RefMan EndNote GB/T7714

 ORCID:

Xi-ming Li

http://orcid.org/0000-0001-8190-5087

-   Go to

Article info.
Open peer comments

Frontiers of Information Technology & Electronic Engineering 

Accepted manuscript available online (unedited version)


Topic modeling for large-scale text data


Author(s):  Xi-ming Li, Ji-hong Ouyang, You Lu

Affiliation(s):  College of Computer Science and Technology, Jilin University, Changchun 130012, China; more

Corresponding email(s):  liximing86@gmail.com, ouyj@jlu.edu.cn

Key Words:  Latent Dirichlet allocation (LDA), Topic modeling, Online learning, Moving average


Share this article to: More <<< Previous Paper|Next Paper >>>


Abstract: 
This paper develops a novel online algorithm, namely moving average stochastic variational inference (MASVI), which applies the results obtained by previous iterations to smooth out noisy natural gradients. We analyze the convergence property of the proposed algorithm and conduct a set of experiments on two large-scale collections that contain millions of documents. Experimental results indicate that in contrast to algorithms named ‘stochastic variational inference’ and ‘SGRLD’, our algorithm achieves a faster convergence rate and better performance.

Open peer comments: Debate/Discuss/Question/Opinion

<1>

Please provide your name, email address and a comment





Journal of Zhejiang University-SCIENCE, 38 Zheda Road, Hangzhou 310027, China
Tel: +86-571-87952783; E-mail: cjzhang@zju.edu.cn
Copyright © 2000 - 2025 Journal of Zhejiang University-SCIENCE