Peacock: Learning Long-Tail Topic Features for Industrial Applications

  title={Peacock: Learning Long-Tail Topic Features for Industrial Applications},
  author={Yi Wang and Xuemin Zhao and Zhenlong Sun and Hao Yan and Lifeng Wang and Zhihui Jin and Liubin Wang and Yang Gao and Ching Law and Jia Zeng},
  journal={ACM TIST},
Latent Dirichlet allocation (LDA) is a popular topic modeling technique in academia but less so in industry, especially in large-scale applications involving search engine and online advertising systems. A main underlying reason is that the topic models used have been too small in scale to be useful; for example, some of the largest LDA models reported in literature have up to 103 topics, which difficultly cover the long-tail semantic word sets. In this article, we show that the number of… CONTINUE READING
Highly Cited
This paper has 39 citations. REVIEW CITATIONS
Recent Discussions
This paper has been referenced on Twitter 5 times over the past 90 days. VIEW TWEETS


Publications citing this paper.
Showing 1-10 of 25 extracted citations


Publications referenced by this paper.
Showing 1-10 of 14 references


  • Ke Zhai, Jordan L. Boyd-Graber, Nima Asadi, Mohamad L. Alkhouja.
  • LDA: A flexible large scale topic modeling…
  • 2012
Highly Influential
5 Excerpts

Similar Papers

Loading similar papers…