Difference between revisions of "Class meeting for 10-605 2013 LDA 2"

From Cohen Courses
Jump to navigationJump to search
Line 12: Line 12:
 
* [http://dl.acm.org/citation.cfm?id=2623756 Reducing the sampling complexity of topic models], Li, Ahmed, Ravi, & Smola, KDD 2014
 
* [http://dl.acm.org/citation.cfm?id=2623756 Reducing the sampling complexity of topic models], Li, Ahmed, Ravi, & Smola, KDD 2014
 
* [http://arxiv.org/abs/1412.1576 LightLDA: Big Topic Models on Modest Compute Clusters], Jinhui Yuan, Fei Gao, Qirong Ho, Wei Dai, Jinliang Wei, Xun Zheng, Eric P. Xing, Tie-Yan Liu, Wei-Ying Ma, 2015
 
* [http://arxiv.org/abs/1412.1576 LightLDA: Big Topic Models on Modest Compute Clusters], Jinhui Yuan, Fei Gao, Qirong Ho, Wei Dai, Jinliang Wei, Xun Zheng, Eric P. Xing, Tie-Yan Liu, Wei-Ying Ma, 2015
 +
 +
=== Things to Remember ===
 +
 +
* Why efficient sampling is important for LDAs
 +
* How sampling can be sped up for many topics by preprocessing the parameters of the distribution
 +
* How the storage used for LDA can be reduced by exploiting the fact that many words are rare.

Revision as of 17:56, 4 December 2015

This is one of the class meetings on the schedule for the course Machine Learning with Large Datasets 10-605 in Spring_2015.

Slides

Readings

Things to Remember

  • Why efficient sampling is important for LDAs
  • How sampling can be sped up for many topics by preprocessing the parameters of the distribution
  • How the storage used for LDA can be reduced by exploiting the fact that many words are rare.