Difference between revisions of "Eisenstein et al 2011: Sparse Additive Generative Models of Text"

From Cohen Courses
Jump to navigationJump to search
Line 8: Line 8:
 
== Summary ==
 
== Summary ==
  
This [[Category::paper]] presents [[UsesMethod::sparse learning]] and [[UsesMethod::additive generative modeling]] approaches for [[UsesMethod::]], which is an alternative training method for general problems in [[AddressesProblem:: Probabilistic Graphical Models]] (e.g. possible applications include [[AddressesProblem::]], [[AddressesProblem::]], [[AddressesProblem::]], and [[AddressesProblem::]]).
+
This recent [[Category::paper]] presents [[UsesMethod::sparse learning]] and [[UsesMethod::additive generative modeling]] approaches for [[AddressesProblem::Topic model]]ing. This is an important alternative approach to [[UsesMethod::Latent Dirichlet Allocation]] (LDA) where sparsity and log-space additive modeling approaches are introduced.
  
 
== Brief Description of the method ==
 
== Brief Description of the method ==

Revision as of 19:43, 28 November 2011

Citation

Sparse Additive Generative Models of Text. Eisenstein, Ahmed and Xing. Proceedings of ICML 2011.

Online version

Eisenstein et al 2011

Summary

This recent paper presents sparse learning and additive generative modeling approaches for Topic modeling. This is an important alternative approach to Latent Dirichlet Allocation (LDA) where sparsity and log-space additive modeling approaches are introduced.

Brief Description of the method

This paper first formulates the problem as xxx, then discusses the use of and the use of xxx to. In this section, we will first summarize the method they use

Dataset and Experiment Settings

[[File:]]

Experimental Results

The authors performed three major experiments. The first experiment is the . The second experiment explores.

Exp

[[File:]]

Exp

[[File:]]

Exp

[[File:]]

Related Papers

This paper is related to many papers in three dimensions.

(1) .

(2) .

(3) .