Annealing Paths for Topic Model Evaluation & Inference Algorithms

annealing paths for the evaluation of topic models n.w
1 / 45
Embed
Share

Explore annealing paths for topic model evaluation by James Foulds & Padhraic Smyth at the University of California, Irvine. Discover motivation, model extensions, structure, inference algorithms, and more for topic models with insightful images and descriptions.

  • Topic Models
  • Model Evaluation
  • Inference Algorithms
  • University Research
  • Data Science

Uploaded on | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author. If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

You are allowed to download the files provided on this website for personal or commercial use, subject to the condition that they are used lawfully. All files are the property of their respective owners.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author.

E N D

Presentation Transcript


  1. Annealing Paths for the Evaluation of Topic Models James Foulds Padhraic Smyth Department of Computer Science University of California, Irvine* *James Foulds has recently moved to the University of California, Santa Cruz

  2. Motivation Topic model extensions Structure, prior knowledge and constraints Sparse, nonparametric, correlated, tree-structured, time series, supervised, focused, determinantal Special-purpose models Authorship, scientific impact, political affiliation, conversational influence, networks, machine translation General-purpose models Dirichlet multinomial regression (DMR), sparse additive generative (SAGE) Structural topic model (STM) 2

  3. Motivation Topic model extensions Structure, prior knowledge and constraints Sparse, nonparametric, correlated, tree-structured, time series, supervised, focused, determinantal Special-purpose models Authorship, scientific impact, political affiliation, conversational influence, networks, machine translation General-purpose models Dirichlet multinomial regression (DMR), sparse additive generative (SAGE) Structural topic model (STM) 3

  4. Motivation Topic model extensions Structure, prior knowledge and constraints Sparse, nonparametric, correlated, tree-structured, time series, supervised, focused, determinantal Special-purpose models Authorship, scientific impact, political affiliation, conversational influence, networks, machine translation General-purpose models Dirichlet multinomial regression (DMR), sparse additive generative (SAGE), Structural topic model (STM), 4

  5. Motivation Inference algorithms for topic models Optimization EM, variational inference, collapsed variational inference, Sampling Collapsed Gibbs sampling, Langevin dynamics, Scaling to ``big data Stochastic algorithms, distributed algorithms, map reduce 5

  6. Motivation Inference algorithms for topic models Optimization EM, variational inference, collapsed variational inference, Sampling Collapsed Gibbs sampling, Langevin dynamics, Scaling to ``big data Stochastic algorithms, distributed algorithms, map reduce 6

  7. Motivation Inference algorithms for topic models Optimization EM, variational inference, collapsed variational inference, Sampling Collapsed Gibbs sampling, Langevin dynamics, Scaling up to ``big data Stochastic algorithms, distributed algorithms, map reduce, sparse data structures 7

  8. Motivation Which existing techniques should we use? Is my new model/algorithm better than previous methods? 8

  9. Evaluating Topic Models Training set Test set 9

  10. Evaluating Topic Models Topic model Training set Test set 10

  11. Evaluating Topic Models Predict: Topic model Training set Test set 11

  12. Evaluating Topic Models Predict: Log Pr() Topic model Training set Test set 12

  13. Evaluating Topic Models (Foulds et al., 2013) Fitting these models only took a few hours on a single core single core machine. Creating this plot required a cluster 13

  14. Why is this Difficult? For every held-out document d, we need to estimate We need to approximate possibly tens of thousands of intractable sums/integrals! 14

  15. Annealed Importance Sampling (Neal, 2001) Scales up importance sampling to high dimensional data, using MCMC Corrects for MCMC convergence failures using importance weights 15

  16. Annealed Importance Sampling (Neal, 2001) low temperature 16

  17. Annealed Importance Sampling (Neal, 2001) high temperature low temperature 17

  18. Annealed Importance Sampling (Neal, 2001) high temperature low temperature 18

  19. Annealed Importance Sampling (Neal, 2001) 19

  20. Annealed Importance Sampling (Neal, 2001) Importance samples from the target An estimate of the ratio of partition functions 20

  21. AIS for Evaluating Topic Models (Wallach et al., 2009) Draw from the prior Anneal towards The posterior 21

  22. AIS for Evaluating Topic Models (Wallach et al., 2009) Draw from the prior Anneal towards The posterior 22

  23. AIS for Evaluating Topic Models (Wallach et al., 2009) Draw from the prior Anneal towards The posterior 23

  24. AIS for Evaluating Topic Models (Wallach et al., 2009) Draw from the prior Anneal towards The posterior 24

  25. Insights 1. We are mainly interested in the relative performance of topic models 2. AIS can provide estimates of the ratio of partition functions of any two distributions that we can anneal between 25

  26. A standard application of Annealed Importance Sampling (Neal, 2001) high temperature low temperature 26

  27. The Proposed Method: Ratio-AIS Draw from Topic Model 2 Anneal towards Topic Model 1 medium temperature medium temperature 27

  28. The Proposed Method: Ratio-AIS Draw from Topic Model 2 Anneal towards Topic Model 1 medium temperature medium temperature 28

  29. The Proposed Method: Ratio-AIS Draw from Topic Model 2 Anneal towards Topic Model 1 medium temperature medium temperature 29

  30. Advantages of Ratio-AIS Ratio-AIS avoids several sources of Monte Carlo error for comparing two models. The standard method estimates the denominator of a ratio even though it is a constant (=1), uses different z s for both models, and is run twice, introducing Monte Carlo noise each time. An easy convergence check: anneal in the reverse direction to compute the reciprocal. 30

  31. Annealing Paths Between Topic Models Geometric average of the two distributions Convex combination of the parameters 31

  32. Efficiently Plotting Performance Per Iteration of the Learning Algorithm (Foulds et al., 2013) 32

  33. Insights 1. Fsf 2. 2sfd 3. We can select the AIS intermediate distributions to be distributions of interest 4. The sequence of models we reach during training is typically amenable to annealing The early models are often low temperature Each successive model is similar to the previous one 33

  34. Iteration-AIS Topic Model at Topic Model at Topic Model at Anneal from Prior Iteration 1 Iteration 2 Iteration N Wallach et al. Ratio AIS Ratio AIS Ratio AIS Re-uses all previous computation Warm starts More annealing temperatures, for free Importance weights can be computed recursively 34

  35. Iteration-AIS Topic Model at Topic Model at Topic Model at Anneal from Prior Iteration 1 Iteration 2 Iteration N Wallach et al. Ratio AIS Ratio AIS Ratio AIS Re-uses all previous computation Warm starts More annealing temperatures, for free Importance weights can be computed recursively 35

  36. Comparing Very Similar Topic Models (ACL Corpus) 36

  37. Comparing Very Similar Topic Models (ACL and NIPS) 100 90 80 70 Left to Right % Accuracy 60 Standard AIS 50 Ratio-AIS (geo.) 40 Ratio-AIS (geo., rev.) 30 Ratio-AIS (convex) 20 Ratio-AIS (convex, rev.) 10 0 NIPS (cheap) NIPS ACL (cheap) ACL (expensive) (expensive) 37

  38. Symmetric vs Asymmetric Priors (NIPS, 1000 temperatures or equiv.) Correlation with longer left-to-right run Variance of the estimate of relative log-likelihood 38

  39. Symmetric vs Asymmetric Priors (NIPS, 1000 temperatures or equiv.) Correlation with longer left-to-right run Variance of the estimate of relative log-likelihood 39

  40. Symmetric vs Asymmetric Priors (NIPS, 1000 temperatures or equiv.) Correlation with longer left-to-right run Variance of the estimate of relative log-likelihood 40

  41. Per-Iteration Evaluation, ACL Dataset 41

  42. Per-Iteration Evaluation, ACL Dataset 42

  43. Conclusions Use Ratio-AIS for detailed document-level analysis Run the annealing in both directions to check for convergence failures Use Left to Right for corpus-level analysis Use Iteration-AIS to evaluate training algorithms 43

  44. Future Directions The ratio-AIS and iteration-AIS ideas can potentially be applied to other models with intractable likelihoods or partition functions (e.g. RBMs, ERGMs) Other annealing paths may be possible Evaluating topic models remains an important, computationally challenging problem 44

  45. Thank You! Questions? 45

More Related Content