Publications
On Smoothing and Inference for Topic Models
Asuncion, A.; Welling, M.; Smyth, P. & Teh, Y.
(2009) [pdf]
A Generic Approach to Topic Models
Heinrich, G.
Machine Learning and Knowledge Discovery in Databases 517-532 (2009) [pdf]
This article contributes a generic model of topic models. To define the problem space, general characteristics for this class
models are derived, which give rise to a representation of topic models as “mixture networks”, a domain-specific compactalternative to Bayesian networks. Besides illustrating the interconnection of mixtures in topic models, the benefit of thisrepresentation is its straight-forward mapping to inference equations and algorithms, which is shown with the derivation andimplementation of a generic Gibbs sampling algorithm.
Named Entity Resolution Using Automatically Extracted Semantic Information
Pilz, A.; Paaß, G. & Augustin, G. S.
(2009) [pdf]
On-line LDA: Adaptive Topic Models for Mining Text Streams with Applications to Topic Detection and Tracking.
AlSumait, L.; Barbará, D. & Domeniconi, C.
, 'ICDM', IEEE Computer Society, 3-12 (2008) [pdf]
Topic and role discovery in social networks with experiments on enron and academic email
McCallum, A.; Wang, X. & Corrada-Emmanuel, A.
Journal of Artificial Intelligence Research, 30() 249-272 (2007) [pdf]
Latent Semantic Analysis: A Road to Meaning
Steyvers, M. & Griffiths, T.
, Laurence Erlbaum, chapter Probabilistic topic models (2007)
Latent Dirichlet Co-Clustering
Shafiei, M. M. & Milios, E. E.
, 'ICDM '06: Proceedings of the Sixth International Conference on Data Mining', IEEE Computer Society, Washington, DC, USA, [http://dx.doi.org/10.1109/ICDM.2006.94], 542-551 (2006)
Clustering with Bregman Divergences.
Banerjee, A.; Merugu, S.; Dhillon, I. S. & Ghosh, J.
Journal of Machine Learning Research, 6() 1705-1749 (2005) [pdf]
Integrating Topics and Syntax
Griffiths, T. L.; Steyvers, M.; Blei, D. M. & Tenenbaum, J. B.
Saul, L. K.; Weiss, Y. & Bottou, Léon., ed., 'Advances in Neural Information Processing Systems 17', 17(), MIT Press, Cambridge, MA, 537-544 (2005)
Finding scientific topics
Griffiths, T. L. & Steyvers, M.
Proceedings of the National Academy of Sciences, 101(Suppl. 1) 5228-5235 (2004)
Learning in graphical models
Jordan, M. I.
2003
Variational extensions to EM and multinomial PCA
Buntine, W.
Lecture notes in computer science 23-34 (2002) [pdf]
Probabilistic inference in graphical models
Jordan, M. & Weiss, Y.
Handbook of neural networks and brain theory (2002) [pdf]
An introduction to graphical models
Murphy, K.
, Web(2001) [pdf]
An Introduction to Variational Methods for Graphical Models
Jordan, M. I.; Ghahramani, Z.; Jaakkola, T. S. & Saul, L. K.
Mach. Learn., 37(2) 183-233 (1999) [pdf]
Latent variable models
Bishop, C.
Learning in graphical models (1998) [pdf]
Learning in graphical models
Jordan, M.
1998, Kluwer Academic Publishers [pdf]
Operations for Learning with Graphical Models
Buntine, W. L.
Journal of Artificial Intelligence Research, 2() 159-225 (1994) [pdf]
This paper is a multidisciplinary review of empirical, statistical learning from a graphical model perspective. Well-known examples of graphical models include Bayesian networks, directed graphs representing a Markov chain, and undirected networks representing a Markov field. These graphical models are extended to model data analysis and empirical learning using the notation of plates. Graphical operations for simplifying and manipulating a problem are provided including decomposition, differentiation, and the manipulation of probability models from the exponential family. Two standard algorithm schemas for learning are reviewed in a graphical framework: Gibbs sampling and the expectation maximization algorithm. Using these operations and schemas, some popular algorithms can be synthesized from their graphical specification. This includes versions of linear regression, techniques for feed-forward networks, and learning Gaussian and discrete Bayesian networks from data. The paper conclu...
Maximum likelihood from incomplete data via the EM algorithm
Dempster, A. P.; Laird, N. M. & Rubin, D. B.
JOURNAL OF THE ROYAL STATISTICAL SOCIETY, SERIES B, 39(1) 1-38 (1977) [pdf]
A broadly applicable algorithm for computing maximum likelihood estimates from incomplete data is presented at various levels of generality. Theory showing the monotone behaviour of the likelihood and convergence of the algorithm is derived. Many examples are sketched, including missing value situations, applications to grouped, censored or truncated data, finite mixture models, variance component estimation, hyperparameter estimation, iteratively reweighted least squares and factor analysis.