Publications
Latent Dirichlet Bayesian Co-Clustering
Wang, P.; Domeniconi, C. & Laskey, K.
Machine Learning and Knowledge Discovery in Databases 522-537 (2009) [pdf]
Co-clustering has emerged as an important technique for mining contingency data matrices. However, almost all existing co-clustering
gorithms are hard partitioning, assigning each row and column of the data matrix to one cluster. Recently a Bayesian co-clusteringapproach has been proposed which allows a probability distribution membership in row and column clusters. The approach usesvariational inference for parameter estimation. In this work, we modify the Bayesian co-clustering model, and use collapsedGibbs sampling and collapsed variational inference for parameter estimation. Our empirical evaluation on real data sets showsthat both collapsed Gibbs sampling and collapsed variational inference are able to find more accurate likelihood estimatesthan the standard variational Bayesian co-clustering approach.
Bayesian Co-clustering.
Shan, H. & Banerjee, A.
, 'ICDM', IEEE Computer Society, 530-539 (2008) [pdf]
Collapsed Variational Inference for HDP
Teh, Y. W.; Kurihara, K. & Welling, M.
, 'Advances in Neural Information Processing Systems', 20() (2008)
Bayesian approach to clustering real value, categorical and network
data: solution via variational methods
Vazquez, A.
(2008) [pdf]
Data clustering, including problems such as finding network communities, can
put into a systematic framework by means of a Bayesian approach. The
plication of Bayesian approaches to real problems can be, however, quite
allenging. In most cases the solution is explored via Monte Carlo sampling or
riational methods. Here we work further on the application of variational
thods to clustering problems. We introduce generative models based on a
dden group structure and prior distributions. We extend previous attends by
ynes, and derive the prior distributions based on symmetry arguments. As a
se study we address the problems of two-sides clustering real value data and
ustering data represented by a hypergraph or bipartite graph. From the
riational calculations, and depending on the starting statistical model for
e data, we derive a variational Bayes algorithm, a generalized version of the
pectation maximization algorithm with a built in penalization for model
mplexity or bias. We demonstrate the good performance of the variational
yes algorithm using test examples.
Hierarchical dirichlet processes
Teh, Y.; Jordan, M.; Beal, M. & Blei, D.
Journal of the American Statistical Association, 101(476) 1566-1581 (2006) [pdf]
Clustering with Bregman Divergences.
Banerjee, A.; Merugu, S.; Dhillon, I. S. & Ghosh, J.
Journal of Machine Learning Research, 6() 1705-1749 (2005) [pdf]
Parameter estimation for text analysis
Heinrich, G.
Web: http://www. arbylon. net/publications/text-est. pdf (2005) [pdf]
Introduction to Bayesian learning
Hertzmann, A.
, 'SIGGRAPH '04: ACM SIGGRAPH 2004 Course Notes', ACM, New York, NY, USA, [http://doi.acm.org/10.1145/1103900.1103922], 22 (2004) [pdf]
Sophisticated computer graphics applications require complex models of appearance, motion, natural phenomena, and even artistic style. Such models are often difficult or impossible to design by hand. Recent research demonstrates that, instead, we can "learn" a dynamical and/or appearance model from captured data, and then synthesize realistic new data from the model. For example, we can capture the motions of a human actor and then generate new motions as they might be performed by that actor. Bayesian reasoning is a fundamental tool of machine learning and statistics, and it provides powerful tools for solving otherwise-difficult problems of learning about the world from data. Beginning from first principles, this course develops the general methodologies for designing learning algorithms and describes their application to several problems in graphics.
An introduction to graphical models
Murphy, K.
, Web(2001) [pdf]
Introduction to inference for Bayesian networks
Cowell, R.
Learning in graphical models 9-26 (1999) [pdf]
An Introduction to Variational Methods for Graphical Models
Jordan, M. I.; Ghahramani, Z.; Jaakkola, T. S. & Saul, L. K.
Mach. Learn., 37(2) 183-233 (1999) [pdf]
Advanced inference in Bayesian networks
Cowell, R.
1998, Learning in Graphical Models. MIT Press [pdf]
Learning in Graphical Models
1998, Jordan, M., ed., MIT Press
Learning in graphical models
Jordan, M.
1998, Kluwer Academic Publishers [pdf]
Operations for Learning with Graphical Models
Buntine, W. L.
Journal of Artificial Intelligence Research, 2() 159-225 (1994) [pdf]
This paper is a multidisciplinary review of empirical, statistical learning from a graphical model perspective. Well-known examples of graphical models include Bayesian networks, directed graphs representing a Markov chain, and undirected networks representing a Markov field. These graphical models are extended to model data analysis and empirical learning using the notation of plates. Graphical operations for simplifying and manipulating a problem are provided including decomposition, differentiation, and the manipulation of probability models from the exponential family. Two standard algorithm schemas for learning are reviewed in a graphical framework: Gibbs sampling and the expectation maximization algorithm. Using these operations and schemas, some popular algorithms can be synthesized from their graphical specification. This includes versions of linear regression, techniques for feed-forward networks, and learning Gaussian and discrete Bayesian networks from data. The paper conclu...
Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images
AN, S. & AN, D.
IEEE Trans. Pattern Anal. Machine Intell, 6() 721-741 (1984) [pdf]