Latent Dirichlet Bayesian Co-Clustering.
Machine Learning and Knowledge Discovery in Databases:522-537, 2009.
Pu Wang, Carlotta Domeniconi und Kathryn Laskey.
[doi]
[Kurzfassung]
[BibTeX]
Co-clustering has emerged as an important technique for mining contingency data matrices. However, almost all existing co-clustering
algorithms are hard partitioning, assigning each row and column of the data matrix to one cluster. Recently a Bayesian co-clusteringapproach has been proposed which allows a probability distribution membership in row and column clusters. The approach usesvariational inference for parameter estimation. In this work, we modify the Bayesian co-clustering model, and use collapsedGibbs sampling and collapsed variational inference for parameter estimation. Our empirical evaluation on real data sets showsthat both collapsed Gibbs sampling and collapsed variational inference are able to find more accurate likelihood estimatesthan the standard variational Bayesian co-clustering approach.
Bayesian Co-clustering..
In:
ICDM, Seiten 530-539.
IEEE Computer Society, 2008.
Hanhuai Shan und Arindam Banerjee.
[doi]
[BibTeX]
Collapsed Variational Inference for HDP.
In:
Advances in Neural Information Processing Systems, Band 20.
2008.
Y. W. Teh, K. Kurihara und M. Welling.
[BibTeX]
Bayesian approach to clustering real value, categorical and network
data: solution via variational methods.
2008. cite arxiv:0805.2689
Comment: 12 pages, 5 figures. New sections added.
Alexei Vazquez.
[doi]
[Kurzfassung]
[BibTeX]
Data clustering, including problems such as finding network communities, can
be put into a systematic framework by means of a Bayesian approach. The
application of Bayesian approaches to real problems can be, however, quite
challenging. In most cases the solution is explored via Monte Carlo sampling or
variational methods. Here we work further on the application of variational
methods to clustering problems. We introduce generative models based on a
hidden group structure and prior distributions. We extend previous attends by
Jaynes, and derive the prior distributions based on symmetry arguments. As a
case study we address the problems of two-sides clustering real value data and
clustering data represented by a hypergraph or bipartite graph. From the
variational calculations, and depending on the starting statistical model for
the data, we derive a variational Bayes algorithm, a generalized version of the
expectation maximization algorithm with a built in penalization for model
complexity or bias. We demonstrate the good performance of the variational
Bayes algorithm using test examples.
Hierarchical dirichlet processes.
Journal of the American Statistical Association, 101(476):1566-1581, 2006.
Y.W. Teh, M.I. Jordan, M.J. Beal und D.M. Blei.
[doi]
[BibTeX]
Clustering with Bregman Divergences..
Journal of Machine Learning Research, 6:1705-1749, 2005.
Arindam Banerjee, Srujana Merugu, Inderjit S. Dhillon und Joydeep Ghosh.
[doi]
[BibTeX]
Parameter estimation for text analysis.
Web: http://www. arbylon. net/publications/text-est. pdf, 2005.
G. Heinrich.
[doi]
[BibTeX]
Introduction to Bayesian learning.
In:
SIGGRAPH '04: ACM SIGGRAPH 2004 Course Notes, Seiten 22.
ACM, New York, NY, USA, 2004.
Aaron Hertzmann.
[doi]
[Kurzfassung]
[BibTeX]
Sophisticated computer graphics applications require complex models of appearance, motion, natural phenomena, and even artistic style. Such models are often difficult or impossible to design by hand. Recent research demonstrates that, instead, we can "learn" a dynamical and/or appearance model from captured data, and then synthesize realistic new data from the model. For example, we can capture the motions of a human actor and then generate new motions as they might be performed by that actor. Bayesian reasoning is a fundamental tool of machine learning and statistics, and it provides powerful tools for solving otherwise-difficult problems of learning about the world from data. Beginning from first principles, this course develops the general methodologies for designing learning algorithms and describes their application to several problems in graphics.
An introduction to graphical models.
Web. 2001.
Kevin Murphy.
[doi]
[BibTeX]
Introduction to inference for Bayesian networks.
Learning in graphical models:9-26, 1999.
R. Cowell.
[doi]
[BibTeX]
An Introduction to Variational Methods for Graphical Models.
Mach. Learn., 37(2):183-233, 1999.
Michael I. Jordan, Zoubin Ghahramani, Tommi S. Jaakkola und Lawrence K. Saul.
[doi]
[BibTeX]
Advanced inference in Bayesian networks.
1998.
R. Cowell.
[doi]
[BibTeX]
Learning in Graphical Models.
1998.
[BibTeX]
Operations for Learning with Graphical Models.
Journal of Artificial Intelligence Research, 2:159-225, 1994.
Wray L. Buntine.
[doi]
[Kurzfassung]
[BibTeX]
This paper is a multidisciplinary review of empirical, statistical learning from a graphical model perspective. Well-known examples of graphical models include Bayesian networks, directed graphs representing a Markov chain, and undirected networks representing a Markov field. These graphical models are extended to model data analysis and empirical learning using the notation of plates. Graphical operations for simplifying and manipulating a problem are provided including decomposition, differentiation, and the manipulation of probability models from the exponential family. Two standard algorithm schemas for learning are reviewed in a graphical framework: Gibbs sampling and the expectation maximization algorithm. Using these operations and schemas, some popular algorithms can be synthesized from their graphical specification. This includes versions of linear regression, techniques for feed-forward networks, and learning Gaussian and discrete Bayesian networks from data. The paper conclu...
Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images.
IEEE Trans. Pattern Anal. Machine Intell, 6:721-741, 1984.
S.G.E.M. AN und D.G.E.M. AN.
[doi]
[BibTeX]