TWiki
>
ANC Web
>
PIGS
>
PIGSTwoThousandAndNine
(11 Aug 2010,
AthinaSpiliopoulou
)
(raw view)
E
dit
A
ttach
---+ Probabilistic Inference Group (!PIGS) - Archive ---++ ---++ Suggested topics * Amos: Sequential Monte-Carlo update. Basically quick summaries of salient papers on [[http://www-sigproc.eng.cam.ac.uk/smc/papers.html][this SMC page]] would be worth doing to get everyone up to speed. I'll probably arrange this for my session on 21 Aug. Charles: I agree and suggest this paper if it hasn't been done already: "Sequential Monte Carlo Samplers", (with P. Del Moral & A. Jasra), <em>J. Royal Statist. Soc. </em>B, vol. 68, no. 3, pp. 411-436, 2006. * NIPS2006 workshop on [[http://nips.cc/Conferences/2006/Program/event.php?ID=535][Dynamical Systems, Stochastic Processes and Bayesian Inference]]. * Compressed sensing, see http://www.dsp.ece.rice.edu/cs/ * Further look at deep belief networks, including the work on human motion which gives a good demo of conditional deep belieft network models. * Submodular functions and optimization (generalizes convexity to functions on sets), see e.g. http://www.mlpedia.org/index.php?title=Submodular_function ---++ Meetings in 2009 ---+++ Tue 12 December (Nicolas Le Roux) Talk by Nicolas Le Roux (Microsoft Research). Please see title and abstract below: How overconfidence slows you down, a learning story. Abstract: Nowadays, for many tasks such as object recognition or language modeling, data is plentiful. As such, the most important challenge has become to find a way to use all the available data rather than to deal with small training sets. In this setting, coined ``large-scale learning'' by Bottou and Bousquet,learning and optimization become different and powerful optimization algorithms are suboptimal learning algorithms. While many only considered optimization algorithms (or approximations thereof) to perform learning, I will show how designing a proper learning algorithm and making use of the covariance of the gradients can yield faster, more robust, convergence. I will also show that this covariance matrix is not an approximation of the Hessian and that the two matrices can be combined in an principled and efficient way. ---+++ Tue 17 November (Jakub Piatkowski) We will discuss the following paper: * Peter Orbanz and Joachim M. Buhmann, Int J Comput Vis 77: 25–45 (2008): [[http://dx.doi.org/10.1007/s11263-007-0061-0][Nonparametric Bayesian Image Segmentation]] ---+++ Wed 11 November (Maurizio Filippone) Talk by [[http://www.dcs.shef.ac.uk/~filippone/][Maurizio Filippone]]. Please see title, abstract and related material below: Information Theoretic Novelty Detection In this talk, we present a novel approach to online change detection problems when the training sample size is small. The proposed method is based on estimating the expected information content of a new data point in the null hypothesis that it has been generated from the same distribution as the training data. In the case of the Gaussian distribution, our approach is analytically tractable and closely related to classical statistical tests, since the expected information content is independent from the statistics of the generating distribution. Such a test naturally takes into account the variability of the statistics due to the finite sample effect, and thus it allows to control the false positive rate even when only a small training set is available. We then discuss two different extensions of the presented method. In the first one, we propose an approximation scheme to evaluate the information content of a new data point when the generating distribution is a mixture of Gaussians. Finally, we study the extension to autoregressive time series with Gaussian noise, thus removing the i.i.d. assumption. The experiments conducted on synthetic and real data sets show that our method maintains a good overall accuracy, but significantly improves the control over the false positive rate. Part of the material covered in the talk can be found here: * M. Filippone and G. Sanguinetti, To appear in Pattern Recognition: [[http://www.dcs.shef.ac.uk/~filippone/Publications/pr09.pdf ][Information Theoretic Novelty Detection]] * M. Filippone and G. Sanguinetti, Technical Report: [[http://www.dcs.shef.ac.uk/~filippone/Publications/tech_rep_ar09.pdf][Novelty detection in autoregressive models using information theoretic measures]] ---+++ Tue 3 November (David Reichert) We will discuss the following papers: * H Larochelle, Y Bengio, ICML (2008): [[http://www.iro.umontreal.ca/~lisa/publications2/index.php/attachments/single/171][Classification using Discriminative Restricted Boltzmann Machines]] * H Larochelle, D Erhan, P Vincent, AISTATS (2009): [[http://jmlr.csail.mit.edu/proceedings/papers/v5/larochelle09a/larochelle09a.pdf][Deep Learning using Robust Interdependent Codes]] ---+++ Tue 20 October (Jyri Kivinen) Jyri will present some of his joint work on statistical modeling of natural images and scenes using a hierarchical nonparametric Bayesian framework (J. J. Kivinen, E. B. Sudderth (Brown), and M. I. Jordan (UC Berkeley)). Please see abstract and background papers below: I will begin by describing the tree-structured latent variable model it employs to generate pyramidally organized multiscale image features, and to couple dependencies between them. I will then describe an extension using Hierarchical Dirichlet Processes to learn data-driven, global statistical image models of unbounded complexity. Finally, we develop effective learning algorithms using Markov chain Monte Carlo methods and belief propagation for categorizing images of novel scenes, and denoising them in a transfer learning-based algorithm. * J. J. Kivinen, E. B. Sudderth, and M. I. Jordan, ICCV (2007): [[http://www.cs.berkeley.edu/~jordan/papers/kivinen-sudderth-jordan-iccv07.pdf][Learning multiscale representations of natural scenes using Dirichlet processes.]] * J. J. Kivinen, E. B. Sudderth, and M. I. Jordan, ICIP (2007): [[http://www.cs.berkeley.edu/~jordan/papers/kivinen-sudderth-jordan-icip07.pdf][* Image denoising with nonparametric hidden Markov trees.]] ---+++ Fri 16 October (Michalis Titsias) Title: Variational Inference for Large Datasets in Gaussian Processes Gaussian processes (GPs) are stochastic processes over real-valued functions that can be used for Bayesian non-linear regression and classification problems. GPs also naturally arise as the solutions of linear stochastic differential equations. However, when the amount of observed or training data is large, the evaluation of posterior GPs is intractable because the computations scale as O(n^3) where n is the number of training examples. Therefore, for large datasets we need to consider approximate or sparse inference methods. In this talk we discuss sparse approximations for GPs based on inducing/support variables and the standard variational inference methodology. We apply this to regression, binary classification and large systems of linear stochastic differential equations. ---+++ Tue 6 October (Michael Dewar) We will discuss Variational Inference in Markov Jump Processes using the following papers: * M. Opper and G. Sanguinetti, NIPS (2007): [[http://www.google.com/url?sa=t&source=web&ct=res&cd=1&url=http%3A%2F%2Fbooks.nips.cc%2Fpapers%2Ffiles%2Fnips20%2FNIPS2007_0067.pdf&ei=o_3BStObEI-D4QabnYCLCA&usg=AFQjCNGBNEJ-E4zTbxaaakXxeckexhsVhQ&sig2=FA4lyRs7xpI6cEBrnBvizQ][Variational Inference for Markov Jump Processes]] * G. Sanguinetti, et al., Bioinformatics 25(10): 1280-1286 (2009): [[http://www.google.com/url?sa=t&source=web&ct=res&cd=2&url=http%3A%2F%2Fbioinformatics.oxfordjournals.org%2Fcgi%2Freprint%2Fbtp138v2.pdf&ei=q_7BSrCgOci14QaLx7mLCA&usg=AFQjCNFWUXFWO8J0Xk1uk8e2pJ-5dAS4mA&sig2=ZYcH-hYgEtiHDJFhHam96w][Switching regulatory models of cellular stress response]] ---+++ Tue 22 September (Jan Antolik) Joint session with joint DevCompNeuro journal club. Short description of the talk: The main goal of the project I'm working at is to predict what image has been presented to an animal based on the activity profile of group of cells (~50) obtain via two-photon imaging. The system would learn this prediction from recordings of pairs of images and activity profiles. Instead of directly predicting the image the goal is to be able to tell from a large set of images which one was the presented one.<br /><br />I have so far applied several simpler approaches to the problem, including the simple linear perceptrons, multi-layer NN with back-propagation and notably the 'gaussian pyramid model' which worked when applied to analogous problem but with fMRI data in a study by (Gallant et al. 2008). I have also tried several approaches to directly determine the receptive field of the neurons.<br /><br />So far these techniques haven't worked. My main aim with this presentation is to get some brainstorming going and perhaps learn from the real machine learning people about latest approaches to fit non-linear models. I would be particularly interested in learning about methods of learning recurrent NN, as it appears that a lot of the neural responses are due to lateral interaction as opposed to the feed-forward receptive field structure. Kay NN, Naselaris T, Prenger RJ and Gallant JL (2008): [[http://www.ncbi.nlm.nih.gov/pubmed/18322462?ordinalpos=1&itool=EntrezSystem2.PEntrez.Pubmed.Pubmed_ResultsPanel.Pubmed_DefaultReportPanel.Pubmed_RVDocSum][Identifying natural images from human brain activity]] ---+++ Tue 25 August (Edwin Bonilla) * Vikash Mansinghka, Daniel Roy, Eric Jonas, Joshua Tenenbaum<strong><em>. </em></strong>[[http://jmlr.csail.mit.edu/proceedings/papers/v5/mansinghka09a/mansinghka09a.pdf][Exact and Approximate Sampling by Systematic Stochastic Search]]. AISTATS 2009. * Ricardo Silva, Zoubin Ghahramani<strong><em>. </em></strong>[[http://jmlr.csail.mit.edu/proceedings/papers/v5/silva09b/silva09b.pdf][Factorial Mixture of Gaussians and the Marginal Independence Model]]. AISTATS 2009. ---+++ Tue 28 July (UAI session) Brief discussions on the following UAI 2009 papers: * KMC: [[http://www.cs.mcgill.ca/~uai2009/papers/UAI2009_0237_ba4d9acc2b44117a6e31e02cf206f196.pdf][Group Sparse Priors for Covariance Estimation]] * KMC: [[http://www.cs.mcgill.ca/~uai2009/papers/UAI2009_0157_72e4f317cc3228a8d51b60ed9ccafc6d.pdf][Multi-Task Feature Learning Via Efficient L2,1-Norm Minimization]] * ASP: [[http://www.cs.mcgill.ca/~uai2009/papers/UAI2009_0167_d5a0043e06398cb511c8f822e5d210a2.pdf][Products of Hidden Markov Models: It Takes N>1 to Tango]] * ASP: [[http://www.cs.mcgill.ca/~uai2009/papers/UAI2009_0149_8d84b5fed914adbf2322fba960ad26ee.pdf][Convexifying the Bethe Free Energy]] * FD: [[http://www.cs.mcgill.ca/~uai2009/papers/UAI2009_0193_6b737c0f7e024139524c730e0a3486c5.pdf][Modeling Discrete Interventional Data using Directed Cyclic Graphical Models]] * AJS: [[http://www.cs.mcgill.ca/~uai2009/papers/UAI2009_0072_2dcf8be912a81c7dffd3bd21da17794a.pdf][Lower Bound Bayesian Networks - Efficient Inference of Lower Bounds on Probability Distributions, Daniel Andrade, Bernhard Sick]] * CKIW [[http://www.cs.mcgill.ca/~uai2009/papers/UAI2009_0134_c9c16e478d8e82ecc3848c5dc76b4925.pdf][Mean Field Variational Approximation for Continuous-Time Bayesian Networks]] * CKIW [[http://www.cs.mcgill.ca/~uai2009/papers/UAI2009_0266_4908509e6c78a93077c5d35e2082a91e.pdf][Virtual Vector Machine for Bayesian Online Classification]] * FA: [[http://www.cs.mcgill.ca/~uai2009/papers/UAI2009_0192_ea3f8ff70d9586675648a0d85eb36323.pdf][Optimization of Structured Mean Field Objectives]] UAI 2009 proceedings at http://www.cs.mcgill.ca/~uai2009/proceedings.html ---+++ Tue 14 July (Kian Ming Chai) We will discuss the following paper: * E. B. Anderes and M. L. Stein, Annals of Statistics, Vol. 36, No. 2, 719-741, (2008): [[http://www.stat.ucdavis.edu/~anderes/papers/AOS0328.pdf][Estimating deformations of isotropic Gaussian random fields on the plane.]] ---+++ Tue 30 June (Amos Storkey) We will discuss Deep Boltzmann Machines using the paper: * R. Salakhutdinov and G.E. Hinton, To appear in Artificial Intelligence and Statistics (2009): [[http://www.google.com/url?sa=t&source=web&ct=res&cd=7&url=http%3A%2F%2Fjmlr.csail.mit.edu%2Fproceedings%2Fpapers%2Fv5%2Fsalakhutdinov09a%2Fsalakhutdinov09a.pdf&ei=xDVCSrexHMS7jAelvoynBg&usg=AFQjCNFdBbIolrKW5K9KKfT4BtPOAEQQhQ&sig2=WyX1NvWtFbPFWDSxNWsKDg][Deep Boltzmann Machines]] If there is enough time, Amos will also give a basic introduction to Martingales using: * [[http://www.stat.duke.edu/courses/Spring06/sta205/lec/mg.pdf][Introduction to Martingales]] by Robert L. Wolpert ---+++ Tue 23 June (2nd ICML session) Brief discussions on the following ICML 2009 papers: * CW: [[http://www.cs.mcgill.ca/~icml2009/papers/447.pdf][Herding Dynamical Weights to Learn]] * JAT: [[http://www.cs.mcgill.ca/~icml2009/papers/481.pdf][Learning Linear Dynamical Systems without Sequence Information]] * JAT: [[http://www.cs.mcgill.ca/~icml2009/papers/323.pdf][Function factorization using warped Gaussian processes]] * MD: [[http://www.cs.mcgill.ca/~icml2009/papers/295.pdf][Learning Nonlinear Dynamic Models]] * MD: [[http://www.cs.mcgill.ca/~icml2009/papers/478.pdf][Dynamic Mixed Membership Block Model for Evolving Networks]] Note: ICML 2009 proceedings at http://www.cs.mcgill.ca/~icml2009/abstracts.html ---+++ Tue 16 June (ICML session) Brief discussions on the following ICML 2009 papers: * KMC: [[http://www.cs.mcgill.ca/~icml2009/papers/284.pdf][Sparse Gaussian Graphical Models with Unknown Block Structure]] * KMC: [[http://www.cs.mcgill.ca/~icml2009/papers/452.pdf][Learning with Structured Sparsity]] * DR: [[http://www.cs.mcgill.ca/~icml2009/papers/223.pdf][Deep Learning from Temporal Coherence in Video]] * ASP: [[http://www.cs.mcgill.ca/~icml2009/papers/119.pdf][Curriculum Learning]] * ASP: [[http://www.cs.mcgill.ca/~icml2009/papers/178.pdf][Factored Conditional Restricted Boltzmann Machines for Modeling Motion Style]] * FA: [[http://www.cs.mcgill.ca/~icml2009/papers/279.pdf][Regression by dependence minimization and its application to causal inference]] * FA: [[http://www.cs.mcgill.ca/~icml2009/papers/573.pdf][Split Variational Inference]] ---+++ Tue 9 June (Michael Dewar) We will discuss Hierarchical HMMs using the paper: * K. Murphy and M. Paskin, NIPS (2001): [[http://www.cs.ubc.ca/~murphyk/Papers/hhmm_nips01.pdf][Linear Time Inference in Hierarchical HMMs]] Note: There is an extended version of the paper: * K. Murphy, November 2001: [[http://www.cs.ubc.ca/~murphyk/Papers/hhmm_tr.ps.gz][Hierarchical HMMs]] ---+++ Tue 26 May (Athina Spiliopoulou) We will discuss two variants from the RBM/DBN literature using the papers: * H. Lee, et al., ICML (2009): [[http://www.google.com/url?sa=t&source=web&ct=res&cd=1&url=http%3A%2F%2Fwww.stanford.edu%2F~hllee%2Ficml09-ConvolutionalDeepBeliefNetworks.pdf&ei=nrUWSquXMsigjAe-1eH1DA&usg=AFQjCNHMNYSIIAI_h8AM5z4Pho_gZkGjJg&sig2=cTaJV1Z8N26rUIuNjTcObw][Convolutional Deep Belief Networks for Scalable Unsupervised Learning of Hierarchical Representations]] * R. Memisevic, G. E. Hinton, CVRP (2007): [[http://www.google.com/url?sa=t&source=web&ct=res&cd=1&url=http%3A%2F%2Fwww.cs.toronto.edu%2F~hinton%2Fabsps%2FgbmTR.pdf&ei=SLUWSr2GIMaMjAfNtJWBDQ&usg=AFQjCNGsT9dyDkTlTutNq07EYfoYFDIjpQ&sig2=b31spP1OixgP6vScD78Eow][Unsupervised Learning of Image Transformations]] ---+++ Tue 21 April (Chris Williams) We will discuss multi-arm bandits and Gittins indices. This is a simple case where the exploration-exploitation tradeoff is seen, and there is an optimal Bayesian solution. The papers J. C. Gittins, Bandit Processes and Dynamic Allocation Indices, Journal of the Royal Statistical Society. Series B (Methodological), Vol. 41, No. 2. (1979), pp. 148-177. J. C. Gittins, D. M. Jones, A Dynamic Allocation Index for the Discounted Multiarmed Bandit Problem, Biometrika, Vol 66, No. 3. (1979), pp. 561-565. are available via http://en.wikipedia.org/wiki/Gittins_index ---+++ Tue 7 April 2009 (Jakub Piatkowski) * L.Li et al., Bayesian Analysis 3:171-196 (2008): [[http://ba.stat.cmu.edu/journal/2008/vol03/issue01/li.pdf][A Method for Avoiding Bias from Feature Selection with Application to Naive Bayes Classification Models]] ---+++ Tue 24 March 2009 (Nicolas Heess) * A. Pelizzola, J.PHYS.A 38:R309 (2005): [[http://arxiv.org/pdf/cond-mat/0508216v1][Cluster Variation Method in Statistical Physics and Probabilistic Graphical Models]] ---+++ Tue 10 March 2009 (Andrew Dai) * D.M. Roy and Y.W. Teh, NIPS (2009): [[http://books.nips.cc/papers/files/nips21/NIPS2008_0849.pdf][The Mondrian Process]] * P. Rai and H. Daume III, NIPS (2009): [[http://books.nips.cc/papers/files/nips21/NIPS2008_0954.pdf][The Infinite Hierarchical Factor Regression Model]] ---+++ Tue 24 February 2009 (Edwin Bonilla) * [[http://www.cogsci.ucsd.edu/%7Etodorov/papers/duality.pdf][General duality between optimal control and estimation]] ( _Emanuel Todorov. IEEE Conference on Decision and Control, 2008_) * [[http://arxiv.org/pdf/0901.0633v2][Optimal control as a graphical model inference problem]] ( _B. Kappen, V. Gomez, M. Opper. arXiv:0901.0633v2 [cs.AI], 2009_) ---+++ Tue 10 February 2009 (Kian Ming Chai) * R. Koenker and K. F. Hallock, The Journal of Economic Perspectives 15:143-156 (2001): [[http://www.jstor.org/stable/2696522][Quantile Regression]] * I. Takeuchi, et al., Journal of Machine Learning Research 7:1231--1264 (2006): [[http://www.jmlr.org/papers/volume7/takeuchi06a/takeuchi06a.pdf][Nonparametric Quantile Estimation]] ---+++ Tue 27 January 2009 (Amos Storkey) * L. Li and R. M. Neal, Bayesian Analysis 3:793-822 (2008): [[http://www.cs.toronto.edu/~radford/comp-param.abstract.html][Compressing parameters in Bayesian high-order models with application to logistic sequence models]] * R. Adams et al., NIPS (2009): [[http://books.nips.cc/papers/files/nips21/NIPS2008_0240.pdf][The Gaussian Process Density Sampler]] ---+++ Tue 13 January 2009 (NIPS) Note: NIPS 21 preproceedings at http://books.nips.cc/nips21.html * NH: I. Murray, R. Salakhutdinov: Evaluating probabilities under high-dimensional latent variable models * NH: I. Sutskever, G. Hinton, G. Taylor: The Recurrent Temporal Restricted Boltzmann Machine * EB: [[http://books.nips.cc/papers/files/nips21/NIPS2008_0120.pdf][Reducing statistical dependencies in natural signals using radial Gaussianization]] ( _Siwei Lyu, Eero Simoncelli_) * EB: [[ftp://ftp.cs.man.ac.uk/pub/ai/neill/spmulti.pdf][Sparse Convolved Gaussian Processes for Multi-ouptut Regression]] ( _M. Alvarez, N. Lawrence_) * CW: [[http://books.nips.cc/papers/files/nips21/NIPS2008_1027.pdf][Shared Segmentation of Natural Scenes Using Dependent Pitman-Yor Processes.]] Erik Sudderth, Michael Jordan * CW: [[http://books.nips.cc/papers/files/nips21/NIPS2008_0635.pdf][The Conjoint Effect of Divisive Normalization and Orientation Selectivity on Redundancy Reduction]] Fabian H. Sinz, Matthias Bethge * CW: [[http://books.nips.cc/papers/files/nips21/NIPS2008_0745.pdf][Bayesian Exponential Family PCA]] Shakir Mohamed, Katherine Heller, Zoubin Ghahramani. See also [[http://www.cs.ualberta.ca/~dale/papers/allerton08.pdf] [Efficient global optimization for exponential family PCA and low-rank matrix factorization.]] Guo, Y. and Schuurmans, D. (2008) in Allerton Conference on Communication, Control, and Computing. * MD: [[http://books.nips.cc/papers/files/nips21/NIPS2008_0536.pdf][Using Bayesian Dynamic Systems for Motion Template Libraries]] Silvia Chiappa, Jens Kober, Jan Peters * MD [[http://books.nips.cc/papers/files/nips21/NIPS2008_0312.pdf][Nonparametric Bayesian Learning of Switching Linear Dynamical Systems]] Emily Fox, Erik Sudderth, Michael Jordan, Alan Willsky * DR: [[http://books.nips.cc/papers/files/nips21/NIPS2008_0060.pdf][Cascaded Classification Models: Combining Models for Holistic Scene Understanding]] Geremy Heitz, Stephen Gould, Ashutosh Saxena, Daphne Koller Some other NIPS 21 papers CW found to be of interest: * The Infinite Factorial Hidden Markov Model, Jurgen Van Gael, Yee Whye Teh, Zoubin Ghahramani * Deep Learning with Kernel Regularization for Visual Recognition. Kai Yu, Wei Xu, Yihong Gong * Cascaded Classification Models: Combining Models for Holistic Scene Understanding. Geremy Heitz, Stephen Gould, Ashutosh Saxena, Daphne Koller
E
dit
|
A
ttach
|
P
rint version
|
H
istory
: r2
<
r1
|
B
acklinks
|
V
iew topic
|
Ra
w
edit
|
M
ore topic actions
Topic revision: r2 - 11 Aug 2010 - 13:02:42 -
AthinaSpiliopoulou
ANC
ANC wiki
Home
Changes
Index
Search
Elsewhere
Main ANC site
DTC site
Copyright © by the contributing authors. All material on this collaboration platform is the property of the contributing authors.
Ideas, requests, problems regarding TWiki?
Send feedback
This Wiki uses
Cookies