I design machine learning algorithms that try to solve some of today's most challenging problems in computer science and statistics.
I adapt ideas from physics and the statistical sciences, and use them in algorithms that can be applied to areas such as: bioinformatics, artificial intelligence, pattern recognition, document information retrieval, and human-computer interaction.
Click on the following topics to see research descriptions and some papers:-
| Nonparametric Bayes | - | powerful nonparametric text/document modelling |
| Variational Bayesian Methods | - | approximate Bayesian learning and inference |
| Bioinformatics | - | microarray analysis using variational Bayes |
| Embedded Hidden Markov Models | - | a novel tool for time series inference |
| Probabilistic Sensor Fusion | - | combining modalities using Bayesian graphical models |
| Collaborators | - | people I have worked with |
Inference using Embedded Hidden Markov Models
Embedded HMMs, invented by Radford Neal, constitute a new type of inference tool for sequential data that will allow many filtering, prediction, and control problems to be tackled and solved in a very novel way. They are an elegant generalisation of the particle filtering and smoothing procedures that are currently used in non-linear systems. Embedded HMMs efficiently perform inference in non-linear systems by temporarily embedding a tractable (finite-state) HMM in the non-linear hidden state-space of the model.
So far the embedded HMM has been applied to the tasks of robot localisation, speech analysis, and recovering 3-dimensional structure from human motion sequences (ongoing work). In the future we hope to apply it to more general inference tasks in the graphical models framework.
Neal, R.M., Beal, M.J. and Roweis, S.T.