T-Space at The University of Toronto Libraries >
School of Graduate Studies - Theses >
Please use this identifier to cite or link to this item:
|Title: ||Non-linear Latent Factor Models for Revealing Structure in High-dimensional Data|
|Authors: ||Memisevic, Roland|
|Advisor: ||Hinton, Geoffrey|
|Department: ||Computer Science|
|Keywords: ||Machine Learning|
|Issue Date: ||28-Jul-2008|
|Abstract: ||Real world data is not random: The variability in the data-sets that arise in computer vision,
signal processing and other areas is often highly constrained and governed by a number of
degrees of freedom that is much smaller than the superficial dimensionality of the data.
Unsupervised learning methods can be used to automatically discover the “true”, underlying
structure in such data-sets and are therefore a central component in many systems that deal
with high-dimensional data.
In this thesis we develop several new approaches to modeling the low-dimensional structure
in data. We introduce a new non-parametric framework for latent variable modelling, that in
contrast to previous methods generalizes learned embeddings beyond the training data and its
latent representatives. We show that the computational complexity for learning and applying
the model is much smaller than that of existing methods, and we illustrate its applicability
on several problems.
We also show how we can introduce supervision signals into latent variable models using
conditioning. Supervision signals make it possible to attach “meaning” to the axes of a latent
representation and to untangle the factors that contribute to the variability in the data. We
develop a model that uses conditional latent variables to extract rich distributed representations
of image transformations, and we describe a new model for learning transformation
features in structured supervised learning problems.|
|Appears in Collections:||Doctoral|
Department of Computer Science - Doctoral theses
This item is licensed under a Creative Commons License
Items in T-Space are protected by copyright, with all rights reserved, unless otherwise indicated.