Course Projects CMPT 880 - Deep Learning
Here are some suggestions for course projects. Feel free to suggest your own. See the section on Additional Readings. You should discuss the topic with me before you start in any case.
The typical project would evaluate a deep learning method on an interesting problem of your choice. "Interesting" means some mix of real-world importance and learning complexity. I assume that many students will already have a learning problem that they are working on, so it is natural to try deep learning methods. I'm also open to other types of projects, for instance surveys or theoretical analysis. Below I list example topics.
Compare with other models for learning latent features
- PCA, kernelized PCA (non-linear).
- Neural nets.
- Hierarchical Clustering, Multiple Clustering.
- SVMs: can we think of deep learning as learning a kernel? How do deep models compare to SVMs with latent variables?
- Consider the experiments and arguments in "Scaling Learning Towards AI".
- How does a Deep Belief net perform in learning parity? The sine function?
- The paper reports a dataset where a convolutional neural net performs much better than an SVM with a local kernel. How about a DBN vs. SVM? Can you find a dataset with the reverse performance, where an SVM outperforms a DBN? If not, why not? What would such a dataset be like?
- Consider the experiments and arguments in "Scaling Learning Towards AI".
Compare Learning Methods for latent feature models
- EM
- gradient descent, e.g. conjugate version.
Compare different deep learning Models
- Number of layers.
- Different types of models, e.g.
- Logistic Regression
- Multilayer perceptron
- Deep Convolutional Network
- Auto Encoders, Denoising Autoencoders
- Stacked Denoising Auto-Encoders
- Restricted Boltzmann Machines
- Deep Belief Networks
Apply Deep Learning
- NLP
- Vision
- Multimodal (see tutorials)
- Transfer Learning.
- Do DBNs achieve the invariance and feature abstraction discussed in the paper? Consider this type of experiment: first learn a DBN that represents the sine function. Then use the weight settings as initial settings for learning the cosine function.
Surveys
- Literature Review
- Theoretical Analysis
- We discussed the equivalence between regularlized neural net learning with 1 hidden layer vs. support vector machine learning. Is it possible to extend this to deep neural nets with several hidden layers?
Updated Mon March 10 2014, 10:23 by oschulte.