8803 Ideas for Projects
One of the course requirements is to do a small project, which you may
do individually or in a group of 2. A project might involve conducting
an experiment or thinking about a theoretical problem, or trying to
relate two problems. It could even just be reading 2 research papers
and explaining how they relate. The end result should be a 5-10 page
report, and a 10 - 15 minute presentation. Here are a few ideas for
possible topics for projects. You might also want to take a look at
recent COLT, ICML,
or
NIPS
proceedings. All the recent COLT proceedings contain a few open
problems, some with monetary rewards!
Project Ideas
Semi-supervised learning and related topics:
- M.F. Balcan and A. Blum.
A Discriminative Model for Semi-Supervised Learning. Journal of
the ACM, 2010.
- M.F. Balcan and A. Blum.
Open Problems in Efficient Semi-Supervised PAC Learning. Open
problem, COLT 2007. [Monetary reward!]
- A. Carlson, J. Betteridge, R. C. Wang, E. R. Hruschka Jr., and T.
M. Mitchell. Coupled
Semi-Supervised
Learning
for
Information
Extraction. International
Conference on Web Search and Data Mining (WSDM), 2010.
- L. Xu, M. White, and D. Schuurmans.
Optimal Reverse Prediction. Twenty-Sixth International Conference
on Machine
Learning (ICML), 2009.
Active learning:
- S. Dasgupta. Coarse
sample complexity bounds for active learning. Advances in Neural
Information Processing Systems (NIPS), 2005.
- A. Beygelzimer, S. Dasgupta, and J. Langford. Importance-weighted
active
learning. Twenty-Sixth International Conference on Machine
Learning (ICML), 2009.
- M.F. Balcan, A. Beygelzimer, J. Langford. Agnostic
active
learning. JCSS 2009.
- S. Fine, Y. Mansour. Active
Sampling for Multiple Output Identification. COLT 2006.
- M.F. Balcan, S. Hanneke, and J. Wortman. The
True
Sample
Complexity
of
Active
Learning. Machine Learning
Journal 2010.
- S. Hanneke's thesis
Theoretical Foundations of Active Learning. CMU 2009
- See also the NIPS
2009 Workshop on
Adaptive
Sensing,
Active,
Learning and
Experimental Design:
Theory, Methods, and Applications.
Clustering and related topics:
Relationship between convex cost functions and discrete loss:
These papers look at relationships between different kinds of objective
functions for learning problems.
Boosting related topics:
Efficient agnostic learning:
Learning with kernel functions:
Learning in Markov Decision Processes: See M. Kearns's home
page and Y.
Mansour's home page for a number of good papers. Also S.
Kakade's
thesis.
PAC-Bayes bounds, shell-bounds, other methods of obtaining
confidence bounds. Some papers:
Learning in Graphical Models (Bayes Nets)