Rich Caruana Multitask Learning Degree Type: Ph.D. in Computer Science Advisor(s): Jaime Carbonell Graduated: May 1998 Abstract: Multitask Learning is an approach to inductive transfer that improves learning for one task by using the information contained in the training signals of other related tasks. It does this by learning tasks in parallel while using a shared representation; what is learned for each task can help other tasks be learned better. In this thesis we demonstrate multitask learning for a dozen problems. We explain how multitask learning works and show that there are many opportunities for multitask learning in real domains. We show that in some cases features that would normally be used as inputs work better if used as multitask outputs instead. We present suggestions for how to get the most out of multitask learning in artificial neural nets, present an algorithm for multitask learning with case-based methods like k-nearest neighbor and kernel regression, and sketch an algorithm for multitask learning in decision trees. Multitask learning improves generalization performance, can be applied in many different kinds of domains, and can be used with different learning algorithms. We conjecture there will be many opportunities for its use on real-world problems. Thesis Committee: Tom Mitchell (Chair) Herb Simon Dean Pomerleau Tom Dietterich (Oregon State) James Morris, Head, Computer Science Department Raj Reddy Dean, School of Computer Science Keywords: Machine learning, neural networks, k-nearest neighbor, multitask learning, inductive bias, medical decision making, pneumonia, ALVINN, autonomous vehicle navigation, pattern recognition, inductive transfer, learning-to-learn CMU-CS-97-203.pdf (1.58 MB) ( 255 pages) Copyright Notice