Skip to main content

Introduction to Semi-Supervised Learning

  • Book
  • © 2009

Overview

This is a preview of subscription content, log in via an institution to check access.

Access this book

eBook USD 29.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book USD 37.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Other ways to access

Licence this eBook for your library

Institutional subscriptions

Table of contents (8 chapters)

About this book

Semi-supervised learning is a learning paradigm concerned with the study of how computers and natural systems such as humans learn in the presence of both labeled and unlabeled data. Traditionally, learning has been studied either in the unsupervised paradigm (e.g., clustering, outlier detection) where all the data are unlabeled, or in the supervised paradigm (e.g., classification, regression) where all the data are labeled. The goal of semi-supervised learning is to understand how combining labeled and unlabeled data may change the learning behavior, and design algorithms that take advantage of such a combination. Semi-supervised learning is of great interest in machine learning and data mining because it can use readily available unlabeled data to improve supervised learning tasks when the labeled data are scarce or expensive. Semi-supervised learning also shows potential as a quantitative tool to understand human category learning, where most of the input is self-evidently unlabeled. In this introductory book, we present some popular semi-supervised learning models, including self-training, mixture models, co-training and multiview learning, graph-based methods, and semi-supervised support vector machines. For each model, we discuss its basic mathematical formulation. The success of semi-supervised learning depends critically on some underlying assumptions. We emphasize the assumptions made by each model and give counterexamples when appropriate to demonstrate the limitations of the different models. In addition, we discuss semi-supervised learning for cognitive psychology. Finally, we give a computational learning theoretic perspective on semi-supervised learning, and we conclude the book with a brief discussion of open questions in the field. Table of Contents: Introduction to Statistical Machine Learning / Overview of Semi-Supervised Learning / Mixture Models and EM / Co-Training / Graph-Based Semi-Supervised Learning / Semi-Supervised Support Vector Machines/ Human Semi-Supervised Learning / Theory and Outlook

Authors and Affiliations

  • University of Wisconsin, Madison, USA

    Xiaojin Zhu, Andrew B. Goldberg

About the authors

Xiaojin Zhu is an assistant professor in the Computer Sciences department at the University of Wisconsin-Madison. His research interests include statistical machine learning and its applications in cognitive psychology, natural language processing, and programming languages. Xiaojin received his Ph.D. from the Language Technologies Institute at Carnegie Mellon University in 2005. He worked on Mandarin speech recognition as a research staff member at IBM China Research Laboratory in 1996-1998. He received M.S. and B.S. in computer science from Shanghai Jiao Tong University in 1996 and 1993, respectively. His other interests include astronomy and geology. Andrew B.Goldberg is a Ph.D. candidate in the Computer Sciences department at the University of Wisconsin-Madison. His research interests lie in statistical machine learning (in particular, semi-supervised learning) and natural language processing. He has served on the program committee for national and international conferences including AAAI, ACL, EMNLP, and NAACL-HLT. Andrew was the recipient of a UW-Madison First-Year Graduate School Fellowship for 2005-2006 and a Yahoo! Key Technical Challenges Grant for 2008-2009. Before his graduate studies, Andrew received a B.A. in computer science from Amherst College, where he graduated magna cum laude with departmental distinction in 2003. He then spent two years writing, editing, and developing teaching materials for introductory computer science and Web programming textbooks at Deitel and Associates. During this time, he contributed to several Deitel books and co-authored the 3rd edition of Internet & World Wide Web How to Program. In 2005, Andrew entered graduate school at UW-Madison and, in 2006 received his M.S. in computer science. In his free time, Andrew enjoys live music, cooking, photography, and travel.

Bibliographic Information

Publish with us