Skip to main content Accessibility help
Internet Explorer 11 is being discontinued by Microsoft in August 2021. If you have difficulties viewing the site on Internet Explorer 11 we recommend using a different browser such as Microsoft Edge, Google Chrome, Apple Safari or Mozilla Firefox.

Chapter 13: Transfer Learning, Multi-task Learning, Continual Learning, and Meta-learning

Chapter 13: Transfer Learning, Multi-task Learning, Continual Learning, and Meta-learning

pp. 503-540

Authors

, King's College London
Resources available Unlock the full potential of this textbook with additional resources. There are free resources and Instructor restricted resources available for this textbook. Explore resources
  • Add bookmark
  • Cite
  • Share

Summary

As discussed so far in this book, the standard formulation of machine learning makes the following two basic assumptions: 1. Statistical equivalence of training and testing. The statistical properties of the data observed during training match those to be experienced during testing – i.e., the population distribution underlying the generation of the data is the same during both training and testing. 2. Separation of learning tasks. Training is carried out separately for each separate learning task – i.e., for any new data set and/or loss function, training is viewed as a new problem to be addressed from scratch.

About the book

Access options

Review the options below to login to check your access.

Purchase options

eTextbook
US$74.00
Hardback
US$74.00

Have an access code?

To redeem an access code, please log in with your personal login.

If you believe you should have access to this content, please contact your institutional librarian or consult our FAQ page for further information about accessing our content.

Also available to purchase from these educational ebook suppliers