Inductive transfer, or transfer learning, is a research problem in machine learning that focuses on storing knowledge gained while solving one problem and applying it to a different but related problem. For example, the abilities acquired while learning to walk presumably apply when one learns to run, and knowledge gained while learning to recognize cars could apply when recognizing trucks. This area of research bears some relation to the long history of psychological literature on transfer of learning, although formal ties between the two fields are limited.
Notably, scientists have developed algorithms for inductive transfer in Markov logic networks and Bayesian networks. Furthermore, researchers have applied techniques for transfer to problems in text classification, spam filtering, and urban combat simulation.
There still exists much potential in this field while the "transfer" hasn't yet led to significant improvement in learning. Also, an intuitive understanding could be that "transfer means a learner can directly learn from other correlated learners". However, in this way, such a methodology in transfer learning, whose direction is illustrated by, is not a hot spot in the area yet.
Famous quotes containing the word transfer:
“No sociologist ... should think himself too good, even in his old age, to make tens of thousands of quite trivial computations in his head and perhaps for months at a time. One cannot with impunity try to transfer this task entirely to mechanical assistants if one wishes to figure something, even though the final result is often small indeed.”
—Max Weber (18641920)