Publication | Open Access
Never-ending learning
499
Citations
29
References
2018
Year
Natural Language ProcessingArtificial IntelligenceKnowledge BaseEngineeringKnowledge AcquisitionData ScienceMachine LearningMachine Learning ModelAlgorithmic LearningCase StudyComputer ScienceAutomated Knowledge AcquisitionAdaptive Learning
People learn diverse knowledge over years, becoming better learners, whereas current machine learning systems learn only a single function from a single dataset. The authors propose that machine learning should emulate humans by enabling agents to learn many knowledge types continuously over years, becoming better learners. They define the never‑ending learning paradigm and present the Never‑Ending Language Learner (NELL) as a case study, describing its design, experimental results, and successes and shortcomings. NELL has been learning from the Web continuously since January 2010, accumulating 120 million confidence‑weighted beliefs, thousands of interrelated functions that improve its reading competence, and it can reason over its knowledge base to infer new beliefs and invent new relational predicates.
Whereas people learn many different types of knowledge from diverse experiences over many years, and become better learners over time, most current machine learning systems are much more narrow, learning just a single function or data model based on statistical analysis of a single data set. We suggest that people learn better than computers precisely because of this difference, and we suggest a key direction for machine learning research is to develop software architectures that enable intelligent agents to also learn many types of knowledge, continuously over many years, and to become better learners over time. In this paper we define more precisely this never-ending learning paradigm for machine learning, and we present one case study: the Never-Ending Language Learner (NELL), which achieves a number of the desired properties of a never-ending learner. NELL has been learning to read the Web 24hrs/day since January 2010, and so far has acquired a knowledge base with 120mn diverse, confidence-weighted beliefs (e.g., servedWith(tea,biscuits) ), while learning thousands of interrelated functions that continually improve its reading competence over time. NELL has also learned to reason over its knowledge base to infer new beliefs it has not yet read from those it has, and NELL is inventing new relational predicates to extend the ontology it uses to represent beliefs. We describe the design of NELL, experimental results illustrating its behavior, and discuss both its successes and shortcomings as a case study in never-ending learning. NELL can be tracked online at http://rtw.ml.cmu.edu, and followed on Twitter at @CMUNELL.
| Year | Citations | |
|---|---|---|
Page 1
Page 1