Never-Ending Language Learning

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search

Never-Ending Language Learning system (NELL) is a semantic machine learning system developed by a research team at Carnegie Mellon University, and supported by grants from DARPA, Google, NSF, and CNPq with portions of the system running on a supercomputing cluster provided by Yahoo!.[1]

Process and goals[edit]

NELL was programmed by its developers to be able to identify a basic set of fundamental semantic relationships between a few hundred predefined categories of data, such as cities, companies, emotions and sports teams. Since the beginning of 2010, the Carnegie Mellon research team has been running NELL around the clock, sifting through hundreds of millions of web pages looking for connections between the information it already knows and what it finds through its search process – to make new connections in a manner that is intended to mimic the way humans learn new information.[2] For example, in encountering the word pair "Pikes Peak", NELL would notice that both words are capitalized and deduce from the second word that it was the name of a mountain, and then build on the relationship of words surrounding those two words to deduce other connections.[1]

The goal of NELL and other semantic learning systems, such as IBM's Watson system, is to be able to develop means of answering questions posed by users in natural language with no human intervention in the process.[3] Oren Etzioni of the University of Washington lauded the system's "continuous learning, as if NELL is exercising curiosity on its own, with little human help".[1]

By October 2010, NELL has doubled the number of relationships it has available in its knowledge base and has learned 440,000 new facts, with an accuracy of 87%.[4][1] Team leader Tom M. Mitchell, chairman of the machine learning department at Carnegie Mellon described how NELL "self-corrects when it has more information, as it learns more", though it does sometimes arrive at incorrect conclusions. Accumulated errors, such as the deduction that Internet cookies were a kind of baked good, led NELL to deduce from the phrases "I deleted my Internet cookies" and "I deleted my files" that "computer files" also belonged in the baked goods category.[5] Clear errors like these are corrected every few weeks by the members of the research team and the system is allowed to continue its learning process.[1]

See also[edit]

References[edit]

  1. ^ a b c d e "Aiming to Learn as We Do, a Machine Teaches Itself". New York Times. October 4, 2010. Retrieved 2010-10-05. Since the start of the year, a team of researchers at Carnegie Mellon University — supported by grants from the Defense Advanced Research Projects Agency and Google, and tapping into a research supercomputing cluster provided by Yahoo — has been fine-tuning a computer system that is trying to master semantics by learning more like a human.
  2. ^ Project Overview, Carnegie Mellon University. Accessed October 5, 2010.
  3. ^ Trader, Tiffany. "Machine Learns Language Starting with the Facts", HPCwire, October 5, 2010. Accessed October 5, 2010.
  4. ^ "NELL: Never-Ending Language Learning", Carnegie Mellon University. Accessed October 5, 2010.
  5. ^ VanHemert, Kyle. "Right Now A Computer Is Reading Online, Teaching Itself Language", Gizmodo, October 6, 2010. Accessed October 5, 2010.
  6. ^ https://www.bbc.com/news/technology-25090534

External links[edit]