Machine perception

From Wikipedia, the free encyclopedia
Jump to navigation Jump to search

Machine perception is the capability of a computer system to interpret data in a manner that is similar to the way humans use their senses to relate to the world around them.[1][2][3] The basic method that the computers take in and respond to their environment is through the attached hardware. Until recently input was limited to a keyboard, or a mouse, but advances in technology, both in hardware and software, have allowed computers to take in sensory input in a way similar to humans.[1][2]

Machine perception allows the computer to use this sensory input, as well as conventional computational means of gathering information, to gather information with greater accuracy and to present it in a way that is more comfortable for the user.[1] These include computer vision, machine hearing, and machine touch.

The end goal of machine perception is to give machines the ability to see, feel and perceive the world as humans do and therefore for them to be able to explain in a human way why they are making their decisions, to warn us when it is failing and more importantly, the reason why it is failing.[4]

Machine vision[edit]

Computer vision is a field that includes methods for acquiring, processing, analyzing, and understanding images and, in general, high-dimensional data from the real world in order to produce numerical or symbolic information, e.g., in the forms of decisions. Computer vision has many applications already in use today such as facial recognition, geographical modeling, and even aesthetic judgment.[5]

Machine hearing[edit]

Machine hearing, also known as machine listening or computer audition, is the ability of a computer or machine to take in and process sound data such as music or speech. This area has a wide range of application including music recording and compression, speech synthesis, and speech recognition.[6] Moreover, this technology allows the machine to replicate the human brain’s ability to selectively focus in a specific sound against many other competing sounds and background noise. This particular ability is called “auditory scene analysis”. The technology enables the machine to segment several streams occurring at the same time.[7] Many commonly used devices such as a smartphones, voice translators, and cars make use of some form of machine hearing.

Machine touch[edit]

Machine touch is an area of machine perception where tactile information is processed by a machine or computer. Applications include tactile perception of surface properties and dexterity whereby tactile information can enable intelligent reflexes and interaction with the environment.

See also[edit]

References[edit]

  1. ^ a b c d Malcolm Tatum (October 3, 2012). "What is Machine Perception".
  2. ^ a b c Alexander Serov (January 29, 2013). "Subjective Reality and Strong Artificial Intelligence" (PDF).
  3. ^ "Machine Perception & Cognitive Robotics Laboratory". www.ccs.fau.edu. Retrieved 2016-06-18.
  4. ^ "Machine Perception Research - ECE - Virginia Tech". www.ECE.VT.edu. Retrieved January 10, 2018.
  5. ^ a b Sagnik Dhar; Vicente Ordonez & Tamara L Berg (2011). "High Level Describable Attributes for Predicting Aesthetics and Interestingness" (PDF). Computer Vision and Pattern Recognition (CVPR), 2011 IEEE. pages 1657 - 1664
  6. ^ a b Richard F. Lyon (September 2010). "Machine Hearing: An Emerging Field". IEEE Signal Processing Magazine.
  7. ^ "Machine Perception & Cognitive Robotics Laboratory". ccs.FAU.edu. Retrieved January 10, 2018.
  8. ^ Turk, Matthew (2000). "Perceptive Media: Machine Perception and Human Computer Interaction" (PDF). Chinese Journal of Computers. 12. pages 1235-1244