Strong AI

From Wikipedia, the free encyclopedia

Strong AI is artificial intelligence that matches or exceeds human intelligence—the intelligence of a machine that can successfully perform any intellectual task that a human being can.[1] It is a primary goal of artificial intelligence research and an important topic for anyone interested in the future, such as science fiction writers and futurists. Strong AI is also referred to as "artificial general intelligence"[2] or as the ability to perform "general intelligent action".[3] Strong AI is also closely related to such traits as sentience, sapience, self-awareness and consciousness.[4]

Some references emphasize a distinction between strong AI and "applied AI"[5] (also called "narrow AI"[1] or "weak AI"[6]): the use of software to study or accomplish specific problem solving or reasoning tasks that do not encompass (or in some cases are completely outside of) the full range of human cognitive abilities.

Contents

[edit] Requirements of strong AI

Many different definitions of intelligence have been proposed (such as being able to pass the Turing test) but there is to date no definition that satisfies everyone.[7] However, there is wide agreement among artificial intelligence researchers that an intelligent machine must at least have the ability to:[8]

Together these skills provide a working definition of what intelligence is, and work is underway to design machines that have these abilities. Strong AI is, at the very least, a system that can perform all these tasks as well as humans do or better.

There are other aspects of the human mind besides intelligence that also bear on the concept of strong AI:

  • consciousness: To say there is something it would "be like" to have strong AI.
  • self-awareness: To be aware of oneself as a separate individual, especially to be aware of one's own thoughts.
  • sentience: The ability to "feel."
  • sapience: the capacity for wisdom.

It is not clear whether any of these are necessary for strong AI—for example, it is not clear if consciousness is necessary for a machine to reason as well as human beings can. It is also not clear whether any of these traits are sufficient for intelligence: if a machine is built with a device that simulates the neural correlates of consciousness, would it automatically have the ability to represent knowledge or use natural language?

It is also possible that some of these properties, such as sentience, naturally emerge from a fully intelligent machine, or that it becomes natural to ascribe these properties to machines once they begin to act in a way that is clearly intelligent. For example, intelligent action may be sufficient for sentience, rather than the other way around.

[edit] Research approaches

[edit] History of mainstream AI research

Modern AI research began in the middle 50s.[9] The first generation of AI researchers were convinced that strong AI was possible and that it would exist in just a few decades. As AI pioneer Herbert Simon wrote in 1965: "machines will be capable, within twenty years, of doing any work a man can do."[10] Their predictions were the inspiration for Stanley Kubrick and Arthur C. Clarke's character HAL 9000, who accurately embodied what AI researchers believed they could create by the year 2001. Of note is the fact that AI pioneer Marvin Minsky was a consultant[11] on the project of making HAL 9000 as realistic as possible according to the consensus predictions of the time, having himself said on the subject in 1967, "Within a generation...the problem of creating 'artificial intelligence' will substantially be solved."[12]

However, in the early 70s, it became obvious that researchers had grossly underestimated the difficulty of the project. The agencies that funded AI became skeptical of strong AI and put researchers under increasing pressure to produce useful technology, or "applied AI".[13] As the eighties began, Japan's fifth generation computer project revived interest in strong AI, setting out a ten year timeline that included strong AI goals like "carry on a casual conversation".[14] In response to this and the success of expert systems, both industry and government pumped money back into the field.[15] However, the market for AI spectacularly collapsed in the late 80s and the goals of the fifth generation computer project were never fulfilled.[16] For the second time in 20 years, AI researchers who had predicted the imminent arrival of strong AI had been shown to be fundamentally mistaken about what they could accomplish.

By the 1990s, AI researchers had gained a reputation for making promises they could not keep. AI researchers became reluctant to make any kind of prediction at all[17] and avoid any mention of "human level" artificial intelligence, for fear of being labeled a "wild-eyed dreamer."[18] This is an unfortunate consequence of developing nascent technologies.

[edit] Mainstream AI research

For the most part, researchers today choose to focus on specific sub-problems where they can produce verifiable results and commercial applications, such as neural nets, computer vision or data mining.[19]

Most mainstream AI researchers hope that strong AI can be developed by combining the programs that solve various subproblems using an integrated agent architecture, cognitive architecture or subsumption architecture. Hans Moravec wrote in 1988 "I am confident that this bottom-up route to artificial intelligence will one day meet the traditional top-down route more than half way, ready to provide the real world competence and the commonsense knowledge that has been so frustratingly elusive in reasoning programs. Fully intelligent machines will result when the metaphorical golden spike is driven uniting the two efforts."[20]

[edit] Artificial general intelligence

Artificial General Intelligence research aims to create AI that can replicate human-level intelligence completely, often called an Artificial General Intelligence (AGI) to distinguish from less ambitious AI projects. (The concept is derived from the psychometric notion of natural general intelligence (often denoted "g")[1], though no adherence to any particular theory of g is implied.) As yet, researchers have devoted little attention to AGI, with some claiming that intelligence is too complex to be completely replicated in the near term. Some small groups of computer scientists are doing AGI research, however. Organizations pursuing AGI include the Adaptive AI, Artificial General Intelligence Research Institute (AGIRI), Singularity Institute for Artificial Intelligence and Texai. One recent addition is Numenta, a project based on the theories of Jeff Hawkins, the creator of the Palm Pilot. While Numenta takes a computational approach to general intelligence, Hawkins is also the founder of the RedWood Neuroscience Institute, which explores conscious thought from a biological perspective.

[edit] Simulated human brain model

Simulated human brain model is one of the quickest means of achieving strong AI, as it doesn't require complete understanding of how intelligence works. Basically, a very powerful computer would simulate a human brain, often in the form of a network of neurons. For example, given a map of all (or most) of the neurons in a functional human brain, and a good understanding of how a single neuron works, it is theoretically possible for a computer program to simulate the working brain over time. Given some method of communication, this simulated brain might then be shown to be fully intelligent. The exact form of the simulation varies: instead of neurons, a simulation might use groups of neurons, or alternatively, individual molecules might be simulated. It's also unclear which portions of the human brain would need to be modeled: humans can still function while missing portions of their brains, and areas of the brain are associated with activities (such as breathing) that might not be necessary to think.

This approach would require three things:

The RIKEN MDGRAPE-3 supercomputer
The RIKEN MDGRAPE-3 supercomputer
  • Hardware. An extremely powerful computer would be required for such a model. Futurist Ray Kurzweil estimates 10 million MIPS, or ten petaflops. At least one special-purpose petaflops computer has already been built (the Riken MDGRAPE-3) and there are nine current computing projects (such as BlueGene/P) to build more general purpose petaflops computers all of which should be completed by 2008, if not sooner.[2] Most other attempted estimates of the brain's computational power equivalent have been rather higher, ranging from 100 million MIPS (100 petaflops) to 100 billion MIPS (100,000 petaflops). Using accurate Top500 projections, it might be estimated that such levels of computing power might be reached using the top-performing CPU-based supercomputers to be by ~2015 (for 100 petaflops), up to a more conservative estimate of ~2025 (for 100,000 petaflops). However, considering that GPU processing and Stream Processing power appears to double every year, these estimates will be reached much sooner using GPGPU processing as high-end GPU's set to arrive in early 2008 are already going to be able to process over 1 teraflop, which is 20x more powerful than a standard quad-core CPU. It should also be noted, however, that the overhead introduced by the modeling of the biological, chemical, and physical details of neural behaviour (especially on a molecular scale) might require a simulator to have access to computational power much greater than that of the brain itself and that current simulations and estimates do not account for the importance of Glial cells which outnumber neurons 10:1.
  • Software. Software to simulate the function of a brain would be required. This assumes that the human mind is the central nervous system and is governed by physical laws. Constructing the simulation would require a great deal of knowledge about the physical and functional operation of the human brain, and might require detailed information about a particular human brain's structure. Information would be required both of the function of different types of neurons, and of how they are connected. Note that the particular form of the software dictates the hardware necessary to run it. For example, an extremely detailed simulation including molecules or small groups of molecules would require enormously more processing power than a simulation that models neurons using a simple equation, and a more accurate model of a neuron would be expected to be much more expensive computationally than a simple model. The more neurons in the simulation, the more processing power it would require.
  • Understanding. Finally, it requires sufficient understanding thereof to be able to model it mathematically. This could be done either by understanding the central nervous system, or by mapping and copying it. Neuroimaging technologies are improving rapidly, and Kurzweil predicts that a map of sufficient quality will become available on a similar timescale to the required computing power. However, the simulation would also have to capture the detailed cellular behaviour of neurons and glial cells, presently only understood in the broadest of outlines.

Once such a model is built, it will be easily altered and thus open to trial-and-error experimentation. This is likely to lead to huge advances in understanding, allowing the model's intelligence to be improved/motivations altered.[dubious ]

The Blue Brain project aims to use one of the fastest supercomputer architectures in the world, IBM's Blue Gene platform, to simulate a single neocortical column consisting of approximately 60,000 neurons and 5km of interconnecting synapses. The eventual goal of the project is to use supercomputers to replicate an entire brain.

The brain gets its power from performing many parallel operations, a standard computer from performing operations very quickly. It should be noted, however, that supercomputers also perform many operations in parallel. Good examples of this are the Cray and NEC vector computers which operate as a single machine but perform many calculations at once. Another example is any form of cluster computing, where multiple single computers operate as one.

The human brain has roughly 100 billion neurons operating simultaneously, connected by roughly 100 trillion synapses.[21] By comparison, a modern computer microprocessor uses only 1.7 billion transistors.[3] Although estimates of the brain's processing power put it at around 1014 neuron updates per second,[22] it is expected that the first unoptimized simulations of a human brain will require a computer capable of 1018 FLOPS. By comparison a general purpose CPU (circa 2006) operates at a few GFLOPS (109 FLOPS). (each FLOP may require as many as 20,000 logic operations).

However, a neuron is estimated to spike 200 times per second (this giving an upper limit on the number of operations).[citation needed] Signals between them are transmitted at a maximum speed of 150 meters per second. A modern 2GHz processor operates at 2 billion cycles per second, or 10,000,000 times faster than a human neuron, and signals in electronic computers travel at roughly half the speed of light; faster than signals in humans by a factor of 1,000,000.[citation needed] The brain consumes about 20W of power whereas supercomputers may use as much as 1MW or an order of 100,000 more (note: Landauer limit is 3.5x1020 op/sec/watt at room temperature).

Neuro-silicon interfaces have also been proposed. [4] [5]

Critics of this approach believe that it is possible to achieve AI directly without imitating nature and often have used the analogy that early attempts to construct flying machines modeled them after birds, but that modern aircraft do not look like birds. The direct approach is used in AI - What is this, where it is shown that if we have a formal definition of AI, it can be found by enumerating all possible programs and then testing each of them to see whether it has produced Artificial Intelligence, or has not.

[edit] Artificial consciousness research

Artificial consciousness research aims to create and study artificially conscious systems. Professor Igor Aleksander[23] argues that the principles for creating a conscious machine already existed but that it would take forty years to train such a machine to understand language.

[edit] Franklin’s Intelligent Distribution Agent

Stan Franklin (1995, 2003) defines an autonomous agent as possessing functional consciousness when it is capable of several of the functions of consciousness as identified by Bernard BaarsGlobal Workspace Theory (Baars 1988), (Baars 1997). His brain child IDA (Intelligent Distribution Agent) is a software implementation of GWT, which makes it functionally conscious by definition. IDA’s task is to negotiate new assignments for sailors in the US Navy after they end a tour of duty, by matching each individual’s skills and preferences with the Navy’s needs. IDA interacts with Navy databases and communicates with the sailors via natural language email dialog while obeying a large set of Navy policies. The IDA computational model was developed during 1996-2001 at Stan Franklin’s "Conscious" Software Research Group at the University of Memphis. It "consists of approximately a quarter-million lines of Java code, and almost completely consumes the resources of a 2001 high-end workstation." It relies heavily on codelets, which are "special purpose, relatively independent, mini-agent[s] typically implemented as a small piece of code running as a separate thread." In IDA’s top-down architecture, high-level cognitive functions are explicitly modeled; see Franklin (1995) and Franklin (2003) for details. While IDA is functionally conscious by definition, Franklin does “not attribute phenomenal consciousness to [his] own 'conscious' software agent, IDA, in spite of her many human-like behaviours. This in spite of watching several US Navy detailers repeatedly nodding their heads saying 'Yes, that’s how I do it' while watching IDA’s internal and external actions as she performs her task."

[edit] Ron Sun's cognitive architecture CLARION

CLARION posits a two-level representation that explains the distinction between conscious and unconscious mental processes.

CLARION has been successful in accounting for a variety of psychological data. A number of well-known skill learning tasks have been simulated using CLARION that span the spectrum ranging from simple reactive skills to complex cognitive skills. The tasks include serial reaction time (SRT) tasks, artificial grammar learning (AGL) tasks, process control (PC) tasks, the categorical inference (CI) task, the alphabetical arithmetic (AA) task, and the Tower of Hanoi (TOH) task (Sun 2002). Among them, SRT, AGL, and PC are typical implicit learning tasks, very much relevant to the issue of consciousness as they operationalized the notion of consciousness in the context of psychological experiments .

The simulations using CLARION provide detailed, process-based interpretations of experimental data related to consciousness, in the context of a broadly scoped cognitive architecture and a unified theory of cognition. Such interpretations are important for a precise, process-based understanding of consciousness and other aspects of cognition, leading up to better appreciations of the role of consciousness in human cognition (Sun 1999). CLARION also makes quantitative and qualitative predictions regarding cognition in the areas of memory, learning, motivation, meta-cognition, and so on. These predictions either have been experimentally tested already or are in the process of being tested.

[edit] Haikonen’s cognitive architecture

Pentti Haikonen (2003) considers classical rule-based computing inadequate for achieving AC: "the brain is definitely not a computer. Thinking is not an execution of programmed strings of commands. The brain is not a numerical calculator either. We do not think by numbers." Rather than trying to achieve mind and consciousness by identifying and implementing their underlying computational rules, Haikonen proposes "a special cognitive architecture to reproduce the processes of perception, inner imagery, inner speech, pain, pleasure, emotions and the cognitive functions behind these. This bottom-up architecture would produce higher-level functions by the power of the elementary processing units, the artificial neurons, without algorithms or programs". Haikonen believes that, when implemented with sufficient complexity, this architecture will develop consciousness, which he considers to be "a style and way of operation, characterized by distributed signal representation, perception process, cross-modality reporting and availability for retrospection." Haikonen is not alone in this process view of consciousness, or the view that AC will spontaneously emerge in autonomous agents that have a suitable neuro-inspired architecture of complexity; these are shared by many, e.g. Freeman (1999) and Cotterill (2003). A low-complexity implementation of the architecture proposed by Haikonen (2004)[verification needed] was reportedly not capable of AC, but did exhibit emotions as expected[citation needed][original research?].

[edit] Self-awareness research

Self-awareness in robots is being investigated by Junichi Takeno [6] at Meiji University in Japan. Takeno is asserting that he has developed a robot capable of discriminating between a self-image in a mirror and any other having an identical image to it[7][8], and this claim has been already reviewed. (Takeno, Inaba & Suzuki 2005)

[edit] Emergence

Main article: emergence

Some[who?] have suggested that intelligence can arise as an emergent quality from the convergence of random, man-made technologies. Human sentience — or any other biological and naturally occurring intelligence — arises out of the natural process of species evolution and an individual's experiences. Discussion of this eventuality is currently limited to fiction and theory.[citation needed][original research?]

[edit] Origin of the term: John Searle's strong AI

See also: philosophy of artificial intelligence and Chinese room

The term "strong AI" was adopted from the name of an argument in the philosophy of artificial intelligence first identified by John Searle as part of his Chinese room argument in 1980.[24] He wanted to distinguish between two different hypotheses about artificial intelligence:[25]

  • An artificial intelligence system can think and have a mind.[26]
  • An artificial intelligence system can (only) act like it thinks and has a mind.

The first one is called "the strong AI hypothesis" and the second is "the weak AI hypothesis" because the first one makes the stronger statement: it assumes something special has happened to the machine that goes beyond all its abilities that we can test. Searle referred to the "strong AI hypothesis" as "strong AI". This usage, which is fundamentally different than the subject of this article, is common in academic AI research and textbooks.[27]

The term "strong AI" is now used to describe any artificial intelligence system that acts like it has a mind,[1] regardless of whether a philosopher would be able to determine if it actually has a mind or not. As Russell and Norvig write: "Most AI researchers take the weak AI hypothesis for granted, and don't care about the strong AI hypothesis."[28] AI researchers are interested in a related statement (that some sources confusingly call "the strong AI hypothesis"):[29]

  • An artificial intelligence system can think (or act like it thinks) as well or better than people do.

This assertion, which hinges on the breadth and power of machine intelligence, is the subject of this article.

[edit] See also

[edit] Notes

  1. ^ a b c (Kurzweil 2005, p. 260) or see Advanced Human Intelligence where he defines strong AI as "machine intelligence with the full range of human intelligence."
  2. ^ Voss 2006
  3. ^ Newell & Simon 1974. This the term they use for "human-level" intelligence in the physical symbol system hypothesis.
  4. ^ These terms are more common in science fiction.
  5. ^ Encyclopedia Britannica Strong AI, applied AI, and cognitive simulation or Jack Copeland What is artificial intelligence? on AlanTuring.net
  6. ^ The Open University on Strong and Weak AI
  7. ^ "We cannot yet characterize in general what kinds of computational procedures we want to call intelligent." John McCarthy, Basic Questions. For a discussion of some definitions of intelligence used by artificial intelligence researchers, see philosophy of artificial intelligence.
  8. ^ This list of intelligent traits is based on the topics covered by major AI textbooks, including: Russell & Norvig 2003, Luger & Stubblefield 2004, Poole, Mackworth & Goebel 1998 and Nilsson 1998.
  9. ^ Crevier 1993, p. 48-50
  10. ^ Simon 1965, p. 96 quoted in Crevier 1993, p. 109
  11. ^ Scientist on the Set: An Interview with Marvin Minsky
  12. ^ Minsky 1967, p. 2 quoted in Crevier 1993, p. 109
  13. ^ The Lighthill report specifically criticized AI's "grandiose objectives" and led the dismantling of AI research in England. (Lighthill 1973) (Howe 1994) In the U.S., DARPA became determined to fund only "mission-oriented direct research, rather than basic undirected research". See (NRC 1999) under "Shift to Applied Research Increases Investment". See also (Crevier 1993, p. 115-117) and (Russell & Norvig 2003, p. 21-22)
  14. ^ Crevier 1993, pp. 211, Russell & Norvig 2003, p. 24 and see also Feigenbaum & McCorduck 1983
  15. ^ Crevier 1993, pp. 161-162,197-203,240, Russell & Norvig 2003, p. 25, NRC 1999 under "Shift to Applied Research Increases Investment"
  16. ^ Crevier 1993, pp. 209-212
  17. ^ As AI founder John McCarthy wrote in his Reply to Lighthill, "it would be a great relief to the rest of the workers in AI if the inventors of new general formalisms would express their hopes in a more guarded form than has sometimes been the case."
  18. ^ "At its low point, some computer scientists and software engineers avoided the term artificial intelligence for fear of being viewed as wild-eyed dreamers."Markoff, John. "Behind Artificial Intelligence, a Squadron of Bright Real People", The New York Times, 2005-10-14. Retrieved on 2007-07-30. 
  19. ^ Russell & Norvig 2003, pp. 25-26
  20. ^ Moravec 1988, p. 20
  21. ^ "nervous system, human." Encyclopædia Britannica. 9 Jan. 2007
  22. ^ Russell & Norvig 2003
  23. ^ Aleksander (1996)
  24. ^ Searle 1980
  25. ^ As defined in a standard AI textbook: "The assertion that machines could possibly act intelligently (or, perhaps better, act as if they were intelligent) is called the 'weak AI' hypothesis by philosophers, and the assertion that machines that do so are actually thinking (as opposed to simulating thinking) is called the 'strong AI' hypothesis." (Russell & Norvig 2003)
  26. ^ The word "mind" is has a specific meaning for philosophers, as used in the mind body problem or the philosophy of mind
  27. ^ Among the many sources that use the term in this way are: Russell & Norvig 2003, Oxford University Press Dictionary of Psychology (quoted in "High Beam Encyclopedia"), MIT Encyclopedia of Cognitive Science (quoted in "AITopics"), Planet Math, Arguments against Strong AI (Raymond J. Mooney, University of Texas), Artificial Intelligence (Rob Kremer, University of Calgary), Minds, Math, and Machines: Penrose's thesis on consciousness (Rob Craigen, University of Manitoba), The Science and Philosophy of Consciousness Alex Green, Philosophy & AI Bernard, Will Biological Computers Enable Artificially Intelligent Machines to Become Persons? Anthony Tongen, and the Usenet FAQ on Strong AI
  28. ^ Russell Norvig, p. 947
  29. ^ A few sources where "strong AI hypothesis" is used this way: Strong AI Thesis, Neuroscience and the Soul

[edit] References

[edit] External links