Misplaced Pages

Artificial consciousness: Difference between revisions

Article snapshot taken from Wikipedia with creative commons attribution-sharealike license. Give it a read and then ask your questions in the chat. We can research this topic together.
Browse history interactively← Previous editNext edit →Content deleted Content addedVisualWikitext
Revision as of 19:28, 6 May 2004 editTkorrovi (talk | contribs)Extended confirmed users1,655 editsm e.g. Chomsky developed a transformational grammar that differed considerably from earlier grammars← Previous edit Revision as of 00:27, 7 May 2004 edit undoPsb777 (talk | contribs)Extended confirmed users9,362 edits =Genuine AC= this section argues the Genuine AC POV. We know Nagel doesn't agree. Essentially you want to have the whole discussion here, in this sectionNext edit →
Line 44: Line 44:
===Genuine AC=== ===Genuine AC===


Proponents of this view believe that artificial consciousness is (or will be) real consciousness, albeit one that has not arisen naturally. The argument in favour of Genuine AC is essentially this: If artificial consciousness is not real consciousness because it is exhibited by a machine then we must assume that the ] is not a ]. If there is something which is not a machine about a human then it must be the ] or a magic spark and the Not-genuine AC argument must then be made in ] or ] terms and science is bypassed. Alternatively, if the human is a machine, then the ] applies and the possibility of Genuine AC must be admitted. This view assumes that anything what (or that) cannot be modelled by AC must be in contradiction with physicalism, but in his "What is it like to be a bat" argues that subjective experience cannot be reduced because it cannot be objectively observed, but subjective experience is not in contradiction with physicalism. Proponents of this view believe that artificial consciousness is (or will be) real consciousness, albeit one that has not arisen naturally. The argument in favour of Genuine AC is essentially this: If artificial consciousness is not real consciousness because it is exhibited by a machine then we must assume that the ] is not a ]. If there is something which is not a machine about a human then it must be the ] or a magic spark and the Not-genuine AC argument must then be made in ] or ] terms and science is bypassed. Alternatively, if the human is a machine, then the ] applies and the possibility of Genuine AC must be admitted.


It is argued that until contradictory evidence is discovered, ] and the ] support the view that AC can be real consciousness and that the building of AC which is real consciousness is likely: The human being is nothing but a machine. The Church-Turing thesis states that we need new ] before two computing machines are different. This means that by machine a virtual machine, like Turing machine, is meant, not AC what (or that?) may not model subjective experience. By Occam's Razor, we should not posit new physics without good reason, but new physics is not necessary when AC is not considered real consciousness. The Copernican principle states that we should claim no special position for human beings without good reason. The only "good" reasons we have are those of arrogance: Humans are supposedly too complicated or special (or some other similar term) for their brains to be built or copied artificially, or for an alternative artificial achitecture to the brain to be truly capable of consciousness. But not considering other than genuine AC a real consciousness would not put human being in a special position, as in that case the model is not complete because of that what we cannot model. It is argued that until contradictory evidence is discovered, ] and the ] support the view that AC can be real consciousness and that the building of AC which is real consciousness is likely: The human being is nothing but a machine. The Church-Turing thesis states that we need new ] before two computing machines are different. This means that by machine a virtual machine, like Turing machine, is meant, not AC what (or that?) may not model subjective experience. By Occam's Razor, we should not posit new physics without good reason, but new physics is not necessary when AC is not considered real consciousness. The Copernican principle states that we should claim no special position for human beings without good reason. The only "good" reasons we have are those of arrogance: Humans are supposedly too complicated or special (or some other similar term) for their brains to be built or copied artificially, or for an alternative artificial achitecture to the brain to be truly capable of consciousness. But not considering other than genuine AC a real consciousness would not put human being in a special position, as in that case the model is not complete because of that what we cannot model.

Revision as of 00:27, 7 May 2004

The neutrality of this article is disputed. Relevant discussion may be found on the talk page. Please do not remove this message until conditions to do so are met. (Learn how and when to remove this message)

Artificial consciousness encompasses digital sentience and simulated consciousness. It is a term that would describe artificial systems (e.g. robots) that simulate some degree of consciousness or sentience.

Simulated consciousness can not be real consciousness, by definition. Yet one school of thought (see below) holds that some AC might be genuinely conscious. Therefore the terms artificial consciousness and simulated consciousness are not equivalent. Digital sentience assumes that the artificial consciousness is exhibited by a computer (or a system with a computer as its "brain"): The possibility of a man-made yet biological (e.g.) system being conscious demonstrates that artificial consciousness is not equivalent to digital sentience.

Description

In computer science, the term digital sentience is used to describe the concept that digital computers could someday be capable of independent thought. Digital sentience, if it ever comes to exist, is likely to be a form of artificial intelligence. A generally accepted criterion for sentience is self-awareness and this is also one of the definitions of consciousness. To support the concept of self-awareness, a definition of conscious can be cited: "having an awareness of one's environment and one's own existence, sensations, and thoughts" (dictionary.com).

In more general terms, an AC system should be theoretically capable of achieving various or by a more strict view all verifiable, known, objective, and observable aspects of consciousness. Another definition of the word conscious is: "being conscious, capable of thought, will, or perception" (dictionary.com).

Aspects of AC

There are various aspects and/or abilities that are generally considered necessary for an AC system, or an AC system should be able to learn them; these are very useful as criteria to determine whether a certain machine is artificially conscious. These are only the most cited, however; there are many others that are not covered.

The ability to predict (or anticipate) foreseeable events is considered a highly desirable attribute of AC by Igor Aleksander: He writes in Artificial Neuroconsciousness: An Update: "Prediction is one of the key functions of consciousness. An organism that cannot predict would have a seriously hampered consciousness." The multiple drafts principle proposed by Daniel Dennett in Consciousness Explained may be useful for prediction: It involves the evaluation and selection of the most appropriate "draft" to fit the current environment.

Consciousness is sometimes defined as self-awareness. While self-awareness is very important, it may be subjective and is generally difficult to test.

Another test of AC, in the opinion of some, should include a demonstration that machine is capable to learn the ability to filter out certain stimuli in its environment, to focus on certain stimuli, and to show attention toward its environment in general. The mechanisms that govern how human attention is driven are not yet fully understood by scientists. This absence of knowledge could be exploited by engineers of AC; since we don't understand attentiveness in humans, we do not have specific and known criteria to measure it in machines. Since unconsciousness in humans equates to total inattentiveness, an AC should have outputs that indicate where its attention is focused at any one time, at least during the aforementioned test.

Awareness could be another required aspect. However, again, there are some problems with the exact definition of awareness. To illustrate this point the philosopher David Chalmers controversially argues that the thermostat could be considered conscious: It knows if it is too hot, too cold, or at the correct temperature. The results of the experiments of neuroscanning on monkeys suggest that a process, not a state or object activates neurons . For such reaction there must be created a model of the process based on the information received through the senses, creating models such way demands a lot of flexibility, and is also useful for making predictions.


Personality is another characteristic that is generally considered vital within consciousness. In the area of behaviorial psychology, there is a somewhat popular theory that personality is an illusion created by the brain in order to interact with other people. It is argued that without other people to interact with, humans (and possibly other animals) would have no need of personalities, and human personality would never have evolved. An artificially conscious machine may need to have a personality capable of expression such that human observers can interact with it in a meaningful way. However, this is often questioned by computer scientists; the Turing test, which measures a machine's personality, is not considered generally useful any more.

Anticipation is the final characteristic that could possibly be used to define artificial consciousness. An artificially conscious machine should be able to anticipate events correctly in order to be ready to respond to them when they occur. The implication here is that the machine needs real-time components, making it possible to demonstrate that it possesses artificial consciousness in the present and not just in the past. In order to do this, the machine being tested must operate coherently in an unpredictable environment, to simulate the real world.

Schools of thought

There are several commonly stated views regarding the plausibility and capability and of AC, and the likelihood that AC will ever be real consciousness. Note that the terms Genuine and Not-genuine refer not to the capability of the artificial consciousness but to its reality (how close it is to real consciousness). Believers in Genuine AC think that AC can (one day) be real. Believers in Not-genuine AC think it never can be real. E.g. Some believers in Genuine AC say the thermostat is really conscious but they do not claim the thermostat is capable of an appreciation of music. This is highly controversial point of view.

Objective less Genuine AC

By "less Genuine" we mean not as real as "Genuine" but more real than "Not-genuine". It is alternative view to "Genuine AC", by that view AC is less genuine only because of the requirement that AC study must be as objective as the scientific method demands, but by Thomas Nagel consciousness includes subjective experience that cannot be objectively observed. It does not intend to restrict AC in any other way.

An AC system must be theoretically capable of achieving all known objectively observable abilities of consciousness possessed by a capable human, even if it does not need to have all of them at any particular moment. Therefore AC is objective and always remains artificial, and is only as close to consciousness as we objectively understand about the subject. Because of the demand to be capable of achieving all these abilities, AC may considered to be a strong artificial intelligence, but this also depends on how strong AI is defined.

Not-genuine AC

Artificial consciousness will never be real consciousness, but merely an approximation of it; it only mimics something that only humans (and some other sentient beings) can truly experience or manifest. Currently, this is the state of artificial intelligence and holders of the Not-genuine AC hypothesis believe that this will always be the case. No computer has been able to pass the somewhat vague Turing test, which would be a first step to an AI that contains a "personality"; this would perhaps be one path to a Genuine AC. By more strict view, subject of another field as AI should not be subject of AC, so by that only a study what(?) cannot be categorized anywhere else, such as artificial emotions, can be considered "Not-genuine AC".

Genuine AC

Proponents of this view believe that artificial consciousness is (or will be) real consciousness, albeit one that has not arisen naturally. The argument in favour of Genuine AC is essentially this: If artificial consciousness is not real consciousness because it is exhibited by a machine then we must assume that the human is not a machine. If there is something which is not a machine about a human then it must be the soul or a magic spark and the Not-genuine AC argument must then be made in religious or metaphysical terms and science is bypassed. Alternatively, if the human is a machine, then the Church-Turing thesis applies and the possibility of Genuine AC must be admitted.

It is argued that until contradictory evidence is discovered, Occam's Razor and the Copernican principle support the view that AC can be real consciousness and that the building of AC which is real consciousness is likely: The human being is nothing but a machine. The Church-Turing thesis states that we need new physics before two computing machines are different. This means that by machine a virtual machine, like Turing machine, is meant, not AC what (or that?) may not model subjective experience. By Occam's Razor, we should not posit new physics without good reason, but new physics is not necessary when AC is not considered real consciousness. The Copernican principle states that we should claim no special position for human beings without good reason. The only "good" reasons we have are those of arrogance: Humans are supposedly too complicated or special (or some other similar term) for their brains to be built or copied artificially, or for an alternative artificial achitecture to the brain to be truly capable of consciousness. But not considering other than genuine AC a real consciousness would not put human being in a special position, as in that case the model is not complete because of that what we cannot model.

Human-like AC

As is to be expected, the Not-genuine and Genuine schools of thought differ on the question of whether artificial consciousness needs to be human-like or whether it could be of an entirely different nature. Proponents of Genuine AC generally hold the view that artificial consciousness does not need to be similar to human consciousness, what is a very questionnable view with no reference to source. Some who hold that artificial consciousness can never be really conscious (i.e., Not-genuine AC proponents) hold that AC, not being real, can only be human-like because that is the only "true" model of consciousness that we will (most likely) ever have, and that this Not-genuine AC will be tested based on what we know about real consciousness.

Nihilistic view

It is impossible to test if anything is conscient. To ask a thermometer to appreciate music is like asking a human to think in five dimensions. It is unnecessary for humans to think in five dimensions, as much as it is irrelevant for thermostats to understand music. Consciousness is just a word attributed to things that appear to make their own choices and perhaps things that are too complex for our mind to comprehend. Things seems to be conscient, but that is just because our morale tells us to believe in it, or because of our feelings for other things. Consciousness is an illusion.

Alternative Views

One alternative view states that it is possible for a human to deny its own existence and thereby, presumably, its own consciousness. That a machine might cogently discuss Descartes' argument "I think, therefore I am", would be some evidence in favour of the machine's consciousness. A conscious machine could even argue that because it is a machine, it cannot be conscious. Consciousness does not imply unfailing logical ability. If we look at the dictionary definition, we find that consciousness is self-awareness: a totality of thought and experience. The richness or completeness of consciousness, degrees of consciousness, and many other related topics are under discussion, and will be so for some time (possibly forever). That one entity's consciousness is less "advanced" than another's does not prevent each from considering its own consciousness rich and complete.

Today's computers are not generally considered conscious. A Unix (or derivative thereof) computer's response to the wc -w command, reporting the number of words in a text file, is not a particularly compelling manifestation of consciousness. However, the response to the top command, in which the computer reports in a real-time continuous fashion each of the tasks it is or is not busy on, how much spare CPU power is available, etc., is a particular if very limited manifestation of self-awareness (and therefore consciousness) by definition.

Testing AC

Unless artificial consciousness can be proven formally, judgments of the success of any implementation will depend on observation. Depending upon one's standpoint regarding what constitues consciousness and the particular attributes of sentience and sapience necessary to demonstrate it, there are various views on what the test acceptance criteria for artificial consciousness should be. Obviously, if the criteria are too stringent then counter-examples citing people who would themselves fail to meet such criteria can be posited to invalidate a particular test. On the other hand, the formulation of a minimum set of requirements, to act as design objectives for any implementation, can be helpful in order to conjecture that the failure to meet all such requirements would result in the failure of the implementation to be deemed conscious by observers.

The Turing test, as mentioned before, is a proposal for identifying machine intelligence as determined by a machine's ability to interact with a person. The Chinese room argument attempts to debunk the validity of the Turing Test by showing that a machine can pass the test and yet not be sapient. In the Turing test one has to guess whether the entity one is interacting with is a machine or a human. An artificially conscious entity could only pass an equivalent test when it had itself passed beyond the imaginations of observers and entered into a meaningful relationship with them, and perhaps with fellow instances of itself.

People would not deny that Christopher Nolan was conscious when he wrote Under The Eye of The Clock, and yet he is unable to move in a coordinated way or to communicate interactively except through eye movement, and then only with people who understand his particular gestures. The book he wrote has been compared with the works of William Butler Yeats and James Joyce, two other highly acclaimed Irish authors. This shows again that AC must be theoretically capable of achieving mental abilities, and test may fail if it doesn't have enough resources like memory or sensors. If test fails on that reason, then it doesn't prove that an entity is not AC, but also doesn't prove otherwise.

Since there is an enormous range of human behaviours, all of which are deemed to be conscious, it is difficult to lay down all the criteria by which to determine whether a machine manifests consciousness. Although objective criteria are being proposed as prerequisites for testing the consciousness of a machine, by the Christopher Nolan argument the failure of any particular test would not disprove consciousness, and it is likely that the final test would be conducted in a similar manner to a Turing test, and the result would hence be subjective.

Anticipation, attentiveness and personality are some of the drivers of the artificially intelligent machine, and there may be others upon which an artful simulation of consciousness will depend.

As artificial consciousness models are defined, designs implemented and models deployed and refined, it is argued that a time will come when artificial consciousness will become indistinguishable from consciousness.

Upon considering whether something qualifies to be called conscious, it may be that mere knowledge of its being a machine would disqualify it (from a human perspective) from being deemed conscious. However, one could then argue that it was conscious of its environment, and it was self-aware of its actual nature: an artificial machine (analogous to our knowledge that we are humans). Artificial consciousness proponents therefore have loosened this constraint and allow that simulating or depicting a conscious machine, such as the robots in Star Wars, could count - not as examples of artificial consciousness, since their personalities are generated by actors, but as models. Hence, some consider that if someone were to produce an artificial C-3PO, which behaved just like the real one, but without needing to be controlled and animated by a human, then it could qualify as being artificially conscious. This argument is very questionnable as it doesn't refer who and where loosened that constraint, it's not known that anybody did that.

Finally, an AC system may fail certain tests simply because the system is not developed to the necessary level or doesn't have enough resources; for example, a computer could fail an attentiveness test simply because it does not have enough memory.

Artificial consciousness as a field of study

Artificial consciousness includes research aiming to create and study artificially conscious systems in order to understand corresponding natural mechanisms.

The term "artificial consciousness" was used by several scientists including Professor Igor Aleksander, a faculty member at the Imperial College in London, England, who stated in his book Impossible Minds that the principles for creating a conscious machine already existed but that it would take forty years to train such a machine to understand language.

Digital sentience has so far been an elusive goal, and a vague and poorly understood one at that. Since the 1950s, computer scientists, mathematicians, philosophers, and science fiction authors have debated the meaning, possibilities and the question of what would constitute digital sentience.

Artificial consciousness in literature and movies

Fictional instances of artificial consciousness:

External links

See also