ArticlePDF Available


Many of the arguments about how to address the hard versus the easy questions of consciousness put by Chalmers (1995) are similar to ones I have developed in Velmans (1991a,b; 1993a). This includes the multiplicity of mind/body problems, the limits of functional explanation, the need for a nonreductionist approach, and the notion that consciousness may be related to neural/physical representation via a dual-aspect theory of information. But there are also differences. Unlike Chalmers I argue for the use of neutral information processing language for functional accounts rather than the term "awareness." I do not agree that functional equivalence cannot be extricated from phenomenal equivalence, and suggest a hypothetical experiment for doing so - using a cortical implant for blindsight. I argue that not all information has phenomenal accompaniments, and introduce a different form of dual-aspect theory involving "psychological complementarity." I also suggest that the hard problem posed by "qualia" has its origin in a misdescription of everyday experience implicit in dualism.
Velmans, Max (1995) The Relation of Consciousness to the Material World Journal of
Consciousness Studies, 2 (3), 255-265
(a commentary on Chalmers, D.(1995) Facing up to the problem of consciousness, JCS, 2(3),
Max Velmans
Department of Psychology
University of London
Lewisham Way
SE14 6NW
KEYWORDS: consciousness, Chalmers, dualism, reductionism, mind/body problem, dual-
aspect, information, functionalism, complementarity, blindsight, cortical implant, qualia
Abstract. Many of the arguments about how to address the hard versus the easy questions
of consciousness put by Chalmers (1995) are similar to ones I have developed in Velmans
(1991a,b; 1993a). This includes the multiplicity of mind/body problems, the limits of
functional explanation, the need for a non-reductionist approach, and the notion that
consciousness may be related to neural/physical representation via a dual-aspect theory
of information. But there are also differences. Unlike Chalmers I argue for the use of neutral
information processing language for functional accounts rather than the term "awareness." I
do not agree that functional equivalence cannot be extricated from phenomenal
equivalence, and suggest a hypothetical experiment for doing sousing a cortical implant
for blindsight. I argue that not all information has phenomenal accompaniments, and
introduce a different form of dual-aspect theory involving "psychological complementarity." I
also suggest that the hard problem posed by "qualia" has its origin in a misdescription of
everyday experience implicit in dualism.
1. Why consciousness is puzzling.
The relation of consciousness to the material world is a puzzle, which has its origin in
dualism, a philosophy of mind which posits their fundamental separation. Dualism, in turn,
has its roots in folk wisdom. The belief that humans are more than bodies and that there is
something in human nature that survives bodily death has its origins in prehistory; it
becomes explicit in the mythology of Ancient Egypt and Assyria and was formulated into a
philosophical position in the Platonic thought of Ancient Greece. But the contemporary view
that the interaction of consciousness with matter poses a problem which may be beyond
scientific understanding can be traced to a clearer formulation of dualism proposed by
According to Descartes (1644) the Universe is composed of two fundamentally different
substances, res cogitans, a substance which thinks, and res extensa, a substance which
extends in space. Res extensa is the stuff of which the material world is made, including
living bodies and brains; res cogitans is the stuff of consciousness. Descartes maintained
that, in humans, res cogitans and res extensa interact via the pineal gland, located in the
centre of the brain. However, even in the seventeenth century, the causal interaction of
substances as different as these was thought by some to pose an insuperable problem.
Leibniz (1686), for example, argued that only physical events could cause physical events
and only mental events could cause mental events. Fortunately, he thought, God has
arranged physical events and mental events into a pre-established harmony so that given
sequences of mental and physical events unfailingly accompany each other ("parallelism").
Consequently, there is an apparent causal interaction of mind with body rather than an
actual one. This view resurfaces in the contemporary assumption that for every conscious
experience there is a distinct neural "correlate." However, attribution of such
correspondences to the workings of a munificent Deity has little appeal to the modern
scientific mind.
Within twentieth century philosophy and science it is far more fashionable to reduce
dualism to a form of materialism, for example to assume or attempt to show that
consciousness is nothing more than a state or function of the brain (physicalism or
functionalism). If either form of reduction is successful the explanatory gap left by dualism
disappears, for the reason that all that needs to be explained can then be explained within
the domain of natural science. Fashion, however, is beginning to change (see, for example,
the debates between Dennett, Fenwick, Gray, Harnad, Humphrey, Libet, Lockwood, Marcel,
Nagel, Searle, Shoemaker, Singer, Van Gulick, Velmans, and Williams in Ciba Foundation
Symposium 174, 1993). The reasons for this are manybut in essence they have to do with
the realization that once one has explained everything that there is to explain about the
material structure and functioning of brains, one will still be left with the problems of
consciousness. To put matters crudely, one cannot find consciousness by any conceivable
histological examination of the brain. Nor, as Nagel (1974) puts it, can one know what it is
like to be something from a physical description alone. In Velmans (1991a) I have considered
functional explanations of consciousness, tracing functional models of the mind through
from input to output and concluded that consciousness cannot be found within any
information processing "box" within the brain. Consciousness accompanies or results from
certain forms of processing but can be dissociated conceptually, and in most cases
empirically from the processes with which it is commonly identified in the cognitive
literature (perception, learning, memory, language, creativity and so on). The same can be
said of models of functioning couched in other terms, such as parallel distributed processing
or the language of neurophysiology (Gray 1995; Velmans 1995a).
In short, while it is likely that consciousness will eventually be found to be associated with
given forms of processing, it looks increasingly likely that consciousness cannot be reduced
to such processing. Or, to put matters another way, "first-person perspective facts" cannot
be fully reduced to "third-person perspective facts" (c.f. Goldman 1993; Velmans 1991a, b,
1993a). In his recent "keynote" article (this issue), Chalmers (1995) comes to the same
But if consciousness cannot be reduced to a state or function of the brain, how might one fill
the explanatory gap left by dualism? Logically, it might be possible to reduce matter to
forms of existence in the mind, for example to argue along with Berkeley (1710) that
material events only exist in so far as they are perceived to exist (idealism). Idealism has its
modern defenders, for example in some interpretations of the observer effect in quantum
mechanics (the view that the Schrödinger wave equation only collapses into an actuality
once an observation is made). In the macro-world it may also be true that the world as-
perceived only exists if there are perceivers (Velmans 1990). However, as a general theory of
the ontology of macro-events this position has its own well-known problems. It might be
that the material world cannot have an appearance without perceivers, but it seems
counterintuitive that its very existence is similarly vulnerable. Closing one's eyes, for
example, does not seem to be enough to make unpleasant events go away.
Given the implausibilities of both materialist and mentalist reductionism, it is important to
consider non-reductionist forms of monism (ways of healing the split induced by dualism,
without reducing either consciousness or the material world to something other than they
appear to be). An early version of this is the "dual-aspect theory" of Spinoza (1677)the
view that mind and matter are manifest aspects of something deeper in Nature, which
appear to interact by virtue of some unfolding, grounding process within Nature itself. This
view may be thought of as a precursor of the contemporary proposal that consciousness
and its correlated brain states may be thought of as dual aspects of a particular kind of
"information," which is in turn, a fundamental property of nature (Velmans 1991b; 1993a;
Chalmers 1995).
2. The interface of consciousness and brain.
We do not yet have precise knowledge of the events which form the interface of
consciousness with the brain. However, in Velmans (1991a, b; 1993a) I considered what the
general character of such events might be. Conscious experiences are representational (they
are about something); consequently it seems likely that the neural or other physical
correlates of conscious experiences will also be representational (i.e. that they will be
representational states). It seems plausible to assume that for every distinct conscious
experience there will be a distinct neural or other physical correlate. If so, any information
(about something) manifest in experience will also be encoded in the correlates in a neural
or other physical form. It follows from these assumptions that given conscious experiences
and their physical correlates encode identical information that is formatted in different
For human beings, of course, the physical correlates of consciousness are formed within
living brains with multiple functions, and there has been much research and theory about
where, in this complex system, consciousness "fits in" (c.f. Velmans 1991a). Information at
the focus of attention usually enters consciousness (like this sentence). Conversely,
information which is not attended to does not enter consciousness (like the pressure of your
tongue against your upper teeth). So the relation of consciousness to focal-attention is
close. But once information at the focus of attention enters consciousness it has already
been analyzed. For example, the processes involved in reading and comprehending this text
are extremely complexinvolving visual pattern recognition, syntactic analysis, semantic
analysis, relating of input information to global knowledge in long-term memory,
judgements about plausibility, conversion into some consciously experienced visual form
and, if it is accompanied by inner speech (verbal thoughts), into phonemic imagery. But all
that enters consciousness is the result of such processing (in the form of seen words,
phonemic imagery, feelings of comprehension and so on).
Given this, I have argued that consciousness relates closely to a late-arising product of focal-
attentive processing. Clues about the nature of this late-arising product come from those
rare situations where focal attention is devoted to input analysis without accompanying
consciousness. One striking example is blindsight, a condition in which subjects are
rendered blind in one half of their visual field as a result of unilateral striate cortex damage.
If stimuli are projected to their blind hemifield subjects cannot see them in spite of the fact
that their full attention is devoted to the task. As they cannot see the stimulus they maintain
that they have no knowledge about it. However, if they are persuaded to make a guess
about the nature of the stimulus in a forced choice task, their performance may be very
accurate. For example, one subject investigated by Weiskrantz et al (1974) was able to
discriminate horizontal from vertical stripes on 30 out of 30 occasions although he could not
see them. In short, the subject has the necessary knowledge but does not know that he
knows. In information processing terms, it is as if one (modular) part of his system has
information which is not generally available throughout the system. On the basis of this and
other evidence I have argued that consciousness relates closely to information
dissemination (Velmans 1991a). Although Chalmers (1995) does not refer to this and related
papers, he again makes the same suggestion (a similar suggestion regarding information
dissemination has also been made by Baars (1988) and Navon (1991).
3. Points of agreement with Chalmers (1995)
In sum, the resemblance of Chalmers' main proposals to my own (in Velmans 1991a, b;
1993a) is strikingalthough the arguments used are sometimes different (and Chalmers
puts the case in a very elegant way). I will not quote from my earlier papers in detail,
however points of agreement include:
1) That the mind/body problem is not one problem but many (see also Velmans 1995b,
2) That exploration of the physical and functional causes of consciousness can be
undertaken by conventional scientific procedures, but that consciousness as such cannot be
captured by such procedures.
3) That the proper route through this is to adopt a non-reductive approach, which relates
first-person accounts of experience to third-person accounts of physical structure and
functioning (psychophysical bridging laws) without seeking to reduce consciousness to brain
or vice-versa.
4) That consciousness may be fundamental, in the sense that matter and energy are
fundamental (if so, a reduction of consciousness to matter is unlikely to be successful).
5) That at the interface of consciousness with the brain, the structure of conscious
representations can be related to the structure of physical representations in the brain
through the notion of information and, in particular, to information dissemination.
6) That this amounts to a form of dual-aspect theory in which conscious experiences and
their physical correlates may be thought of as dual manifestations of something more
fundamental than eitherthe information structure of some thing-itself (there are also
some differences between us on this point, but more of this in section 4.4 below).
4. Points of disagreement with Chalmers (1995).
4.1 Definitions.
Within his keynote article Chalmers suggests that it would be useful to distinguish
"consciousness" from "awareness." The term "consciousness" refers to "experience",
"qualia," or "phenomenal experience." "Awareness," he suggests should refer to
phenomena associated with consciousness, such as the ability to discriminate, categorize,
and react to environmental stimuli, the integration of information by a cognitive system,
and so on (p6). Contributors to this issue have been urged to follow this usageand, if there
is consensus, this might become standard within the field as a whole. Given this, it is
important to examine Chalmers' usage with care.
As noted earlier, I have also argued that consciousness in the sense of "experience" should
be clearly distinguished from the information processing functions with which it is
commonly associated (Velmans 1991a), however Chalmers' choice of the term "awareness,"
for such functions is unfortunate. In common usage, the terms "consciousness,"
"awareness," and "experience" are often interchangeable as are the terms "conscious
awareness" and "conscious experience." More seriously, the use of the term "awareness"
for information processing functions is not theoretically neutral. It suggests (subtly) that
information processes associated with consciousness are themselves in some sense
"sentient." Indeed, Chalmers go on to argue that information processing of certain kinds is
inevitably associated with consciousness of certain kinds, whether or not such processes are
embodied in the brain (p19). This might, or might not be so (we return to this below). But
the matter cannot be decided by defining information processing functions in such a way
that they already have something of the quality of consciousness (in the form of
Blurring of the functioning/sentience distinction also leads to confusion about what
scientific or conceptual advances might illuminate our understanding of consciousness. For
example, many of the "easy" problems of consciousness listed by Chalmers (pp. 5-6) are,
strictly speaking, not problems of consciousness at all. The ability to discriminate,
categorize, and react to environmental stimuli, the integration of information in a cognitive
system, and the ability of a system to access its own internal states, can be accounted for (in
principle) in information processing terms which make no reference to consciousness or
awareness. For example, personal computers which print error messages on a screen have
access, of a kind, to their own internal states. But this rudimentary form of meta-
representation reveals nothing about PC consciousness or awareness.
For reasons such as this a number of other recent reviewers of the field have tended to use
the terms "consciousness," "awareness" and "experience" interchangeably, and to express
information processing functions in neutral information processing terms (see Farthing
1992; readings in Velmans 1996b). This retains the convention within cognitive psychology
that human information processing can be investigated without making any presumptions
about the extent to which it enters consciousness or awareness. It allows the possibility that
various forms of information processing can take place in humans without "conscious
awareness" (in blindsight, visual masking experiments etc.). It also remains open to the
possibility that machines can perform many of the functions performed by humans without
any awareness of what they do.
Chalmers notes that, "The ambiguity of the term 'consciousness' is often exploited by both
philosophers and scientists writing on the subject" enabling them to present theories of
functioning as if they are theories of consciousness (p7). His suggestion that the term
"awareness" should be reserved for such information processing functions perpetuates this
exploitation for theoretical reasons of his own.
4.2 Do all systems with the same functional organization have qualitatively identical
According to Chalmers systems with the same fine-grained functional organization have
identical experiences and, if the functional organization associated with consciousness is
defined to be the system's "awareness," this controversial claim becomes more plausible.
But Chalmers' case is not simply a matter of definition. His principal argument is a thought
experiment which demonstrates that systems which are functionally isomorphic could not
themselves distinguish between having experiences of different kinds. Conversely, if they
could make such distinctions (and noticed or reported different experiences) they would not
be functionally isomorphic.
However, whether functionally isomorphic systems could notice experiential differences
must be distinguished from whether experiential differences in such systems exist. A totally
non-conscious machine, for example, would have no way of noticing that it was totally non-
conscious (the conscious/non-conscious distinction would have no meaning for it). It could
nevertheless be made to respond as if it were consciousfor example, it could be
programmed to report only on information at the focus of attention being disseminated
throughout its system (assuming that in humans only such information is conscious). If the
machine could be made functionally isomorphic to a conscious human being there would be
no way of distinguishing the machine from a human by its behaviour alone. But that would
not alter the fact that in the machine there is nobody at home.
According to Chalmers there is no empirical way to distinguish between such functionally
isomorphic systems. Perhaps for completely isomorphic systems he is right. However, in
human beings, that is not how one would attempt to distinguish conscious from non-
conscious functioning.
A cortical implant for blindsight. In typical psychology experiments on consciousness, one
requires subjective reports, for example about whether a subject has experienced a
stimulus or not. One has to accept that these two potential outcomes are associated with
functional differences, if only for the reason that they are associated with different verbal
reports. Usually, there are additional functional differences in these two situations.
Suppose, however, that we are able to eliminate these additional differences so that
the only functional difference between the two situations is that in one case there is a
conscious experience (of the stimulus) with a corresponding verbal report, and in the other
case there is not.
Suppose also that consciousness is most closely associated with information dissemination
(as suggested above). Recall that in blindsight, subjects can identify stimuli projected to
their blind hemifields without conscious experience, but information about the stimuli does
not appear to be disseminated throughout the brain. Consequently, subjects do not "know
that they know" what the stimuli are and have to be forced to guess.
Imagine that we could implant a cortical network constructed out of silicon chips which
restores information dissemination (of stimuli in the blind hemifield) so that in terms of all
functions other than the appearance of consciousness and its associated verbal
report blindsighted patients can no longer be distinguished from normals. Equipped with
their new implants blindsighted patients now identify stimuli presented to their blind
hemifields with confidence, learn from them, behave appropriately towards them and so
on. But suppose that they still report that only stimuli presented to their non-affected
hemifields are consciously experienced. If so, information dissemination is dissociable from
We cannot second-guess the outcome of this experiment, and in practice this bit of medical
wizardry may turn out to be impossible. But the experiment is not unthinkable. And if it is
not unthinkable, then Chalmers is wrong to argue that functional equivalence (in this case,
in terms of information dissemination) cannot be empirically extricated from phenomenal
4.3 Is all information associated with consciousness?
Chalmers is open to the possibility that not all information has a phenomenal aspect but
argues that it is both plausible and elegant to suggest that it does so. If so, not just mice, but
even thermostats might have some maximally simple experience (p 22). If his argument that
functional equivalence entails phenomenal equivalence were correct, it would follow that
machines that function like humans would also have the experiences of humans.
While I have also argued that consciousness may be related to certain forms of information
in a fundamental way, there are very good empirical grounds for supposing that this does
not apply in every case. In the human brain, for example, there is massive information, but
at any given moment only a small proportion of this has any manifestation in conscious
experience. There is, for example, a vast store of knowledge in long-term memory which at
any given moment remains unconscious. Of the currently processed information, only that
which is at the focus of attention ultimately reaches consciousness (like this sentence); the
details of current information processing do not themselves enter consciousness (like the
processing required to read this sentence). In fact, very few information processing details
ever enter consciousness (Velmans 1991a). Given that human information processing
operates in largely non-conscious fashion it is plausible to assume that it operates in a
similarly non-conscious fashion in machines.
It follows from this that we need to constrain the theory that information (always) has
phenomenal aspects. This leaves open the possibility that information of certain kinds has
phenomenal aspects. But what characterizes such information?
This is an empirical question to which there are no confident answers (as yet). It could be
that information only has a phenomenal aspect if it is embodied in a neurochemical form.
But this would not be a sufficient constraint as most neural representations have no
manifestation in consciousness. An additional constraint might be that neural
representations only become conscious when they have been activated above a given
threshold. Another common hypothesis is that information only becomes conscious when it
is passed to a particular region or circuit in the brain. A third possibility is that information
only becomes conscious when it is subjected to a certain kind of processing (such as
information dissemination, as suggested above). A fourth possibility is that information in
the brain becomes conscious only when the conditions above combine in some wayfor
example, if information activated above some threshold is disseminated throughout the
These suggestions share the assumption that for information to become conscious
something added has to happen. There is, however, another intriguing possibility. It could
be that information has a phenomenal aspect unless it is prevented from doing so. In the
human brain, for example, it could be that information and accompanying consciousness is
massively inhibited, to prevent information (and consciousness) overload. Much of neural
activity is known to be inhibitory. To enable coherent, adaptive response it might be that
only the information at the focus of attention which is disseminated throughout the system
is released from inhibition.
Release from inhibition has been hypothesised to explain certain psychological effects, for
example, the sudden improvement in short-term memory performance, if after a series of
trials with similar stimuli, one changes the features of the to-be-remembered stimuli
(Wickens 1972). There is also evidence that selective attention operates, in part, by the
inhibition of non-attended stimuli (c.f. Arbuthnott 1995). This is consistent with the view
that the brain may act as a filter (as well as an organizer) of information (e.g. Broadbent
1958). If consciousness is a naturally occurring accompaniment of neurally encoded
information (which is massively inhibited) Chalmers' suggestion that
information generally has a phenomenal aspect could not be so easily dismissed.
One would still, of course, have to establish whether consciousness can be decoupled from
neurochemistryfor example, by experiments such as the cortical implant for blindsight
suggested above. And, even if it were found that consciousness accompanies information
embodied in silicon and other physical substances, one would still have to be cautious about
the forms of consciousness that might accompany different forms of information.
Chalmers, for example, suggests that thermostats have some minimal experience. But, of
what? For human purposes, thermostats convey information about temperature. Feelings of
hot and cold also convey information about temperature. Given this, it is possible (in
principle) to device a thermostat which controls temperature in a closed environment in a
way that would be indistinguishable from that of a human controller responding to feelings
of hot and cold. So for these purposes one could create a form of man/machine functional
equivalence. But it would be facile to assume that thermostats (in this situation) feel hot
and cold in the way that humans do.
In humans, feelings of hot and cold are mediated by free nerve endings embedded in the
skin which pass their information to complex circuitry in the thalamus and cerebral cortex. A
simple thermostat might be constructed out of a bimetal strip. Because the expansion
coefficient of one metal is greater than the other, the strip bends as temperature increases.
If the metal bends enough it opens (or closes) a circuit which controls a source of heat. In
short, differential expansion in metals is used by human beings for the purposes of human
beingsto convey information about and control temperature. But the bimetal strip simply
expands and bends. If it does experience something that relates to how it expands and
bends, there is no reason to suppose that this will relate in any way to the feelings of hot
and cold experienced by human beings. As I have argued above, functional equivalence
does not guarantee phenomenal equivalence.
4.4. Dual-aspect theory or naturalistic dualism?
One can only speculate about thermostat consciousness, but few would deny the existence
of human consciousness, or its close association with activities in brains. As noted above, it
is also plausible to link different forms of consciousness to different forms of neural or other
physical encoding (via the notion of information). But this does not explain how, in an
otherwise physical system, consciousness comes to exist. We have seen that the neural
conditions for the appearance of consciousness can in principle be determined empirically.
For example, it might turn out that given combinations of activation, brain location, and
information processing are involved. Isolation of the necessary and sufficient conditions for
the appearance of consciousness in the brain would provide a form of causal explanation.
But that would not provide an understanding of why these neural activities are
accompanied by something so unlike neural activities.
It is worth bearing in mind that nature does not always bother to arrange things so that they
are immediately obvious to humans. Sometimes, for example, apparently dissimilar
phenomena (to our eyes) are fundamentally linked. For example, electric current passing
through a wire produces a surrounding, spatially distributed magnetic field. Conversely,
moving a wire through an existing magnetic field produces current in the wire. In Velmans
(1991b; 1993a) I have suggested that there is some similarly fundamental process taking
place in the brain that unifies certain forms of neural functioning with conscious experience.
As noted in section 1 above, there are different ways in which one can construe the nature
of this consciousness/ brain relationship, for example in dualist or reductionist terms.
However, the simplest non-reductionist way to capture the intimate linkage of
consciousness to its neural correlates is to suppose that they are two aspects of one
representational process (dual-aspect theory). As noted in section 2 above, it is plausible to
think of conscious experiences and their neural correlates as representational states, and to
assume that the information encoded in a given conscious experience and its physical
correlates is identical. Given this underlying identity, why does the information embodied in
conscious experiences and accompanying neural states appear so different? In Velmans
(1991b, 1993a) I have argued that the appearance of this information depends entirely on
the perspective from which it is viewedi.e. it depends on whether the information in
question is being viewed from an external observer's third-person perspective or from the
first-person perspective of the subject.
Suppose a subject S is asked to focus his attention on a cat, while an experimenter E tries to
determine what is going in the subject's brain. Using his visual system aided by sophisticated
physical equipment E might, in principle, be able to view the information about the cat
being formed in S's brain. Given our current knowledge of brains, it is reasonable to assume
that this information will be embodied in a neurophysiological or other physical form. We
also know that for S, the same information will be displayed in the form of a phenomenal
cat out in the world. We do not know how representations in S's brain are translated into a
conscious experience, but we can safely assume that the biological arrangements whereby
information (about the cat) in S's brain is translated into S's experience are very different to
the experimental arrangements used by E for accessing the same information in S's brain.
Consequently, it is not surprising that for S and E the same information appears to be
formatted in very different ways. Note that there is nothing unusual about identical
information appearing in very different formats. This happens, for example, to the
information encoded in the form of magnetic variations on videotape once it is displayed as
a moving picture on a TV screen.
But this is only an analogy. The manner in which different arrangements for observing the
information in S's brain, produce such dramatically different phenomenal consequences, is
more closely reminiscent of wave/particle duality in quantum mechanics. In Velmans
(1991b, 1993a) I have sketched out the beginnings of a "psychological complementarity"
principle (operating at the interface of consciousness with the brain). This resembles
complementarity in physics in its stress on the dependence of the observation on the entire
observation arrangements. Psychological and physical complementarity are also similar in
requiring both (complementary) descriptions for completeness. Both a wave and a particle
description are required for a complete understanding of photons. Likewise, both a (third-
person) neural/physical and a (first-person) phenomenal description are required for a
complete psychological understanding of subjects' representations.
But there the similarities end. I do not suggest for example that consciousness is wavelike,
or that neural correlates are like particles. And crucially, wave/particle manifestations are
both accessible to an external observer, whereas it is a unique feature of the consciousness
-brain relationship that the complementarity obtains between what is accessible to an
external observer versus an experiencing subject.
This account of psychological complementarity is only a first sketch, but attempts to grapple
with how a fundamental representational process might have such different manifestations.
While Chalmers also argues that experience is a fundamental feature of the world, his
analysis of how it relates to physical features is quite different. According to Chalmers,
experience is a fundamental feature alongside mass, charge, and space-time. Consequently
he suggests that the relationship of experience to physical processes may be thought of as a
kind of "naturalistic dualism" (p15). Although elsewhere (p4, p20) Chalmers considers his
theory to be a form of "double-aspect" theory, and he refers to the intrinsic versus the
extrinsic nature of information, the relation of this double-aspect theory to naturalistic
dualism is not made clear.
5. How to bridge a nonexistent gap.
There can be little doubt that some of the questions about consciousness are hardand
that not all the questions are just the result of conceptual confusion. For example, there
really are conscious experiences and we have every reason to suppose that there really are
neural or other physical accompaniments of these. Working out the details of the
brain/consciousness interface is consequently important.
But I believe that there are some "hard" questions that are the result of conceptual
confusion. It must be remembered that the "mind/body problem" has resulted, in part,
from the splitting of mind from body implicit in dualism. Consciousness, in Descartes'
formulation, is nonmaterial and has no location or extension in space. Conscious states are
commonly thought to be private and subjective in contrast to physical states which are
public and objective. Given this, it is hardly surprising that consciousness has been thought
to pose an insuperable problem for sciencewith a consequent shift towards materialist
reductionism within twentieth century philosophy of mind.
In Velmans (1990, 1993b) I have reviewed evidence that Descartes' way of describing
consciousness and its relation to the brain accords neither with science (the nature of
perception) nor with the everyday experiences of human beings. That is to say, the
mind/body problem results in part from a description of everyday experience that does not
correspond to everyday experience. I have also taken issue with the private vs. public and
subjective vs. objective distinctions in the forms that they usually appear.
It is not possible to enter into the detailed analysis relating to these basic points in the
limited space available here. But, to put matters very briefly, very few aspects of
consciousness have the non-extended character suggested by Descartes (thoughts, some
feelings, some images and so on). Classical "mental" events such as pains nearly always have
a location and distribution in one's body, and the same applies to other tactile sensations.
Auditory and visual phenomena are generally perceived to be outside the body and
distributed in three-dimensional space. In short, the contents of consciousness include not
just "inner events" such as thoughts, but body events, and events perceived to be in the
external phenomenal world. This external phenomenal world is what we normally think of
as the "physical world." In short, what we normally think of as the "physical world" is part-
of the contents of consciousness (part-of what we experience). Consequently, there is no
unbridgeable divide separating "physical phenomena" from the "phenomena we
Consider how this description applies to the situation described in 4.4 above, in which an
experimenter observes the brain of a subject. While the experimenter focuses on the
subject, his phenomenal world includes the subject, and if his experimental arrangement is
successful, the representational states in the brain of the subject. While the subject focuses
on the cat his phenomenal world includes the cat. It is fashionable (at present) to think of
E's "observations" (of the subject's brain) as public and objective. S's "experiences" of the
cat, by contrast, are private and subjective. Indeed this radical difference in the status of E
and S is enshrined in the different terminology applied to what they perceive; that is, E
makes "observations," whereas S merely has "subjective experiences."
But suppose they turn their heads, so that E switches his attention to the cat, while S
switches his attention to what is going on in E's brain. Now E is the "subject" and S is the
"experimenter." Following the same convention, S would now be entitled to think of his
observations (of E's brain) as public and objective and to regard E's observations of the cat
as private and subjective. But this would be absurdas nothing has changed in the
character of the observations of E and S other than the focus of their attention.
I cannot do more in a few lines than to sew a few seeds of doubt. A full pursuit of such
simple points requires a reanalysis of idealism versus realism, of private versus public, of
subjectivity, intersubjectivity and objectivity, of the relation of psychology to physics, and of
the proper nature of a science of consciousness (Velmans 1990, 1993b; Harman 1993).
Suffice it to say that, in the new analysis, the problem of how to incorporate conscious
"qualia" into empirical science disappears. All phenomena in science are seen to be aspects
of the phenomenal worlds of observers, and therefore part-of the world that they
experience. "Qualia" are in there from the beginning. Indeed, the whole of science may be
seen as an attempt to make sense of the phenomena that we observe or experience.
Arbuthnott, K.D. 1995. Inhibitory mechanisms in cognition: Phenomena and models. Cahiers
de Psychologie Cognitive 14(1):3-45.
Baars, B.J. 1988. A Cognitive Theory of Consciousness. Cambridge University Press.
Berkeley, G. 1710/1942. The principles of human knowledge. Reissued by T. Nelson & Sons:
Broadbent, D.E. 1958. Perception and communication. Pergamon Press.
Chalmers, D. 1995. Facing up to the problem of consciousness. Journal of Consciousness
Studies 1(2):5-23.
Ciba Foundation Symposium 174. 1993. Theoretical and Experimental Studies of
Consciousness. Wiley, Chichester.
Descartes, R. 1644/1972. Treatise on Man. Trans. by T.S.Hall. Harvard University Press.
Farthing, W. 1992. The Psychology of Consciousness. Prentice-Hall.
Goldman, A.J. 1993. Consciousness, folk psychology, and cognitive science. Consciousness
and Cognition 2:364-382.
Gray, J. 1995. The contents of consciousness: A neuropsychological conjecture. Behavioral
and Brain Sciences 18: 659-722.
Harman, W. 1993. Towards an adequate epistemology for the scientific exploration of
consciousness. Journal of Scientific Exploration 7(2):133-143.
Leibniz, G.W. 1686/1923. Discourse of Metaphysics, Correspondence with Arnauld, and
Monadology. Trans. by M. Ginsberg, London: Allen & Unwin.
Nagel, T. 1974. What is it like to be a bat? Philosophical Review 83: 435-451.
Navon, D. 1991. The function of consciousness or of information? Behavioral and Brain
Sciences 14: 690-691.
Neill, W.T. 1993. Consciousness, not focal attention, is causally effective in human
information processing. Behavioral and Brain Sciences 16(2): 406-407.
Spinoza, B. 1677/1876. The Ethics. Reprinted in The Ethics of Benedict Spinoza, New York:
Van Nostrand.
Velmans, M. 1990. Consciousness, brain, and the physical world. Philosophical
Psychology 3(1): 77-99.
Velmans, M. 1991a. Is human information processing conscious? Behavioral and Brain
Sciences 14(4): 651-669.
Velmans, M. 1991b. Consciousness from a first-person perspective. Behavioral and Brain
Sciences 14(4): 702-726.
Velmans, M. 1993a. Consciousness, causality, and complementarity, Behavioral and Brain
Sciences 16(2): 404-416.
Velmans, M. 1993b. A reflexive science of consciousness. In Experimental and theoretical
studies of consciousness. Ciba Foundation Symposium No.174, Wiley, Chichester.
Velmans, M. 1995a. The limits of neurophysiological models of consciousness. Behavioral
and Brain Sciences 18(4): 702-703.
Velmans, M. 1995. Consciousness, Theories of. In M. Arbib (ed) The Handbook of Brain Theory
and Neural Networks, MIT Press, pp. 247-250.
Velmans, M. 1996a. An introduction to the science of consciousness. In M. Velmans
(ed.) The Science of Consciousness: Psychological, Neuropsychological and Clinical Reviews,
Routledge, pp. 1-22.
Velmans, M. (ed.) 1996b. The Science of Consciousness: Psychological, Neuropsychological
and Clinical Reviews, Routledge.
Weiskrantz, L., Warrington, E.K., Sanders, M.D. & Marshall,J. 1974. Visual capacity in the
hemianopic field, following a restricted occipital ablation. Brain 97:709-728.
Wickens, D.D. 1972. Characteristics of word encoding. In A.W. Melton & E. Martin,
(eds.) Coding Processes in Human Memory, John Wiley & Sons.
... One problem with this is that without any consciousness at all, there is no possibility of awareness or any psychological reality (Searle 1997). Velmans (1995) thinks Chalmers's choice of the term 'awareness' for information processing functions is unfortunate. In common usage, the terms 'consciousness', 'awareness' and 'experience' are often interchangeable as are the terms 'conscious awareness' and 'conscious experience'. ...
... It seems objectively insane that it should, and yet it does. (p. 2) Velmans (1995) rightly disagrees: Many of the 'easy' problems of consciousness listed by Chalmers … are, strictly speaking, not problems of consciousness at all. Velmans (1995:5) does not exclude functionality from the phenomenal and he does not use the term 'awareness' used by Chalmers. ...
Full-text available
The theory of evolution makes sense of the emergence of consciousness. Reduction is not wrong as such, but must not be totalised. The fact that we are star stuff does not preclude the novelty of consciousness. Materialism is naturalism, but naturalism need not be materialism. Neural pathways are relevant but are not the total picture. The central thesis is about David Chalmers’s philosophy being based on an exclusive disjunction. An inclusive disjunction is, when explained, more appropriate. Functionalism is appropriate. Thomas Nagel’s philosophy on first person ontology can still be maintained. Quantum and complexity theories’ hypothesis on consciousness is more compatible with freedom of decision than classical theories.
... At higher level (universe level), Coarse matter is absent and only two things responsible for all the activities i.e. subtle matter (mind), and pure consciousness. At the highest level, both coarse matter and subtle matter are absent and it is made by pure consciousness [Velmans 1997]. This pure consciousness is responsible for all the activities there. ...
The purpose of the book is to explain the basic concepts of Engineering Ethics, moral values and consciousness. Taking ethical decision is not only depending upon the circumstances, values and culture, but depend upon the individual consciousness. This book provides an introduction to the professional ethics more specific the engineering ethics and moral dilemma which an engineer faces in his professional and personal life. These issues are illustrated with the relevant case studies of engineering practices. This book provides sufficient material for a full course to engineering ethics and moral values with sufficient flavour for consciousness. It can also be used for a 2-3 week modular course on engineering ethics specially related to general professional ethics, human values, applied philosophical ethics, ethics and modern technology, etc. Contents : Contents 1. Human Values 2. Professional Ethics 3. Engineering Ethics 4. Corporate and Business Ethics 5. Consciousnesses, Mind and Brain Study 6. Measurement of Consciousness 7. All Round Development for Well Being
... Res extensa is the substance of which the material world is made, while res cogitans is the substance of consciousness, the non-material, the thinking entity. According to Descartes res cogitans and res extensa interact in the brain, and are located in the centre of the brain (Velmans, 1995). Very often the mind-body problem is traced back to René Descartes and many leave out a reference to Galileo Galilei and his influential book the Assayer (Galilei, 1623). ...
Full-text available
To understand the experience of being present somewhere else, via a digital environment, we start by considering how we can experience being anywhere. We present several different philosophical and psychological perspectives on this, stressing the importance of perception. Each has something to offer and add to our understanding of digital travel. We compare four philosophical views: representationalism, relationism, enactivism and the sense-data view. Each has its strengths and weaknesses, but relationism is best placed to accommodate perceptual illusions, which is a prevalent view of the psychological nature of telepresence experiences. As suggested by enactivism and the direct perception approach, the possibilities for action in the world are important to the nature of our experience of places. This, in turn, is influenced by the characteristics of the world in which we act, through affordances.
... The target article and my reply to commentaries in Velmans (1991b) also introduced a dual-aspect theory of information processing-a form of ontological monism combined with epistemological dualism reminiscent of the work of Fechner (1860) to which we return below. This, and many other non-reductive features of my BBS analysis were also later adopted (and popularised) by the philosopher David Chalmers (1995) in ways outlined in Velmans (1995)-an invited commentary on his target article in the Journal of Consciousness Studies. 8 This form of ontological monism combined with epistemological dualism also allows one to understand why, in many contexts, it makes perfect sense to speak of consciousness-brain causal interactions, while at the same time accepting that, viewed from a purely third-person perspective, the physical world appears causally closed. ...
Full-text available
This chapter examines the integrative nature of reflexive monism (RM), a psychological/philosophical model of a reflexive, self-observing universe that can accommodate both ordinary and extraordinary experiences in a natural, non-reductive way that avoids both the problems of reductive materialism and the (inverse) pitfalls of reductive idealism. To contextualize the ancient roots of the model, the chapter touches briefly on classical models of consciousness, mind and soul and how these differ in a fundamental way from how mind and consciousness are viewed in contemporary Western philosophy and psychological science. The chapter then travels step by step from such contemporary views towards reflexive monism, and towards the end of the chapter, to more detailed comparisons with Hindu Vedanta and Samkhya philosophy and with Cosmopsychism (a recently emergent, directly relevant area of philosophy of mind). According to RM there never was a separation between what we normally think of as the “physical world” and what we think of as our “conscious experience”. In terms of its phenomenology, the phenomenal physical world is part-of conscious experience not apart-from it. This phenomenal world can be thought of as a biologically useful representation of what the world is like, although it is not the world as-described-by modern physics, and it is not the thing itself—supporting a form of indirect (critical) realism. The analysis then outlines how 3D phenomenal worlds are constructed by the mind/brain, focusing specifically on perceptual projection, and then demonstrates how normal, first-person conscious experiences (e.g. of phenomenal worlds) and their associated, third-person viewable neural correlates can be understood as dual manifestations of an underlying psychophysical mind, which can, in turn, be understood as a psychophysical form of information processing. This dual-aspect monism combines ontological monism with a form of epistemological dualism in which first- and third-person perspectives on the nature of mind are complementary and mutually irreducible—a principle that turns out to have wide-ranging applications for the study and understanding of consciousness. The chapter then considers the evolution and wider distribution of consciousness (beyond humans) through a brief analysis of the many forms of discontinuity theory versus continuity theory and argues that to avoid the “hard problem” of consciousness one may need to treat its existence as fundamental, and, as co-evolving with the evolution of its associated material forms. This, in turn, takes one to a central issue: What does consciousness actually do? The analysis argues that its central function is to real-ize existence (to know it in a way that makes it subjectively real). With these foundations in place we then come to the heart of the essay—the ways in which reflexive monism provides a very different view of the nature of the universe to those offered either by dualism or materialist reductionism. As summarised in the last paragraph of this section, “In this vision, there is one universe (the thing-itself), with relatively differentiated parts in the form of conscious beings like ourselves, each with a unique, conscious view of the larger universe of which it is a part. In so far as we are parts of the universe that, in turn, experience the larger universe, we participate in a reflexive process whereby the universe experiences itself.” The essay then considers the precise ways in which this reflexive monist understanding of “consciousness” and “mind” relates to later developments in Vedic philosophy and suggests a way of bridging contemporary Western and classical Vedic ways of understanding consciousness and mind. Finally, the chapter considers what can be said of mystical experience and the ground of being, following the principle that this ground must have the power to both manifest the universe in the form that science shows it to be and our ability to experience the universe in the way that we do. In this, RM is shown to be a dual-aspect monist form of cosmopsychism—a recent area of development within philosophy of mind. The essay compares and contrasts this with idealist versions of cosmopsychism and argues that RM allows for an integrated understanding of realism versus idealism, dualism versus monism, how ordinary experience relates to mystical experience, and how consciousness relates to mind. RM also provides an ‘open’ conceptual system that can, in principle, incorporate a range of parapsychological effects.
... Unfortunately, so far, no distinct answer has been found for this question . What is distinctive about how modern scientists and scholars are approaching this issue is that they are treating it as a scientific problem rather than a spiritual one and they approach the neuroscience for answers [4][5][6][7][8][9][10][11][12][13][14]. On the other hand, consciousness cannot be faced using only the scientific approach saying that consciousness is a physical state, as it sounds like saying software is a laptop. ...
Full-text available
Does pure consciousness exist without being hooked to a physical mechanism (e.g. human body)? Can such claim be proven logically? The magnitude of asking this sort of question is similar to asking: Is it logical that matter exists out of the total void? The answer to both questions is yes. The aim of this paper is to show that, the existence of pure consciousness is a logical state, it is not energy, and it exists timelessly and can be experienced beyond the physical body. Furthermore, we show that pure consciousness (the soul) is a virtual junction on the metaphysical virtual net where each junction incorporates a set of cords of consciousness. The set can be limited, unlimited or uncountable. Cords of consciousness is treated as brain neural patterns and as such the brain can easily be manipulated by an external device which may transmit neural patterns in order to generate artificial scenes. A computation example was provided to elaborate the neural pattern and cord of consciousness. Finally, we discuss the opportunity rather than the possibility to visit the metaphysical virtual space. In that sense, if it is possible, should we grab the opportunity and expend our science beyond physics, into the metaphysics?
... The target article and my reply to commentaries in Velmans (1991b) also introduced a dual-aspect theory of information processing-a form of ontological monism combined with epistemological dualism reminiscent of the work of Fechner (1860) to which we return below. This, and many other non-reductive features of my BBS analysis were also later adopted (and popularised) by the philosopher David Chalmers (1995) in ways outlined in Velmans (1995)-an invited commentary on his target article in the Journal of Consciousness Studies. 8 This form of ontological monism combined with epistemological dualism also allows one to understand why, in many contexts, it makes perfect sense to speak of consciousness-brain causal interactions, while at the same time accepting that, viewed from a purely third-person perspective, the physical world appears causally closed. ...
Full-text available
This is an Introduction to "Is the Universe Conscious? Reflexive Monism and the Ground of Being" an invited chapter for E. Kelly and P. Marshall (eds.) Consciousness Unbound. Rowman and Littlefield (2021--in press). The presentation introduces ancient dualist and monist attempts to understand how consciousness, mind, self and soul relate to brain, body and the material world. It then contrasts modern dualist and materialist- reductionist visions of the universe with reflexive monism, the view that the universe is reflexive and self-observing. It explains ways in which the latter requires a paradigm shift in which consciousness and its material accompaniments are entirely natural consequences of what the universe is like, exemplified by the dual-aspect monist nature of the human mind and that of other sentient beings. The presentation then introduces different ways to explore the ground of being, and makes connections with Eastern as well as Western philosophy. The presentation is publicly available at
Wow! What a title! An author daring to write something under such a title immediately makes himself look ridiculous. He must either be totally ignorant of what is going on in the field, or have up his sleeve something like a panacea or ‘philosophical stone’ — a recipe for the solution of a problem on which the greatest minds of mankind have broken their brains. The author’s arrogant impudence is even more emphasized by his double affiliation, which appears rather suspicious and not suited to an attack on such an impregnable bulwark that has excited and tried human intellectual capacity since the very outset of civilization. Moreover, if this is just another essay or review, it is doubtful that something new in principle can be found: everything has already been masticated many times over. Therefore, it may seen that only a lazy curiosity could make a reader go on reading the paper. The author can only declare that the title’s intention is quite serious and this presentation is not a review.
Full-text available
This overview of Consciousness Studies examines the conditions that one has to satisfy to establish a scientific investigation of phenomenal consciousness. Written from the perspective experimental psychology, it follows a two-pronged approach in which traditional third-person methods for investigating the brain and physical world are complementary to first-person methods for investigating subjective experience allowing the possibility of finding “bridging laws” that relate such first- and third-person data to each other. Mindful of the relative sophistication of third-person methods the chapter focuses on the problems of developing similarly sophisticated first-person methods. The problems are of three kinds: (1) Epistemological problems: How can one obtain public, objective knowledge about private, subjective experiences? (2) Methodological problems: Given that one cannot attach measuring instruments directly up to experiences, what psychological “instruments” and procedures are appropriate to their study? (3) The relation of the observer to the observed: The more closely coupled an observer is with an observed, the greater the potential influence of the act of observation on the nature of the observed (“observer effects”). Given this, how can one develop introspective and phenomenological methods where the observer is the observed? The chapter argues that the epistemological problems are more apparent than real, although this requires one to construe what is private versus public, and what is subjective or intersubjective versus what is objective in a slightly different way—with some enabling consequences for a science of consciousness. Methodological problems are real, but not fundamentally different to the problems traditionally faced in experimental psychological investigations of mental phenomena. The close-coupling of observer with the observed in first-person investigations can also be a problem, producing “observer effects” that are more acute than in most third-person investigations. The chapter suggests that one can either try to minimise such effects or to harness them, depending on the purpose of the investigation.
Full-text available
This paper reviews two authors who are relevant in contemporary philosophy of mind because of their antireductionist arguments and theses regarding mind-body problem. Colin McGinn is the best representative of new mysterianism, position that considers the mind-body connection as a mystery in virtue of a cognitive closure which prevents us from understanding that link. On the other hand, Max Velmans proposes a double-aspect theory in the tradition of Spinoza called reflexive monism. After reviewing these approaches, some objections will be reviewed as well.
Empirical genetic research strongly indicates that particular proteins comprise physical substrates of the qualities which, in turn, constitute consciousness. From the perspective of a theorist, the proteins are simply molecules, but from the perspective of a conscious individual, the characteristics of the proteins constitute the basic elements of qualities that, after processing by the cognitive structures of the brain, reach the level of consciousness. As consciousness, the qualities are normally taken for granted. The development of consciousness is described as beginning with noumenal environmental objects which convey energies that give rise to sensory inputs and qualities. Qualities, in turn, are organized into phenomenal objects that are then referred to the environment as the objects of perception and consciousness. Thus, consciousness is considered to be physical and multiple, a collection of qualities, rather than a single, ill-defined, all-encompassing entity.
Full-text available
This introductory chapter was written in 1995, for a new book of review articles on the emerging science of consciousness, specifically aimed at undergraduate and postgraduate students by experts in the relevant fields. Following on a brief history, the chapter moves on to definitions of consciousness and background philosophical issues, and then introduces a unified, non-reductionist scientific approach. It then summarises major issues for studies of consciousness in cognitive psychology, including studies of attention, memory, the extent of preconscious analysis and the relation of consciousness more generally to human information processing. It then turns to the neuropsychology of consciousness, starting with some apparent neural requirements for the transition from preconscious to conscious states, various clinical dissociations of consciousness, conditions for integration (or binding) and, finally, clinical applications, including different forms of mind/body interaction and evidence for the causal efficacy of mental states. The chapter concludes that while some of the ancient problems of consciousness remain unsolved, its study has become the subject of a rapidly developing science.
Full-text available
This commentary elaborates on Gray's conclusion that his neurophysiological model of consciousness might explain how consciousness arises from the brain, but does not address how consciousness evolved, affects behaviour or confers survival value. The commentary argues that such limitations apply to all neurophysiological or other thirdperson perspective models. To approach such questions the first-person nature of consciousness needs to be taken seriously in combination with third-person models of the brain.
To make progress on the problem of consciousness, we have to confront it directly. In this paper, I first isolate the truly hard part of the problem, separating it from more tractable parts and giving an account of why it is so difficult to explain. I critique some recent work that uses reductive methods to address consciousness, and argue that these methods inevitably fail to come to grips with the hardest part of the problem. Once this failure is recognized, the door to further progress is opened. In the second half of the paper, I argue that if we move to a new kind of nonreductive explanation, a naturalistic account of consciousness can be given. I put forward my own candidate for such an account: a nonreductive theory based on principles of structural coherence and organizational invariance and a double-aspect view of information.
Drawing on previous models of anxiety, intermediate memory, the positive symptoms of schizophrenia, and goal-directed behaviour, a neuropsychological hypothesis is proposed for the generation of the contents of consciousness. It is suggested that these correspond to the outputs of a comparator that, on a moment-by-moment basis, compares the current state of the organism's perceptual world with a predicted state. An outline is given of the information-processing functions of the comparator system and of the neural systems which mediate them. The hypothesis appears to be able to account for a number of key features of the contents of consciousness. However, it is argued that neitherthis nor any existing comparable hypothesis is yet able to explain why the brain should generate conscious experience of any kind at all.
Reviews behavioral evidence of inhibition in cognitive processing and addresses the question of what inhibitory mechanisms are responsible for these effects. Inhibitory effects from 5 paradigms (negative priming, lexical priming of ambiguous words, negative error priming, directed forgetting, and stop-signal tests) in the domains of selective attention, language and number processing, and memory and response management are reviewed. Two mechanisms, lateral inhibition and self-inhibition, are described as potential sources of the reviewed effects, and possible means by which intention could influence these inhibitory processes are considered. (PsycINFO Database Record (c) 2012 APA, all rights reserved)
The psychology of consciousness covers a wide range of topics related to normal waking consciousness and altered states of consciousness, including: characteristics of consciousness, factors that influence the stream of consciousness, the distinction between conscious and nonconscious mind, the relationship between the brain and consciousness, introspection, daydreaming, sleep, dreams, hypnosis, meditation, psychedelic drug states, and other topics listed in the first chapter. I approach the psychology of consciousness from a natural science and cognitive psychology viewpoint. In this view, consciousness is a natural phenomenon—a product of the brain's functioning. I emphasize research on topics of interest to cognitive psychologists, cognitive neuropsychologists, and personality and social psychologists. I also discuss clinical applications for a number of topics, such as hypnosis and meditation. The incentive to write this textbook developed over a period of years during which I have taught a course on the psychology of consciousness at the University of Maine. . . . It became apparent that a comprehensive textbook on the psychology of consciousness was needed. Thus, the major reason for writing this book is to provide, in a single volume, a review of research and theory on the psychology of consciousness at a level suitable for advanced college courses. (PsycINFO Database Record (c) 2012 APA, all rights reserved)