ArticlePDF Available

Abstract and Figures

In 4 experiments, we investigated the effects of object affordance in reach-to-grasp actions. Participants indicated whether a depicted small or large object was natural or manmade by means of different object-grasping responses (i.e., with a power or a precision grip). We observed that the size of the depicted object affected the grasping kinematics (grip aperture) and the reach-onset times of compatible and incompatible actions. Additional experiments showed that the effect of perceived object size on motor response was modulated by contextual action information and the observation of others' actions with the object. Thus, beyond the observation of object affordance effects in natural grasping actions, this study suggests that the coupling between object perception and action is not static and obligatory. Behavioral effects of action-relevant object features seem rather to depend on contextual action information.
Content may be subject to copyright.
Context Effects on the Processing of Action-Relevant Object Features
Giovanna Girardi
University of Rome “La Sapienza”
Oliver Lindemann and Harold Bekkering
Radboud University Nijmegen
In 4 experiments, we investigated the effects of object affordance in reach-to-grasp actions. Participants
indicated whether a depicted small or large object was natural or manmade by means of different
object-grasping responses (i.e., with a power or a precision grip). We observed that the size of the
depicted object affected the grasping kinematics (grip aperture) and the reach-onset times of compatible
and incompatible actions. Additional experiments showed that the effect of perceived object size on
motor response was modulated by contextual action information and the observation of others’ actions
with the object. Thus, beyond the observation of object affordance effects in natural grasping actions, this
study suggests that the coupling between object perception and action is not static and obligatory.
Behavioral effects of action-relevant object features seem rather to depend on contextual action infor-
mation.
Keywords: object affordance, action context, object grasping, action observation, action intention
Over the past decade, cognitive science has shown increased
interest in understanding the relation between the functional pro-
cesses necessary to initiate a goal-directed action and the processes
essential for perception and thought. For example, compatibility
effects between object perception and motor response have been
shown to be bidirectional. That is, stimulus features can affect the
characteristics of potential actions (i.e., stimulus–response com-
patibility; e.g., Hommel, 1995; Kornblum, Hasbroucq, & Osman,
1990; Simon, 1969), and characteristics of a prepared or executed
action can influence the perception of stimulus features (i.e.,
response–stimulus compatibility; e.g., Craighero, Fadiga, Rizzo-
latti, & Umilta`, 1999; Fagioli, Ferlazzo, & Hommel, 2007; Fagioli,
Hommel, & Schubotz, 2007; Mu¨sseler & Hommel, 1997). Two
dominant theoretical views of the coupling between object percep-
tion and action can be distinguished: theories of direct perception
and theories of ideomotor action.
Theories of direct perception assume that perceptual processes
are intimately related to motor processes and claim that people
perceive each object in their environment in terms of potentially
afforded behaviors (e.g., Gibson, 1979). Gibson (1979) argued that
the affordances of objects are based on their intrinsic perceptual
properties, registered automatically and without the need for fur-
ther cognitive processes such as object recognition. To test this
notion, Tucker and Ellis (2001) required their participants to
indicate the semantic category of natural and manmade objects by
mimicking either a full or a precision handgrip. They found a
compatibility effect between the size (large or small) of the pre-
sented object and the required response (full or precision hand-
grip). They interpreted their findings as an object affordance effect
reflecting the directly perceived relation between certain visual
object properties and possible motor responses (see also Derby-
shire, Ellis, & Tucker, 2006; Ellis & Tucker, 2000; Ellis, Tucker,
Symes, & Vainio, 2007; Vainio, Symes, Ellis, Tucker, & Ottoboni,
2008).
Interestingly, Glover, Rosenbaum, Graham, and Dixon (2004)
demonstrated that interference effects between object properties
and motor responses are not only present for simple button-press
responses but can also be found in natural object– directed reach-
to-grasp movements. To be precise, they investigated the influence
of object words on the movement kinematics of grasping actions
and observed larger maximum grip apertures after reading words
representing relatively large objects (e.g., APPLE) than after read-
ing words representing relatively small objects (e.g., GRAPE). As
a more detailed analysis revealed, the object affordance effects in
word reading on grasping kinematics were already present very
early in the reach, suggesting that the effect of the object words
emerged during action planning and online motor control (see also
Glover, 2004). Taking into account the finding of automatic word-
reading effects, it seems plausible to assume that the processing of
visual object features (e.g., object size) also affects the kinematics
of natural reach-to-grasp actions. However, until now there has
been no behavioral evidence for such an impact of visual action-
related object information on grasping kinematics that went be-
yond response latency measurements and a facilitated execution of
compatible motor response.
The finding of object affordance effects on button-press laten-
cies has been interpreted as support for the idea that perceived
object affordances automatically and obligatorily affect the plan-
ning of subsequent motor responses. However, the notion that the
processing of visual object information takes place in an automatic
fashion does not imply that simply viewing graspable objects
automatically potentiates components of the actions they afford.
Giovanna Girardi, Department of Psychology, University of Rome “La
Sapienza”; Oliver Lindemann and Harold Bekkering, Donders Institute for
Brain, Cognition and Behavior, Radboud University Nijmegen.
We thank Gabriella Antonucci and Shirley-Ann Rueschemeyer for their
comments on earlier versions of the article. This research was supported by
the ICIS project sponsored by the Dutch Ministry of Economic Affairs
(Grant BSIK03024). Giovanna Girardi and Oliver Lindemann put an equal
amount of work into this project and therefore share first authorship.
Correspondence concerning this article should be addressed to Giovanna
Girardi, Department of Psychology, University of Rome “La Sapienza,” via
dei Marsi, no. 78, 00185 Rome, Italy. E-mail: giovanna.girardi@uniroma1.it
Journal of Experimental Psychology: © 2010 American Psychological Association
Human Perception and Performance
2010, Vol. 36, No. 2, 330–340
0096-1523/10/$12.00 DOI: 10.1037/a0017180
330
Empirical evidence for this is coming, for instance, from a study of
Tipper, Paul, and Hayes (2006), which recently demonstrated that
action affordance effects on grasping were larger when partici-
pants were presented with an active action state of the object, such
as a door handle depressed by 45%, than when they were presented
with the same object in a passive state (i.e., horizontal). According
to Tipper et al., this benefit of the active state suggests that the
activation of affordance from object perception is context depen-
dent and might be mediated by mental simulations of another
person’s action with the object. Moreover, Bub and Masson (2006)
have recently demonstrated that object affordance effects emerge
only if the observer attends to the object. Passive viewing without
the intention to act does not evoke hand gesture knowledge. Taken
together, recent observations have argued against the idea that
processing of action-related object features obligatorily activates
consistent action plans. That is, even though action-relevant infor-
mation is probably automatically extracted and processed and
object perception is not, this knowledge does not obligatorily
affect processes of action planning and execution. Rather, the
behavioral impact of perceived object affordance seems to depend
heavily on the action context in which the object is presented as
well as on the concurrent motor intentions of the observer.
Another approach to the coupling of perception and action is
provided by theories of ideomotor action, which basically hold that
movements are exhaustively coded in terms of their sensory con-
sequences (e.g., Greenwald, 1970), and by theories of common
coding, which assume that representations of perception and action
are based on the same cognitive codes and thus operate on the
same representational domain (Hommel, Mu¨sseler, Aschersleben,
& Prinz, 2001; Prinz, 1990). The central assumption shared by
both views is that motor actions and perceptual effects are highly
interrelated and mutually dependent. Experimental evidence for
this idea comes from studies on imitation showing that the per-
ception of a hand or finger movement facilitates the execution of
consistent motor actions (Bertenthal, Longo, & Kosobud, 2006;
Brass, Bekkering, & Prinz, 2001; Brass, Bekkering, Wohlschla¨ger,
& Prinz, 2000; Stu¨rmer, Aschersleben, & Prinz, 2000). For exam-
ple, Stu¨rmer et al. (2000) required participants to open or close
their hand in response to the color of a picture of an opening or
closing hand. Although the depicted action was irrelevant to the
assigned task, participants responded faster when they had to
execute the same action as that in the picture. Likewise, Brass et al.
(2001) demonstrated that people are faster in initiating a finger
movement to an arbitrary cue when an irrelevant but response-
compatible visual finger movement is shown simultaneously. To-
gether, the reports of automatic imitation effects show that per-
ceived sensory action consequences automatically activate the
representations of associated actions.
In contrast to the direct perception view, the common coding
approach predicts perceptual effects on action planning and as-
sumes a bidirectional connection between perception and ac-
tion—a notion that also implies the possibility of action-induced
effects on perception (Hommel et al., 2001). Because visual ob-
jects and motor actions are assumed to be represented by shared
features codes, it is expected that action planning affects percep-
tual processing by biasing the cognitive system toward feature
dimensions that are relevant for the preparation of the intended
response (Hommel et al., 2001). For instance, the intention to
grasp an object should prepare the visual system for the processing
of grasp-related object features. Such enhanced activation of codes
of features defined on motor-relevant dimensions can be under-
stood as a sort of intentional weighting process (see, e.g., Hommel,
in press). In fact, several studies have reported evidence for action-
induced effects and demonstrated that the preparation of a motor
response affects visual processing of objects and events that is
consistent with the currently intended action (Craighero et al.,
1999; Fagioli, Hommel, & Schubotz, 2007; Hamilton, Wolpert, &
Frith, 2004; Lindemann & Bekkering, 2009; Schubo¨, Prinz, &
Aschersleben, 2004; Wohlschla¨ger, 2000; Zwickel, Grosjean,
& Prinz, 2007). As one example, Craighero et al. (1999) demon-
strated that the processing of a visual stimulus is facilitated if it
affords the same type of grasping response as the participant
concurrently intends to perform. In their paradigm, participants
were instructed to prepare to grasp differently oriented wooden
bars but to delay the response execution until a visual go-signal
was presented. They observed faster detections of go-signals that
afforded the same type of grip as that involved in the prepared
action, indicating that the preparation of an object-directed motor
response facilitates the visual processing of action-consistent stim-
uli and showing that the intention to grasp an object is sufficient to
constitute an action context that modulates visual perceptual pro-
cessing.
Interestingly, as we know from recent research on the represen-
tation of functional object knowledge, the mere observation of a
grasping action or hand posture strongly influences semantic judg-
ments of graspable objects (Paulus, Lindemann, & Bekkering, in
press; Vainio et al., 2008; Yoon & Humphreys, 2005). Yoon and
Humphreys (2005), for instance, presented pictures of tools to-
gether with hands that were holding the objects in different ways
and showed that this contextual action information strongly influ-
enced the time taken to identify how the object is typically used.
Taking into account the view of the bidirectional perception–
action coupling and the finding that task-irrelevant action infor-
mation modulates the semantic processing of familiar objects, it
might be speculated that action-induced effects on the perceptual
processing of objects and their affordances might be influenced by
contextual action information associated with the grasp and type of
use. However, so far we do not know much about the role of
others’ actions on the representation of affordances and a possible
interplay between processes of action observation and object per-
ception.
The major aim of this research was therefore to study object
affordance effects in natural reach-to-grasp actions while focusing
on the role of the action context in the processing of object
affordances. In four experiments, we required participants to judge
the semantic category (i.e., natural or manmade) of presented
objects. In contrast to previous studies on stimulus–response com-
patibility effects in object perception (e.g., Tucker & Ellis, 2001),
we required participants to indicate their decisions by performing
different types of reach-to-grasp movements. The aim of the first
experiment was to determine whether the perception of visual
object properties interferes with the planning and execution of
natural grasping movements by investigating the effects of object
affordances on the reach onset times and movement kinematics of
grasping actions. In subsequent experiments, we applied this par-
adigm to examine effects of action context on object perception. In
particular, we focused on the role of the perception of another’s
action by examining the potential influence of action observation
331
ACTION CONTEXT AND OBJECT PERCEPTION
on the presence of the object affordance effect, that is, the com-
patibility effect between presented object and executed grasping
response.
Experiment 1
In Experiment 1, we examined the presence of object affordance
effects in natural grasping actions and moreover tested whether the
processing of action-related object features has an impact on both
components of a reach-to-grasp action, that is, on the reaching
(movement initiation times) and on the grasping (maximum grip
aperture [MGA]). To this end, we required participants to judge
the semantic category (natural or manmade) of a visually presented
object (fruit or tool) and to reach out for an object (manipulandum)
placed in front of them. More important, the decisions had to be
indicated by grasping the manipulandum with either a power grip
or a precision grip. We expected the time to initiate reaching to be
shorter when it was cued by an object affording the same type of
grip rather than a different type of grip, that is, an object affor-
dance effect (Tucker & Ellis, 2001). Taking into account previous
research showing an impact of semantic magnitude information on
grasping kinematics (Glover et al., 2004; Lindemann, Abolafia,
Girardi, & Bekkering, 2007; Lindemann, Stenneken, van Schie, &
Bekkering, 2006), we also expected the perceived object size to
affect the aperture of the hand during the reaching phase, as
revealed by an enlarged MGA for pictures of large as compared
with small objects.
Method
Participants. Twenty-one students of Radboud University
Nijmegen took part in the experiment in return for 4.50 ($6.39) or
course credit. All were right handed, had normal or corrected-to-
normal vision, and were naive with respect to the purpose of the
study.
Setup. In a dimly lit room, participants sat in front of a table
and were required to reach out for a wooden object (i.e., manipu-
landum). The manipulandum consisted of two parts: a large cyl-
inder (diameter 6 cm, height 7 cm) at the bottom and a small
cylinder (diameter 0.7 cm, height 1.5 cm) attached to top of
the large cylinder. It could be grasped in one of two ways: either
with a power grip at the large cylinder or with a precision grip at
the small cylinder (see Figure 1B). The manipulandum was placed
on the right side of the table behind an opaque screen (height 44
cm, width 45 cm), allowing participants to reach it comfortably
with their right hand without visual control (see Figure 1A). At a
distance of 30 cm from the center of the object, we fixated a small
pin (height 0.5 cm, diameter 0.5 cm) that served as a marker
for the starting position of the reach-to-grasp movements.
Stimuli. All stimuli were displayed on a gray background
using a 17-in. (43.2-cm) monitor (refresh rate 100 Hz). Each
target stimulus consisted of a color photograph of a small or large
manipulable object. At a viewing distance of 50 cm, horizontal and
vertical visual angle ranged from about 3° (small objects such as a
paperclip) to 30° (a large object such as a saw). We used 20
manmade objects and 20 natural objects (see Appendix A for a list
of all objects). Half of the objects were small and consequently
afforded a precision grip action (e.g., a sharpener or a grape), and
the other half were large and required a power grip action (e.g., a
hammer or a banana).
1
Each object was depicted in two different
horizontal orientations. One orientation required a right-hand grasp
(e.g., handle on the right side), and the other required a left-hand
grasp (e.g., handle on the left side). Pictures of objects with
opposite alignments were obtained by mirroring the photographs.
Procedure. Before the experiment started, participants per-
formed a short preexperimental block in which they were required
to grasp the manipulandum with either a precision or a power grip,
depending on which colored dot was presented on the screen.
Specifically, they were required to reach out for the manipulandum
and to grasp it either at its large bottom part, using all of the fingers
on one hand (i.e., power grip response), or at its small top part,
using only their thumb and index finger (i.e., precision grip re-
sponse). The actual experiment started only when participants
were able to perform the grasping movements correctly and flu-
ently without vision.
At the beginning of each experimental trial, a gray fixation cross
was presented at the center of the screen; it indicated that partic-
ipants should place their index finger in the starting position. As
soon as their hand was placed correctly, the fixation cross turned
black and disappeared 1,500 ms later. After a random interval of
500 –2,000 ms, the target stimulus (i.e., tool or fruit) was pre-
sented. The participants’ task was to judge the semantic category
of the depicted object and to indicate their decision by performing
one of the two practiced actions (i.e., precision or power grip).
Responses had to be made as quickly and accurately as possible.
The target stimulus remained visible until the onset of reaching or
until 3,000 ms had elapsed. After performing the reach-to-grasp
movement, participants were required to grasp the object until the
gray fixation cross appeared to indicate the start of the next trial.
A stop sign, together with a beep sound (4400 Hz, lasting 200 ms),
1
Large and small objects were chosen considering the kind of grip
required to pick them up. We presented 56 pictures of manmade and
natural objects to 18 participants and asked them to indicate whether the
object required a precision or power grip to handle it appropriately. The 40
objects selected for the study were classified to a rate of 100% of the
responses as either a precision or a power grip object.
Precision Grip
Power Grip
B
A
Figure 1. Basic experimental setup. A: Participants sat at a table with a
computer screen and a manipulandum. An opaque screen blocked the view
of the manipulandum and the right hand. B: The manipulandum consisted
of two segments: a large cylinder at the bottom, affording a power grip, and
a small cylinder at the top, affording a precision grip.
332 GIRARDI, LINDEMANN, AND BEKKERING
was presented as error feedback if reach-to-grasp movements were
initiated before onset of the target stimulus.
Design. The mapping between the semantic object category
and the required response was counterbalanced between partici-
pants; that is, half of the participants performed a power grip action
in response to natural objects and a precision grip action in
response to manmade objects. For the other half, the stimulus–
response mapping was reversed. The experimental block consisted
of 160 trials (20 objects 2 semantic categories 2 object
orientations 2 repetitions). All trials were presented in a ran-
domized order. In addition, at the beginning of the experiment we
presented 20 practice trials consisting of four sample objects
2
that
were not used in the experimental trials. The experiment lasted
about 45 min.
Data acquisition. To record the hand movements, we used an
electromagnetic position tracking system (miniBIRD 800TM, As-
cension Technology Corporation, Burlington, VT). Three sensors
were attached to the participants’ thumb, index finger, and right
wrist. Sensor positions were tracked at a sampling rate of 103 Hz.
The movement kinematics were analyzed offline. A fourth-order
Butterworth low-pass filter with a cutoff frequency of 10 Hz was
applied to the raw data. The onset of a reach-to-grasp movement
was defined as the first moment in time at which the tangential
velocity of the index finger sensor exceeded the threshold of 10
cm/s. For the offset, we used the opposite criterion, taking the time
of the first sample in which the velocity decreased below this
threshold.
As dependent variables for the statistical tests, we calculated the
mean response latencies of the reaching movements (response time
[RT], defined as the mean time elapsed between onset of picture
presentation and onset of reach-to-grasp movement), and the MGA
(defined as the maximum distance between the thumb and the
index finger during reaching) of the grasping. Anticipation re-
sponses (response before onset of go-picture presentation and
RTs 100 ms), missing responses (no reactions and RTs 1,000
ms), and incorrectly performed actions (e.g., incorrect grasping,
wrong type of grip, movement stopped while reaching) were
considered to be errors and excluded from further analyses. A
Type I error rate alpha of .05 was used in all statistical tests
reported here.
Results
The rate of anticipations (1%) and errors (1.8%) was low,
showing that participants had carefully executed the object cate-
gorization task.
Reach onset latencies. We subjected the mean RTs to a
repeated measures analysis of variance (ANOVA) with object size
(large or small) and motor response (power grip or precision grip)
as within-subjects factors. The analysis revealed a significant main
effect of object size, F(1, 20) 39.53, p.001, indicating that
large objects were identified faster (469 ms) than small objects
(493 ms). There was a tendency for power grip responses to be
initiated faster (477 ms) than precision responses (485 ms), F(1,
20) 3.37, p.08. The two-way interaction between object size
and motor response was significant, F(1, 20) 6.76, p.02 (see
Figure 2). As revealed by post hoc t-test comparisons, when large
objects were presented, participants initiated power grip responses
(459 ms) faster than precision grip responses (479 ms), t(1, 19)
2.84, p.001; when small objects were presented, RTs for the
precision grip responses (491 ms) were descriptively shorter than
for the power grip responses (496 ms). However, this contrast
failed to reach significance (t1).
Grasping kinematics. We conducted a repeated measures
ANOVA with the factors object size (large or small) and object
category (natural or manmade) on the mean MGAs. The analysis
showed a main effect of object size, F(1, 20) 4.58, p.05. That
is, grip apertures were larger when grasping the manipulandum
while viewing large objects (109.7 mm) than while viewing small
objects (108.5 mm). Neither the main effect of object category nor
the interaction between the two factors reached significance (both
Fs1).
Discussion
Experiment 1 demonstrates that reach-to-grasp movements are
affected by visual processing of objects affording different types of
actions. The object compatibility effect on the reach onset latencies
reflects that participants initiated power grip actions faster in
response to large objects (e.g., a hammer) than in response to small
objects (e.g., a sharpener). This finding provides the first direct
empirical support for the notion of object affordance effects in
complex natural grasping actions and thus extends previous find-
ings of object affordance effects on the execution of finger move-
ments that are involved in power or precision grip actions (Tucker
& Ellis, 2001). Moreover, we observed that the MGA was larger
while viewing a large object than while viewing a small object.
The object size effect in the grasping kinematics demonstrates an
impact of task-irrelevant magnitude information on motor behav-
ior, as has previously been reported for word-reading and number-
processing tasks (Glover et al., 2004; Lindemann et al., 2007), and
shows that similar effects also emerge during processing of visual
2
A cherry tomato, a tangerine, a carving fork, and a nail were used as
sample objects for the training trials.
Precision Grip Power Grip
430
440
450
460
470
480
490
500
510
520
530
Reaction Times (ms)
Small Objects
Large Objects
Figure 2. Mean response latencies of the grasping movements in Exper-
iment 1 as a function of the factors motor response and object size.
333
ACTION CONTEXT AND OBJECT PERCEPTION
information. The object affordance effect in the reach-onset laten-
cies was significant only for large objects. This dissociation be-
tween small and large objects was possibly driven by the fact that
small objects were more difficult for the participants to discrimi-
nate, and as a result, they were not strongly associated with a
particular motor representation. The idea of object affordances and
the observation of response compatibility effects in object percep-
tion has led many researchers to conclude that the processing and
representation of action-relevant visual information in object per-
ception takes place in a rather automatic way (see, e.g., Tucker &
Ellis, 2001). The finding of interference effects between object
perception and motor actions, however, does not inevitably imply
that the detection of action-relevant object features automatically
activates motor representations suited for manipulation of objects
(Phillips & Ward, 2002). Alternatively, it might be possible that
affordances are automatically processed but do not obligatorily
activate motor actions.
As mentioned in the introduction, ample evidence has shown
that the intention to perform a motor movement facilitates the
processing of action-related perceptual features such as object size
(Fagioli, Hommel, & Schubotz, 2007) or orientation (Craighero et
al., 1999) or the detection of action-consistent events (Lindemann
& Bekkering, 2009). It has therefore been argued that the activa-
tion of action representations has a direct impact on subsequent
attentional and perceptual processes by facilitating the processing
of action-relevant features and dimensions. Along the lines of this
intentional weighting hypothesis (see Hommel, in press), one
might also speculate that the mere observation of another person’s
object grasping modulates the processing of action-relevant object
features. If so, one might expect that object affordance effects
depend on concurrently activated representations of others’ motor
behavior and contextual information about others’ action inten-
tions. A possible means of testing this hypothesis is to manipulate
the action context, that is, the scenario in which the object is
perceived, and to investigate whether the presence of another
person’s grasping action affects the perceptual processing of
action-related object features and, thus, the activation of afforded
actions. We therefore conducted another experiment, the aim of
which was to examine the influence of contextual information on
the presence of object affordance effects.
Experiment 2
In Experiment 2, we investigated whether the action context in
which an object is perceived affects the presence of object affor-
dance effects on grasping responses. As in Experiment 1, partici-
pants were instructed to reach out and grasp the manipulandum
differently depending on the semantic decision task regarding the
nature of visually presented large and small objects. To investigate
the influence of observing another’s action on processing of object
affordances, we presented the objects in the context of different
grasping actions by also showing pictures of hands approaching
the object with either a power or a precision grip. Because the
depicted objects were either small or large, the grip size of pre-
sented hand postures was thus either appropriate or inappropriate
with respect to the action afforded by the object. Because of the
manipulation of observed hand– object relations, it was possible to
test whether object affordance effects are modulated by action
context.
As we know from previous research on ideomotor compatibility
effects (Brass et al., 2000, 2001; Stu¨rmer et al., 2000), the obser-
vation of hand postures or hand movements has a direct impact on
the motor system and facilitates the execution of congruent motor
actions. If object affordances also automatically potentiate com-
ponents of actions, the object compatibility effect on grasping
actions should not be mediated by contextual information. In this
case, we expect to find two independent but not interfering effects,
that is, an effect of the object’s affordance and an ideomotor effect
of the hand posture. If the processing of object affordances, how-
ever, is modulated by the action context, we expect that the effects
of object affordance on motor responses will be modulated by the
depicted hand posture and its relation to the required response.
Method
Participants. Twenty-four students of Radboud University
Nijmegen took part in the experiment. All were right handed, had
normal or corrected-to-normal vision, and were naive with respect
to the purpose of the study. They received 4.50 ($6.39) or course
credit for their participation.
Setup stimuli and data acquisition. The experimental setup
and the data acquisition were identical to those of Experiment 1. A
new set of color photographs of 20 small or large tools (i.e.,
manmade objects) and 20 small or large fruits (i.e., natural objects)
were used as target stimuli (see Appendix B for a list of all
objects). In contrast to previous studies, each object was always
presented along with a photograph of the left or right hand (see
Figure 3A). Each object subtended a visual angle of between 3°
(small objects such as a paperclip) and 25° (large objects such as
a teapot). The pictures of the hands (visual angle 18°) were
presented randomly to the left or the right of the object. More
important, all stimuli were assembled in such a way that the
depicted hand appeared to approach the object. That is, right hands
were always shown to the left of the object and left hands to the
right. The left-hand picture was obtained by mirroring the photo-
graph of the right hand. For the no-go trials, the picture of the hand
was tinted blue.
Procedure. The procedure was basically identical to that of
Experiment 1. Participants indicated the semantic category of the
presented object (i.e., manmade or natural) by means of different
grasping responses (i.e., precision or power grip). To ensure that
participants paid attention to both object and hand, we introduced
additional no-go trials in which the hand was tinted blue. In these
trials, participants had to refrain from responding irrespective of
the object category.
Design. The mapping between the semantic object category
and the required response was again counterbalanced between
participants. The experimental block consisted of 160 randomized
experimental trials (20 objects 2 semantic categories 2
grasping postures 2 object orientations) and 32 no-go trials.
Again, motor responses were trained in a short preexperimental
block. The experiment lasted about 45 min.
Because of the factorial combinations of the required grasping
responses and the two orthogonal stimulus features of object size
and depicted hand posture, each trial was compatible or incom-
patible with respect to object size and posture. That is, depending
334 GIRARDI, LINDEMANN, AND BEKKERING
on whether the grip was the same or different from that of the
hand in the photograph, a response could be considered as
posture compatible or incompatible. Moreover, each response
was, depending on whether the grip size matched the object
size, either object compatible or incompatible, as was the case
in Experiment 1.
Results
The low percentage of anticipations (1%) and incorrectly
performed actions (3.0%) indicated that participants had per-
formed the task carefully.
We subjected the mean RTs to an ANOVA with the within-
subjects factors motor response (power grip or precision grip),
object compatibility (compatible or incompatible), and posture
compatibility (compatible or incompatible). The analysis revealed
a effect for the factor motor response, F(1, 23) 6.98, p.01,
showing that power grip responses (572 ms) were initiated faster
than precision grip responses (596 ms). Also, the main effect of
object compatibility reached significance, F(1, 23) 9.39, p
.01, indicating that grasping actions compatible with the action
afforded by the object (578 ms) were initiated faster than incom-
patible grasping actions (589 ms). There was no effect for the
factor posture compatibility, F(1, 23) 1.6. More important,
however, there was a significant interaction between the factors
object compatibility and posture compatibility, F(1, 23) 13.6,
p.001 (see Figure 4). No other interaction of the ANOVA
reached significance (all Fs1).
As post hoc t-test comparisons revealed, responses were faster
toward compatible objects only if the depicted posture was com-
patible with the required response (572 ms vs. 593 ms), t(23)
4.32, p.001. For posture-incompatible trials, however, there
was no effect of object compatibility (585 ms vs. 586 ms; |t|1).
Discussion
Experiment 2’s results show that object affordance effects were
present only if the depicted hand posture was compatible with the
required grasping action. That is, the execution of grasping was
facilitated only if both stimulus features (i.e., object size and hand
posture) were compatible with the response. However, in condi-
tions in which one or both stimulus features were incompatible,
RTs did not differ from one another. This finding clearly excludes
the possibility of two independent effects of object size and hand
posture; rather, it suggests that the effect of object affordance on
action execution depends on the concurrent action intention and
the relation between depicted hand posture and required motor
action. Experiment 2 shows consequently that ideomotor compat-
ibility effects between perceived and performed actions (Brass et
al., 2000) modulate the processing of action-related object features
and its impact on the motor system.
However, we cannot exclude at this point that the interaction
between object compatibility and posture compatibility originates
from interference at the perceptual stage of stimulus identification
and is not driven, as interpreted earlier, by stimulus–response
Figure 3. Examples of stimuli in Experiments 2, 3 (A), and 4 (B) for the two types of stimulus–response
compatibility— object and hand posture—in the case of a required power grip action. The side of hand
presentation (left or right) was counterbalanced. Note that because pointing actions were unrelated to hands
and objects, responses in Experiment 3 could not be classified as object or posture compatible (see text for
details).
Object Compatible Object Incompatible
560
570
580
590
600
610
620
630
640
650
660
Reaction Times (ms)
Posture Compatible
Posture Incompatible
Figure 4. Mean response latencies of the grasping movements in Exper-
iment 2 as a function of the factors object compatibility and posture
compatibility.
335
ACTION CONTEXT AND OBJECT PERCEPTION
relations. It might be possible that effects of object perception on
the grasping actions emerge only if the object is part of a mean-
ingful grasping action scenario in which the depicted hand posture
is shaped appropriately to grasp an object (i.e., a small object next
to a precision handgrip or a large object next to a power handgrip).
The observed interaction would in this case merely be the result of
facilitated processing of objects presented next to appropriately
shaped hands, which would reflect an effect of visual familiarity
based on people’s experiences in action observation. To control for
this alternative explanation, we conducted another experiment.
Experiment 3
In Experiment 3, we aimed to examine the origin of the inter-
action between posture and object compatibility in greater detail.
We tested in particular whether the interference effect observed in
Experiment 2 reflected an overlap between action stimulus and
response features (ideomotor stimulus–response compatibility) or
whether it was driven by an intrastimulus consistency of two
stimulus features, that is, the fit of the depicted grasping hand and
the object size (stimulus–stimulus congruency effect; see Korn-
blum et al., 1990). A straightforward way to disentangle these two
explanations is to minimize the ideomotor compatibility between
observed and executed actions. To do so, we modified the required
action responses and instructed participants to perform pointing
movements, which are, in contrast to the grasping actions in the
previous experiment, unrelated to the depicted hand postures. If
the observed interference effect in Experiment 2 reflects a facili-
tated perceptual processing of consistent action scenarios in which
the grip size of the presented hand posture fits the object size, RT
effects should be independent of the required type of motor re-
sponse and thus also be present for pointing movements. On the
contrary, no difference should be expected in pointing latencies if
the observed effects depend on the compatibility between the
selected responses and the depicted action scenarios.
Method
Participants. Twenty-four right-handed students of Radboud
University Nijmegen took part in the experiment in return for
4.50 ($6.39) or course credit. All had normal or corrected-to-
normal vision and were naive with respect to the purpose of the
study.
Setup, stimuli, and procedure. The experimental setup, stim-
uli, and procedure were almost identical to those of Experiment 2.
The only modification was that participants had to indicate their
judgments by pointing movements. That is, depending on the
semantic category of the depicted object, participants were re-
quired to point either to the small top or to the large bottom part of
the manipulandum. The actions were executed in a training phase
until participants attained sufficient expertise. Half of the partici-
pants pointed to the small top cylinder when viewing a natural
object and to the large bottom cylinder when viewing a manmade
object. The response mapping was reversed for the other partici-
pants.
Data acquisition and design. The data acquisition and exper-
imental design were identical to those of Experiment 3. Because
pointing actions were not related to the hands and objects, re-
sponses could not be classified as object or posture compatibility.
We therefore analyzed the latencies with respect to possible intra-
stimulus effects of the congruency between the depicted hand
posture and the size of the object (i.e., stimulus congruency).
Results
The anticipation rate was less than 1%; incorrectly performed
actions occurred in only 2.2% of the trials. A repeated measures
two-way ANOVA on the mean RTs revealed no main effect for the
factors stimulus congruency (congruent or incongruent), F(1,
23) 2.03, p.1, and motor response (pointing to the bottom or
pointing to the top), F(1, 23) 3.07, p.05. Also, the interaction
between the two factors failed to reach significance (F1). This
finding suggests that semantic judgments and motor responses
were not facilitated for stimuli in which the grip size of the
depicted hand posture fit the object (667 ms) as compared with
incongruent stimuli (674 ms).
To compare the congruency effects for grasping (Experiment 2)
and pointing actions (Experiment 3) directly, we performed a
mixed ANOVA with the between-subjects factor experiment and
the within-subjects factor stimulus congruency. As the significant
interaction between the two factors confirmed, F(1, 46) 3.90,
p.05, stimulus consistency effects were present only in Exper-
iment 2, t(23) 5.05, p.001, and not in Experiment 3, t(23)
1.42, p.15. Furthermore, RTs in Experiment 3 were on average
slower (671 ms) than RTs in Experiment 2 (585 ms), F(1, 46)
10.80, p.01. It might consequently be possible that stimulus
congruency effects emerge only if RT intervals are relatively short
and vanishing congruency effects were thus the result of an overall
RT difference. To exclude this possibility, we performed a RT
distribution analysis and investigated the time course of a possible
stimulus congruency effect in Experiment 3 (see Ratcliff, 1979).
RTs for motor response and stimulus congruency conditions of
each participant were rank ordered, divided into quintiles, aver-
aged, and submitted to a repeated measures ANOVA. The inter-
action between the factors stimulus congruency and quintiles
(F1) failed to reach significance, indicating an absence of
stimulus congruency effects for the slow and the fast responses.
We can therefore exclude that the vanishing of the congruency
effect was the result of the overall RT difference between Exper-
iments 2 and 3.
Discussion
When participants performed pointing movements, we observed
no congruency effect between depicted object and depicted hand
postures. This finding argues against the possibility that the out-
come of Experiment 2 was driven by stimulus–stimulus congru-
ency effect, that is, a facilitated perceptual processing of congruent
action scenarios in which the grip size of the presented hand
posture fitted the size of the object. We can therefore exclude that
the modulation of object affordance effects by observed hand
postures had a mere perceptual origin.
Our experiments demonstrate that the compatibility between
observed and performed grasping actions modulates the perception
and the cognitive effects of the object affordances. Interestingly,
Experiment 2 revealed no main effect for the factor posture com-
patibility, that is, faster initiations of grasping responses consistent
336 GIRARDI, LINDEMANN, AND BEKKERING
with the depicted hand postures. At first glance, this finding might
be unexpected because several previous studies have reported the
presence of ideomotor compatibility effects for motor response
that are consistent with observed hand movements (Brass et al.,
2000; Stu¨rmer et al., 2000; but see also Vainio, Tucker, & Ellis,
2007). In contrast to these studies, the hands in Experiment 2’s
pictures were never presented alone without an object. Neverthe-
less, we can exclude that they were not processed because hand
posture compatibility modulated the presence of object affordance
effects. The lack of an ideomotor compatibility effect can be
interpreted as an indication that participants processed the hand
and the object as one integrated action scenario and not as two
separate stimuli. Alternatively, the dominance of object affordance
effects might be because the semantic categorization task required
participants to focus their attention on the object and less on the
depicted hand. This raises the question of whether both hand
posture compatibility and object affordance effects emerge simul-
taneously while viewing an object and a hand that are not per-
ceived as two related entries. In Experiment 4, therefore, we tested
whether two independent stimulus–response compatibility effects
for object and hand can be observed if they are presented as
spatially separated visual stimuli that are not part of one action
context.
Experiment 4
Experiment 4 investigated the presence of stimulus–response
compatibility effects of the depicted object and hand posture under
conditions in which they are not elements of one integrated object-
directed action. The same pictures were presented as in the previ-
ous experiments. However, objects and hand postures were now
presented at two separate locations on the screen so that they did
not appear to be two related elements in one grasping action
scenario. We expected that if objects and hands are perceived as
two separate stimuli, RTs would reflect two independent effects,
that is, an effect of the object affordance and an effect of ideomotor
compatibility.
Method
Participants. Twenty-four right-handed students of Radboud
University Nijmegen took part in the experiment in return for
4.50 ($6.39) or course credit. All had normal or corrected-to-
normal vision and were naive with respect to the purpose of the
study. None of the students had participated in the other experi-
ments.
Setup, stimuli, and procedure. The experimental setup, stim-
uli, and procedure were basically the same as in Experiment 2. The
only modification was the arrangement of the hand and object
stimuli on the screen. Instead of presenting the two pictures in
spatial proximity to one another, object and hand posture were
presented in the left and right visual field (4° eccentricity), respec-
tively, so that the hand did not appear to approach the object (see
Figure 3B). As in Experiments 2 and 3, the pictures of right hands
were presented to the left of the object and left hands were
presented to the right.
Data acquisition and design. The data acquisition and exper-
imental design were identical to those of Experiment 2.
Results
The rate of anticipations was less than 1%. The error rate was
5.2%. The three-way repeated measures ANOVA (Motor Re-
sponse Object Compatibility Posture Compatibility) of the
mean RTs revealed significant effects for both compatibility fac-
tors. Object-compatible responses (627 ms) were initiated faster
than object-incompatible responses (648 ms), F(1, 23) 23.85,
p.001, and responses compatible with the depicted posture (633
ms) were initiated faster than the posture-incompatible responses
(641 ms), F(1, 23) 3.95, p.05 (see Figure 5). Interestingly,
the interaction of the two compatibility factors did not reach
significance, F(1, 23) 1.2, indicating that the effects of object
and of posture compatibility occurred independently of one an-
other and did not interfere.
Again, participants showed a tendency to initiate power grip
responses (629 ms) faster than precision grip responses (645 ms),
F(1, 23) 3.83, p.07. The two-way interactions between motor
response and object compatibility and between motor response and
posture compatibility were both significant, F(1, 23) 9.39, p
.005, and F(1, 23) 11.46, p.003, respectively. These inter-
action effects indicate that both object and posture compatibility
effects were larger for power grip responses (34 ms and 19 ms,
respectively) than for precision grip responses (8 ms and 3 ms,
respectively), ts(23) 3.32 and 2.94, ps.05.
To test whether the influence of the depicted hand posture on the
object compatibility effect was different in Experiment 4 than in
Experiment 2, in which object and hand posture were presented
close together, we performed a between-experiments comparison.
To this aim, we calculated the size of the object compatibility
effects, defined as the RT difference between object-incompatible
and object-compatible trials, for each participant in both experi-
ments. We entered the resulting RT effects into a mixed design
ANOVA with the between-subjects factor experiment (Experiment
2 or Experiment 4) and the within-subject factor posture compat-
ibility. Besides the main effect of posture compatibility, F(1, 46)
Object Compatible Object Incompatible
560
570
580
590
600
610
620
630
640
650
660
Reaction Times (ms)
Posture Compatible
Posture Incompatible
Figure 5. Mean response latencies of the grasping movements in Exper-
iment 4 as a function of the factors object compatibility and posture
compatibility.
337
ACTION CONTEXT AND OBJECT PERCEPTION
11.49, p.001, the analysis yielded a significant interaction
between the two factors, F(1, 46) 4.34, p.05. Post hoc ttests
showed that the depicted hand postures had an impact on the
presence of the object compatibility effect only in Experiment 2
(object compatibility effect for hand-compatible trials, 21.4 ms; for
hand-incompatible trials, 0.31 ms), t(23) 3.68, p.001, but not
in Experiment 4 (18.3 ms vs. 23.4 ms), t(23) 1.
Discussion
Experiment 4 demonstrates the presence of two independent
stimulus–response compatibility effects, object affordance and
hand posture, if both were presented simultaneously but as two
separate stimuli. That is, grasping actions were initiated faster in
response to pictures of objects affording a compatible grip and in
response to pictures of compatible hand postures. More important,
the results revealed that the stimulus–response compatibility effect
of object and posture did not interact if the pictures were not
arranged so that the hand appeared to approach the object. The lack
of cross-talk under these circumstances shows that the relation
between the hand posture and the object modulates the presence of
object compatibility effects. As the comparison of the object
compatibility effects in Experiments 2 and 4 confirmed, the per-
ceptual processing of the object and hand posture interacted only
if they were perceived as two integrated parts of one action
scenario. This outcome points to the important role of the action
context in object perception and the processing of object affor-
dances.
General Discussion
Four experiments explored the cognitive interference between
perception and action and investigated, in particular, the role of
contextual information on the behavioral effects of perceived ob-
ject affordances. As the analyses of stimulus–response compati-
bility effects revealed, perceived affordances of an object had an
influence on the planning and execution of natural reach-to-grasp
actions. This observation is in line with several previous studies
showing that the perception of an object activates representations
of motor actions that are best suited for a manipulation of the
object (Glover et al., 2004; Phillips & Ward, 2002; Tucker & Ellis,
2001). More important, this study aimed to investigate the influ-
ence of the action context, that is, the influence of observation of
others’ actions on the processing of action-relevant object features.
It thus extends the literature on object perception and action
affordances in at least two aspects.
First, we demonstrated that the object affordance effects on
motor response are modulated by the action context in which the
object is perceived. We only found stimulus–response compatibil-
ity effects between object features and required motor responses
when the object was presented close to the hand grasping posture
so that they formed a meaningful grasping action (Experiment 2).
We did not, however find effects of the object affordance if an
object-inconsistent hand posture was presented. This interaction
shows that object affordance effects strongly depend on the action
context, that is, the observed action that is performed with the
object, suggesting that object-related actions are not automatically
activated in the observer. Interestingly, the compatibility effects
could not be observed when participants executed finger-pointing
movements (Experiment 3). We can consequently reject the alter-
native account of the outcome of Experiment 2 as a stimulus–
stimulus congruency effect and exclude the possibility that object
affordance effects were driven by a facilitated perceptual process-
ing of congruent action scenarios in which the grip size of the
presented posture fitted the size of the object. We therefore inter-
pret the dependence of the object affordance effect on the depicted
hand posture as evidence for an influence of contextual informa-
tion on object processing. This outcome is in line with other
studies showing that the associated functional object knowledge
becomes activated only if people intend to use the object with that
specific purpose (Bub & Masson, 2006; Lindemann et al., 2006) or
if the object is perceived in an active action context (Tipper et al.,
2006). We thence conclude that the activation of action represen-
tations and motor codes is not completely automatic and obligatory
(see, e.g., Tucker & Ellis, 2001).
Second, this study is among the first to demonstrate the
presence of object affordance effects in reaching and grasping
components of natural object-directed grasping actions. So far,
object affordance effects have mostly been reported only for
button-press responses or simple finger movements that are
involved in grasping actions (e.g., Derbyshire et al., 2006; Ellis
& Tucker, 2000; Tucker & Ellis, 2001). The major advantage to
using natural object-directed grasping action is the possibility
of dissociating between effects in reaching and grasping com-
ponents of motor response. Interestingly, object affordance
effects were evident in both the onset of the reaching move-
ments and the movement kinematics of the grasping. The effect
in the movement latencies shows, in line with motor theories
(Rosenbaum, Meulenbroek, Vaughan, & Jansen, 2001), that the
intended grip size at the end of the movement is anticipated and
planned before the hand has reached the target object. This
interference between object perception and action intention
clearly demonstrates that object affordance effects emerge at
the level of motor planning. Also, the finding of an enlarged
MGA indicates an interaction between object perception and
action planning. The effect on the grasping kinematics, more-
over, parallels research demonstrating effects of size-related
semantic information on grasping actions (Gentilucci, Benuzzi,
Bertolani, Daprati, & Gangitano, 2000; Glover et al., 2004;
Lindemann et al., 2007). Glover et al. (2004) have shown that
the modulation of the grip aperture by magnitude information
reflects interference at the level of action planning. The kine-
matics effect in this study consistently provided additional
support for our notion that the perception of action-relevant
object features interferes only with action planning but not with
motor control.
As mentioned before, the ideomotor theory suggests that
action representations are activated when observing visual
events that correspond to effects of own and others’ motor
actions (Hommel et al., 2001). With this study, we provide new
support for this view by demonstrating that ideomotor compat-
ibility effects are not restricted to simple finger movements and
also emerge in connection with complex action scenarios and
goal-directed reach-to-grasp movements. When the object and
the hand grasping posture were presented far away from one
another (Experiment 4), we found two effects. Observing a
grasping hand facilitates the execution of similar grasping ac-
tions (i.e., ideomotor compatibility effect; Brass et al., 2000;
338 GIRARDI, LINDEMANN, AND BEKKERING
Stu¨rmer et al., 2000), whereas observing a manipulable object
primes the execution of the grasping action associated with it
(cf. Tucker & Ellis, 2001). Interestingly, the behavioral effects
of object perception and action observation have mostly been
investigated in isolation. As this study now shows, object
affordance and ideomotor compatibility effects can arise under
certain conditions independently and do not necessarily inter-
fere. This independence indicates that the mere perception of
others’ actions is not sufficient to modulate the processing of
action-relevant object features and suggests that object and
action observation interact only if both aspects are conceived as
part of one meaningful action scenario (Experiment 2).
Taken together, these data argue against the view that visual
objects automatically and obligatorily potentiate components of
action they afford and rather provide evidence for the notion
that the processing of action-relevant object features depends
greatly on the contextual correspondence of perception and
action. The finding that object affordance effects depend on the
action context in which the object is perceived supports the idea
of an intentional weighting of action-relevant dimensions in
perceptual processing.
References
Bertenthal, B. I., Longo, M. R., & Kosobud, A. (2006). Imitative response
tendencies following observation of intransitive actions. Journal of
Experimental Psychology: Human Perception and Performance, 32,
210 –225.
Brass, M., Bekkering, H., & Prinz, W. (2001). Movement observation
affects movement execution in a simple response task. Acta Psycho-
logica, 106, 3–22.
Brass, M., Bekkering, H., Wohlschla¨ger, A., & Prinz, W. (2000). Com-
patibility between observed and executed finger movements: Comparing
symbolic, spatial and imitative cues. Brain Cognition, 44, 124 –143.
Bub, D. N., & Masson, M. E. J. (2006). Gestural knowledge evoked by
objects as part of conceptual representations. Aphasiology, 20, 1112–
1124.
Craighero, L., Fadiga, L., Rizzolatti, G., & Umilta`, C. (1999). Action for
perception: A motor-visual attentional effect. Journal of Experimental
Psychology: Human Perception and Performance, 25, 1673–1692.
Derbyshire, N., Ellis, R., & Tucker, M. (2006). The potentiation of two
components of the reach-to-grasp action during object categorization in
visual memory. Acta Psychologica, 122, 74 –98.
Ellis, R., & Tucker, M. (2000). Micro-affordance: The potentiation of
components of action by seen objects. British Journal of Psychology, 9,
451– 471.
Ellis, R., Tucker, M., Symes, E., & Vainio, L. (2007). Does selecting one
visual object from several require inhibition of the actions associated
with nonselected objects? Journal of Experimental Psychology: Human
Perception and Performance, 33, 670 – 691.
Fagioli, S., Ferlazzo, F., & Hommel, B. (2007). Controlling attention
through action: Observing actions primes action-related stimulus dimen-
sions. Neuropsychologia, 45, 3351–3355.
Fagioli, S., Hommel, B., & Schubotz, R. I. (2007). Intentional control of
attention: Action planning primes action-related stimulus dimensions.
Psychological Research, 71, 22–29.
Gentilucci, M., Benuzzi, F., Bertolani, L., Daprati, E., & Gangitano, M.
(2000). Language and motor control. Experimental Brain Research, 13,
468 – 490.
Gibson, J. J. (1979). The ecological approach to visual perception. Boston:
Houghton-Mifflin.
Glover, S. (2004). Separate visual representations in the planning and
control of actions. Behavioral and Brain Sciences, 27, 3–24.
Glover, S., Rosenbaum, D. A., Graham, J., & Dixon, P. (2004). Grasping
the meaning of words. Experimental Brain Research, 154, 103–108.
Greenwald, A. G. (1970). Sensory feedback mechanisms in performance
control: With special reference to the ideomotor mechanism. Psycho-
logical Review, 77, 73–99.
Hamilton, A., Wolpert, D. M., & Frith, U. (2004). Your own action
influences how you perceive another person’s action. Current Biology,
14, 493– 498.
Hommel, B. (1995). Stimulus–response compatibility and the Simon ef-
fect: Toward an empirical clarification. Journal of Experimental Psy-
chology: Human Perception and Performance, 21, 764 –775.
Hommel, B. (in press). Grounding attention in action control: The inten-
tional control of selection. In B. J. Bruya (Ed.), Effortless attention: A
new perspective in the cognitive science of attention and action. Cam-
bridge, MA: MIT Press.
Hommel, B., Mu¨sseler, J., Aschersleben, G., & Prinz, W. (2001). The
theory of event coding (TEC): A framework for perception and action
planning. Behavioral and Brain Sciences, 24, 849 –937.
Kornblum, S., Hasbroucq, T., & Osman, A. (1990). Dimensional overlap:
Cognitive basis for stimulus–response compatibility: A model and tax-
onomy. Psychological Review, 97, 253–270.
Lindemann, O., Abolafia, J. M., Girardi, G., & Bekkering, H. (2007).
Getting a grip on numbers: Numerical magnitude priming in object
grasping. Journal of Experimental Psychology: Human Perception and
Performance, 33, 1400 –1409.
Lindemann, O., & Bekkering, H. (2009). Object manipulation and motion
perception: Evidence of an influence of action planning on visual pro-
cessing. Journal of Experimental Psychology: Human Perception and
Performance, 35, 1062–1071.
Lindemann, O., Stenneken, P., van Schie, H. T., & Bekkering, H. (2006).
Semantic activation in action planning. Journal of Experimental Psy-
chology: Human Perception and Performance, 32, 633– 643.
Mu¨sseler, J., & Hommel, B. (1997). Blindness to response-compatible
stimuli. Journal of Experimental Psychology: Human Perception and
Performance, 23, 861– 872.
Paulus, M., Lindemann, O., & Bekkering, H. (in press). Motor simulation
in verbal knowledge acquisition. Quarterly Journal of Experimental
Psychology.
Phillips, J. C., & Ward, R. (2002). S-R compatibility effects of irrelevant
visual affordance: Timecourse and specificity of response activation.
Visual Cognition, 9, 540 –558.
Prinz, W. (1990). A common coding approach to perception and action. In
O. Neumann & W. Prinz (Eds.), Relationships between perception and
action: Current approaches (pp. 167–201). Berlin: Springer.
Ratcliff, R. (1979). Group reaction time distributions and an analysis of
distribution statistics. Psychological Bulletin, 86, 446 – 461.
Rosenbaum, D. A., Meulenbroek, R. J., Vaughan, J., & Jansen, C. (2001).
Posture-based motion planning: Applications to grasping. Psychological
Review, 108, 709 –734.
Schubo¨, A., Prinz, W., & Aschersleben, G. (2004). Perceiving while acting:
Action affects perception. Psychological Research, 68, 208 –215.
Simon, J. R. (1969). Reactions toward the source of stimulation. Journal of
Experimental Psychology, 81, 174 –176.
Stu¨ rmer, B., Aschersleben, G., & Prinz, W. (2000). Correspondence effects with
manual gestures and postures: A study on imitation. Journal of Experimental
Psychology: Human Perception and Performance, 26, 1746 –1759.
Tipper, S., Paul, M. A., & Hayes, A. E. (2006). Vision-for-action: The
effects of object property discrimination and action state on affordance
compatibility effects. Psychonomic Bulletin & Review, 13, 493– 498.
Tucker, M., & Ellis, R. (2001). The potentiation of grasp types during
visual object categorization. Visual Cognition, 8, 769 – 800.
Vainio, L., Symes, E., Ellis, R., Tucker, M., & Ottoboni, G. (2008). On
the relations between action planning, object identification, and mo-
339
ACTION CONTEXT AND OBJECT PERCEPTION
tor representations of observed actions and objects. Cognition, 108,
444 – 465.
Vainio, L., Tucker, M., & Ellis, R. (2007). Precision and power grip
priming by observed grasping. Brain and Cognition, 65, 195–207.
Wohlschla¨ger, A. (2000). Visual motion priming by invisible actions.
Vision Research, 40, 925–930.
Yoon, E. Y., & Humphreys, G. W. (2005). Direct and indirect effects
of action on object classification. Memory & Cognition, 33, 1131–
1146.
Zwickel, J., Grosjean, M., & Prinz, W. (2007). Seeing while moving:
Measuring the online influence of action on perception. Quarterly Jour-
nal of Experimental Psychology, 60, 1063–1071.
Appendix A
Natural and Manmade Objects Presented in
Experiment 1
Small objects affording a precision grip action: almond, brussels
sprout, cranberry, garlic, grape, mushroom, nut, pepper, radish,
string bean, clothespin, key, lighter, paperclip, pen, pencil, sharp-
ener, screw, teaspoon, tweezers.
Large objects affording a power grip action: avocado, banana, carrot,
cucumber, eggplant, leek, mango, pear, paprika, potato, brush, corkscrew,
cup, hairbrush, hairdryer, hammer, knife, rake, saw, screwdriver.
Appendix B
Natural and Manmade Objects Presented in
Experiments 2, 3, and 4
Small objects affording a precision grip action: almond,
brussels sprout, cranberry, garlic, grape, mushroom, nut,
pepper, radish, string bean, chess piece, clothespin, dart,
hairspray, needle, paperclip, pushpin, sharpener, teaspoon,
tweezers.
Large objects affording a power grip action: avocado, banana,
carrot, cucumber, eggplant, leek, mango, pear, paprika, potato,
book, bottle, cup, iron, joystick, key, soft drink can, tea box,
teapot, thermos.
Received May 9, 2008
Revision received June 16, 2009
Accepted June 19, 2009
E-Mail Notification of Your Latest Issue Online!
Would you like to know when the next issue of your favorite APA journal will be available
online? This service is now available to you. Sign up at http://notify.apa.org/ and you will be
notified by e-mail when issues of interest to you become available!
340 GIRARDI, LINDEMANN, AND BEKKERING
... In the field of affordance perception, there has been debate regarding whether it is an automated process or not. Some researchers argue that affordance perception is an automatic process because of its fast and effortless nature (Bonner & Epstein, 2017;Goslin et al., 2012;Harel et al., 2022;Tucker & Ellis, 1998), whereas others suggest that it is not automated but rather highly contextualised and can be influenced by biases and expectations (Girardi et al., 2010;Kalénine et al., 2016;Mustile et al., 2021;Pellicano et al., 2010;Tipper et al., 2006;Wokke et al., 2016). However, a synthesis perspective proposes that affordance automaticity should be understood as a dynamic process that changes over time, whereby affordance perception may initially occur automatically but is later modulated by higher-level cognitive processes (Borghi & Riggio, 2015;Djebbara et al., 2022;Gastelum, 2020;Kourtis et al., 2018). ...
... Other studies have challenged the view that affordances are automatically activated, based on findings that perception of object affordances is modulated by task and context (Mazzuca et al., 2021). Context, in this case, refers to the scenario in which the object is perceived, including the presence of other objects or situations in which objects are embedded (Girardi et al., 2010;Mazzuca et al., 2021;Mustile et al., 2021). For instance, studies by Pellicano et al. (2010) and Tipper et al. (2006) demonstrated that the affordance effects to grasp objects as proposed by Tucker and Ellis (1998) only emerge when the task requires deeper processing of the object's characteristics, such as object categorisation and shape recognition, but not in simpler perceptual tasks like colour discrimination (Pellicano et al., 2010;Tipper et al., 2006). ...
... For instance, studies by Pellicano et al. (2010) and Tipper et al. (2006) demonstrated that the affordance effects to grasp objects as proposed by Tucker and Ellis (1998) only emerge when the task requires deeper processing of the object's characteristics, such as object categorisation and shape recognition, but not in simpler perceptual tasks like colour discrimination (Pellicano et al., 2010;Tipper et al., 2006). In a behavioural study, Girardi et al. (2010) aimed to study object affordance effects in natural reach-to-grasp actions while focussing on the role of the action context in the processing of object affordances. They presented the participants with a priming picture consisting of a hand and an object, and required them to make a power or precision grasp to classify the object as manmade or animate. ...
Article
Full-text available
Affordances, the opportunities for action offered by the environment to an agent, are vital for meaningful behaviour and exist in every interaction with the environment. There is an ongoing debate in the field about whether the perception of affordances is an automated process. Some studies suggest that affordance perception is an automated process that is independent from the visual context and bodily interaction with the environment, whereas others argue that it is modulated by the visual and motor context in which affordances are perceived. The present paper aims to resolve this debate by examining affordance automaticity from the perspective of sensorimotor time windows. To investigate the impact of different forms of bodily interactions with an environment, that is, the movement context (physical vs. joystick movement), we replicated a previous study on affordance perception in which participants actively moved through differently wide doors in an immersive 3D virtual environment. In the present study, we displayed the same environment on a 2D screen with participants moving through doors of different widths using the keys on a standard keyboard. We compared components of the event‐related potential (ERP) from the continuously recorded electroencephalogram (EEG) that were previously reported to be related to affordance perception of architectural transitions (passable and impassable doors). Comparing early sensory and later motor‐related ERPs, our study replicated ERPs reflecting early affordance perception but found differences in later motor‐related components. These results indicate a shift from automated perception of affordances during early sensorimotor time windows to movement context dependence of affordance perception at later stages, suggesting that affordance perception is a dynamic and flexible process that changes over sensorimotor stages.
... In the field of affordance perception, there has been debate regarding whether it is an automated process or not. Some researchers argue that affordance perception is an automatic process due to its fast and effortless nature (Tucker & Ellis, 1998;Goslin, Dixon, Fischer, Cangelosi, & Ellis, 2012;Bonner & Epstein, 2017;Harel, Nador, Bonner, & Epstein, 2022), whereas others suggest that it is not automated but rather highly contextualized and can be influenced by biases and expectations (Tipper, Paul, & Hayes, 2006;Girardi, Lindemann, & Bekkering, 2010;Pellicano, Iani, Borghi, Rubichi, & Nicoletti, 2010;Kalénine, Wamain, Decroix, & Coello, 2016;Wokke, Knot, Fouad, & Ridderinkhof, 2016;Mustile, Giocondo, Caligiore, Borghi, & Kourtis, 2021). ...
... Context, in this case, refers to the scenario in which the object is perceived, including the presence of other objects or situations in which objects are embedded (Girardi, Lindemann, & Bekkering, 2010;Mazzuca et al., 2021;Mustile, Giocondo, Caligiore, Borghi, & Kourtis, 2021). For instance, studies by Pellicano et al. (2010) and Tipper et al. (2006) demonstrated that the affordance effects to grasp objects as proposed by Tucker and Ellis (1998) only emerge when the task requires deeper processing of the object's characteristics, such as object categorization and shape recognition, but not in simpler perceptual tasks like color discrimination (Pellicano, Iani, Borghi, Rubichi, & Nicoletti, 2010;Tipper, Paul, & Hayes, 2006). ...
... For instance, studies by Pellicano et al. (2010) and Tipper et al. (2006) demonstrated that the affordance effects to grasp objects as proposed by Tucker and Ellis (1998) only emerge when the task requires deeper processing of the object's characteristics, such as object categorization and shape recognition, but not in simpler perceptual tasks like color discrimination (Pellicano, Iani, Borghi, Rubichi, & Nicoletti, 2010;Tipper, Paul, & Hayes, 2006). In a behavioral study, Girardi et al. (2010) aimed to study object affordance effects in natural reach-to-grasp actions while focusing on the role of the action context in the processing of object affordances. They presented participants with a priming picture consisting of a hand and an object, and required them to make a power or precision grasp to classify the object as manmade or animate. ...
Preprint
Full-text available
Affordances, the opportunity for action offered by the environment to an agent, are vital for meaningful behavior and exist in every interaction with the environment. Regarding its temporal mechanism, some studies suggest that affordance perception is an automated process that is independent from the visual context and bodily interaction with the environment, while others argue that it is modulated by the visual and motor context in which affordances are perceived. We aims to resolve this debate by examining affordance automaticity from the perspective of sensorimotor time windows. We replicated a previous study on affordance perception in which participants actively moved through doors of different width in VR environments. To investigate the impact of different forms of bodily interactions with an environment, i.e., the movement context (physical vs. joystick movement), we used the identical virtual environment from Djebbara and colleagues (2019) but displayed it on a 2D screen with participants moving through different wide doors using the laptop keyboard. We compared components of the event-related potential (ERP) from the continuously recorded electroencephalogram (EEG) that were previously reported to be related to affordance perception of architectural transitions (passable and impassable doors). Comparing early sensory and later motor-related ERPs, our study replicated ERPs reflecting early affordance perception but found differences in later motor-related components. It indicates a shift from automated perception of affordances during early sensorimotor time windows to movement context dependence of affordance perception at later stages suggesting that affordance perception is a dynamic and flexible process that changes over sensorimotor stages.
... In the field of affordance perception, there has been debate regarding whether it is an automated process or not. Some researchers argue that affordance perception is an automatic process due to its fast and effortless nature (Tucker & Ellis, 1998;Goslin, Dixon, Fischer, Cangelosi, & Ellis, 2012;Bonner & Epstein, 2017;Harel, Nador, Bonner, & Epstein, 2022), whereas others suggest that it is not automated but rather highly contextualized and can be influenced by biases and expectations (Tipper, Paul, & Hayes, 2006;Girardi, Lindemann, & Bekkering, 2010;Pellicano, Iani, Borghi, Rubichi, & Nicoletti, 2010;Kalénine, Wamain, Decroix, & Coello, 2016;Wokke, Knot, Fouad, & Ridderinkhof, 2016;Mustile, Giocondo, Caligiore, Borghi, & Kourtis, 2021). ...
... Context, in this case, refers to the scenario in which the object is perceived, including the presence of other objects or situations in which objects are embedded (Girardi, Lindemann, & Bekkering, 2010;Mazzuca et al., 2021;Mustile, Giocondo, Caligiore, Borghi, & Kourtis, 2021). For instance, studies by Pellicano et al. (2010) and Tipper et al. (2006) demonstrated that the affordance effects to grasp objects as proposed by Tucker and Ellis (1998) only emerge when the task requires deeper processing of the object's characteristics, such as object categorization and shape recognition, but not in simpler perceptual tasks like color discrimination (Pellicano, Iani, Borghi, Rubichi, & Nicoletti, 2010;Tipper, Paul, (which was not certified by peer review) is the author/funder. ...
... ;https://doi.org/10.1101/2023.09.07.556516 doi: bioRxiv preprint & Hayes, 2006. In a behavioral study, Girardi et al. (2010) aimed to study object affordance effects in natural reach-to-grasp actions while focusing on the role of the action context in the processing of object affordances. They presented participants with a priming picture consisting of a hand and an object, and required them to make a power or precision grasp to classify the object as manmade or animate. ...
Preprint
Full-text available
Affordances, the opportunity for action offered by the environment to an agent, are vital for meaningful behavior and exist in every interaction with the environment. There is an ongoing debate in the field about whether the perception of affordances is an automated process. Some studies suggest that affordance perception is an automated process that is independent from the visual context and bodily interaction with the environment, while others argue that it is modulated by the visual and motor context in which affordances are perceived. The present paper aims to resolve this debate by examining affordance automaticity from the perspective of sensorimotor time windows. We replicated a previous study on affordance perception in which participants actively moved through doors of different width in immersive 3D virtual environments. To investigate the impact of different forms of bodily interactions with an environment, i.e., the movement context (physical vs. joystick movement), we used the identical virtual environment from Djebbara and colleagues (2019) but displayed it on a 2D screen with participants moving through different wide doors using the keys on a standard keyboard. We compared components of the event-related potential (ERP) from the continuously recorded electroencephalogram (EEG) that were previously reported to be related to affordance perception of architectural transitions (passable and impassable doors). Comparing early sensory and later motor-related ERPs, our study replicated ERPs reflecting early affordance perception but found differences in later motor-related components. These results indicate a shift from automated perception of affordances during early sensorimotor time windows to movement context dependence of affordance perception at later stages suggesting that affordance perception is a dynamic and flexible process that changes over sensorimotor stages.
... Such a size coding might be related to the size of the grasped response device. Indeed, in previous studies, participants had to respond either by pressing a small or a large switch (e.g., Ellis and Tucker 2000;Ellis 2001, 2004) or by grasping the small or large part of a single response device (Girardi et al. 2010). In addition, the Theory of Event Coding (TEC; Hommel et al. 2001; Hommel 2019 see also Ansorge and Whür, 2004) highlighted another critical requirement for the spatial coding of responses in a two-alternatives-forced-choice task: alternative responses are coded as a function of distinctive features. ...
... Second, it should encourage researchers to better theorize the kind of motor representations really simulated when objects are merely seen (e.g., the usual or more frequent or other kinds) and their influence at a motor level (i.e., what is the exact range of actions that can be potentiated by the motor representation of an usual action). Third, our contribution should also encourage to systematically use a control condition in which participants have to perform pointing movements directed toward the large vs. small component of the response device (e.g., Girardi et al. 2010) to decide between an embodied and a size-coding account. ...
Article
Full-text available
People are faster to perform a precision grip when they see a cherry (i.e., a small graspable object) than to perform a power grip, and the reverse holds true when they see an apple (i.e., a large graspable object). This potentiation effect supports that object representations could include motor components that would be simulated when a graspable object is seen. However, the nature of these motor components remains unclear. The embodied account posits that seeing an object only potentiates the most frequent actions associated with it (i.e., usual actions). In contrast, the size-coding account posits that seeing an object potentiates any actions associated to spatial codes compatible with those associated to the objects. We conducted three experiments to disentangle these two alternative accounts. We especially varied the nature of the responses while participants saw either large or small graspable objects. Our results showed a potentiation effect when participants performed the usual grasping actions (Experiment 1: power and precision grip) but also when they performed unusual grasping actions (Experi-ment 2: grasping between the thumb and little finger) and even when they had to perform non-grasping actions (Experiment 3: pointing actions). By supporting the size-coding account, our contribution underlines the need for a better understanding of the nature of the motor components of object representations and for using a proper control condition (i.e., pointing action) before arguing that the embodied account convincingly explains the potentiation effect of grasping behaviors.
... Prior to undertaking the affordances task, participants viewed four brief video clips, each lasting two seconds, in which a male/female hand (consistent with the participant's gender) reached and grasped a teapot or a cup by the handle, each turning left in one clip and right in a second clip (for similar procedures see: Garrido-Vásquez & Schubö, 2014;Littman & Kalanthroff, 2021Tipper et al., 2006). This was done in line with previous suggestions according to which affordances tendencies may become more prominent under conditions that emphasize the object's graspability or the contextual correspondence of perception and action (Girardi et al., 2010;Lu & Cheng, 2013;Netelenbos & Gonzalez, 2015). Next, participants completed a practice block and then performed an experimental block consisting of 288 trials. ...
... Importantly, a wide body of brain imaging studies has demonstrated the activation of premotor areas when participants view manipulable objects (Chao & Martin, 2000;Creem-Regehr & Lee, 2005;Grafton et al., 1997;Grezes & Decety, 2002;Proverbio et al., 2011), an activation which is absent in classic Simon tasks (e.g., Kerns, 2006), and unique patterns of brain activity for manipulable objects that go beyond the effects of spatial correspondence (Buccino et al., 2009;Rice et al., 2007). Nonetheless, the findings of recent studies refined the initial concept of complete automaticity of the affordances effect and suggested that the affordances effect becomes more behaviorally evident when objects are presented in their functional orientation (Bub et al., 2018;Masson et al., 2011), and under conditions which emphasize the object's graspability (Girardi et al., 2010;Lu & Cheng, 2013). To ascertain the emergence of an affordances effect, we followed the specific suggestions made by these studies. ...
Article
The affordances task serves as an important tool for the assessment of cognition and visuomotor functioning, and yet its test–retest reliability has not been established. In the affordances task, participants attend to a goal-directed task (e.g., classifying manipulable objects such as cups and pots) while suppressing their stimulus-driven, irrelevant reactions afforded by these objects (e.g., grasping their handles). This results in cognitive conflicts manifesting at the task level and the response level. In the current study, we assessed the reliability of the affordances task for the first time. While doing so, we referred to the “reliability paradox,” according to which behavioral tasks that produce highly replicable group-level effects often yield low test–retest reliability due to the inadequacy of traditional correlation methods in capturing individual differences between participants. Alongside the simple test–retest correlations, we employed a Bayesian generative model that was recently demonstrated to result in a more precise estimation of test–retest reliability. Two hundred and ninety-five participants completed an online version of the affordances task twice, with a one-week gap. Performance on the online version replicated results obtained under in-lab administrations of the task. While the simple correlation method resulted in weak test–retest measures of the different effects, the generative model yielded a good reliability assessment. The current results support the utility of the affordances task as a reliable behavioral tool for the assessment of group-level and individual differences in cognitive and visuomotor functioning. The results further support the employment of generative modeling in the study of individual differences.
... Colour judgements are strongly action-irrelevant and recent action priming results suggest that although fruit and vegetables are manipulable objects, visual presentation of natural manipulable objects may not evoke grasps as much as manufactured manipulable objects (Godard et al., 2019). Moreover, several studies indicate that reach-and-grasp responses are more inclined (if not necessary) to elicit the evocation of action components from visual objects (e.g., Bub et al., 2018;Girardi et al., 2010). Interestingly, Matheson & Thompson-Schill, (2019) also stressed the importance of the task rather than prior motor experience with objects in the emergence of compatibility effects, at least with object position. ...
... Three RTs were considered: the latency from stimulus presentation to finger release from starting position, i.e., initiation RTs, the latency from release to device grasp, i.e., movement RTs, and the sum of initiation and movement RTs, i.e., total RTs. Several authors showed evocation of action components by analysing initiation and movement times (Bub et al., 2018;Girardi et al., 2010). ...
Article
When the size of visual objects matches the size of the response required to perform the task, a potentiation effect has been reported, with faster responses in compatible than incompatible situations. Size compatibility effects have been taken as evidence of close perception-action interrelations. However, it is still unclear whether the effect arises from abstract coding of the size of stimulus and response or from the evocation of grasp affordances from visual objects. We aimed to disentangle the two interpretations. Two groups of 40 young adults categorised small and large objects presented in standardised size as natural or artefacts objects. One group categorised manipulable objects that may be associated with small or large size properties and evoke power and precision grasp affordances. The other groups categorised non-manipulable objects that may only be associated with small or large size properties. Categorisation responses were made by reaching and grasping a monotonic cylindric device with a power and precision grip in a grasping condition and with large or small touch responses in a control condition. Compatibility effects were found in both grasping and control conditions, independently of the manipulability or category of objects. Participants were faster when the size of the expected response matched the size of the object than when they mismatched, especially for power grasp or whole-hand touch responses. Overall findings support the abstract coding hypothesis and suggest that compatibility between the conceptual size of the object and the size of the hand response is sufficient to facilitate semantic categorization judgements.
... On the behavioral side, an important line of evidence comes from studies showing stimulus-response compatibility effects. Numerous studies (Ellis & Tucker, 2000;Girardi et al., 2010;Makris et al., 2014;Tucker & Ellis, 2001) have reported facilitating effects of executing a congruent grasping gesture when participants perform semantic categorization tasks on objects (e.g., natural vs. artifact). Participants classically categorize small objects (e.g., cherry) faster when they respond by executing a precision grasping gesture rather than a power grasping gesture, and vice versa for large objects (e.g., bottle). ...
Article
The present study addressed the role of motor representations in declarative memory (i.e., semantic and episodic). Based on embodied and grounded theories of cognition, it is often suggested that motor representations contribute to declarative memory. According to the action priming effect, graspable objects are categorized faster when primed by pictures of a congruent hand grip, as motor representations (how to grasp it) and semantic information (what it is) are closely related. Moreover, motor representations may contribute to episodic memory functioning. We immobilized participants’ dominant hand for 24 hours to impair their processing of hand-related motor representations. This method is known to elicit rapid updating of cortical hand representations, and a slowdown in cognitive tasks linked to hand-related motor cognition. We expected to observe a decreased action priming effect following short-term hand nonuse. We further predicted that in a subsequent recognition task, objects that had been encoded following congruent action priming would be recognized faster by controls, but not by previously immobilized participants. Results did not show any effect of hand nonuse on action priming, suggesting that motor representations are not a decisive factor for this effect. Nonetheless, prime congruence influenced subsequent recognition. Immobilized participants were slower to recognize objects previously seen with an unrelated hand grip prime compared to a congruent one. This result suggests a contribution of motor representation to declarative memory, in particular when the sensorimotor system has previously been impaired.
... It has been demonstrated that facilitation in compatible conditions between stimulus and response may at least partially arise from a match between abstract codes associated with the target (e.g., "large" object) and the response (e.g., "large" grasp; Azaad et al., 2019;Cho & Proctor, 2010;Heurley et al., 2020;Proctor et al., 1990;Proctor & Miles, 2014). Yet studies have evidenced that in some specific action-relevant situations when task demands and response conditions are sufficiently relevant for action, the affordance of the objects can be evoked (Bub et al., 2018;Girardi et al., 2010). Importantly, the effect of affordance evocation and abstract coding may coexist in the stimulus-response compatibility paradigm. ...
Article
Full-text available
Object perception and action are closely interrelated: Various grasping components are evoked when perceiving visual objects (“object affordances”). Yet little is known about the impact of the evocation of multiobject affordances on object perceptual processing. This study aimed to determine whether object processing may be affected by the similarity of affordances evoked by multiple objects and whether semantic relations between objects modulate this effect. Adult students were presented with three-dimensional scenes involving pairs of graspable objects. Each object evoked grasp size affordances (precision or power grasps). Affordances of the two objects could be similar or dissimilar and objects could be thematically related (spatula–pan) or unrelated (spatula–snow globe). Participants had to judge the color of a target object by performing power and precision grasps compatible or incompatible with the target evoked grasp. Results showed slower responses on compatible targets when unrelated distractors evoked similar compared to dissimilar affordances. This cost of similar affordances disappeared when objects were thematically related. Findings corroborate predictions of recent models hypothesizing automatic inhibition of distractor affordances when selecting one object among others. We further provide novel evidence for a role of thematic relations between objects in the perception of multiple affordances. Findings have implications for object processing in naturalistic scenes.
... The SDA-M approach has been applied in several different studies, such as manual actions [35], rehabilitation settings [36], cognitive robotics [37], and complex actions in sports contexts [38,39], to measure the structure and dimension of mental representation in motor actions. However, only a few studies examined the sensorimotor aspects of perceptual actions to understand how sensorimotor representations relate to grasping and how grasping movements are affected by object features [40]. So far, we do not know much about the role of grasping actions on the sensorimotor representation and the possible interplay between object features and goal-directed grasping actions. ...
Article
Full-text available
This study investigated the structure of sensorimotor representations during goal-directed grasping actions and explored their relationship with object features. Sixteen 3D-printed spheres that varied in size (i.e., a diameter of 20 mm, 40 mm, 60 mm, 80 mm) and weight (i.e., 40 g, 52 g, 76 g, 91 g) were used as experimental stimuli. The Structural Dimensional Analysis of Mental Representation (SDA-M) method was used to assess the sensorimotor representation structure during grasping. Participants were instructed in each trial to weigh, lift, or transport sets of two different spheres and to judge the similarity of the objects’ features, taking into account the executed grasping movement. Each participant performed a total of 240 trials, and object presentation was randomized. The results suggest that the functional interplay between object features and goal-directed actions accounts for the significant variations in the structure of sensorimotor representations after grasping. Specifically, the relevance of the perceived objects’ size and weight is closely interrelated to the grasping task demands and movement dynamics of the executed action. Our results suggest that distinct sensorimotor representations support individual grasping actions according to top-down influences modulated by motor intentions, functional task demands, and task-relevant object features.
... Participants were tested individually. In line with previous suggestions that affordances tendencies may become more prominent under conditions which emphasize the object's graspability or the contextual correspondence of perception and action (Girardi et al., 2010;Lu & Cheng, 2013;Netelenbos & Gonzalez, 2015), we primed participants' reaching and grasping tendencies by using the following procedure: before performing the affordances task, participants were presented with an actual cup and an actual pot, each placed once with its handle facing right and once facing left. In each trial, the participants were asked to demonstrate grasping the object by its handle with their congruent hand. ...
Article
Task conflict emerges when a stimulus triggers two or more competing tasks. To date, task conflict has been studied mainly using the color-word Stroop task. We hypothesized that task conflict also emerges in the affordances task between the goal-directed relevant task (e.g., classifying manipulable objects such as cups and pots), and the automatic, stimulus-driven, irrelevant task afforded by these objects (e.g., grasping their handles). Thus, we expected task conflict to manifest in both congruent and incongruent trials, separately from the well-known affordances response conflict that manifests in incongruent trials between responding with the right vs. the left hand. To this end, we aimed to identify a neutral condition for the affordances task. In Experiment 1, participants performed an affordances task that included images of manipulable objects and houses. While manipulable objects evoke automatic grasping tendencies, house images were hypothesized to serve as neutral, conflict-free stimuli. House images yielded shorter reaction time (RT) than incongruent trials, indicating that they may serve as neutral stimuli for the task. House images also yielded shorter RT than congruent trials, suggesting that task conflict manifests in congruent (as well as in incongruent) affordances trials. In Experiment 2 we manipulated cognitive control in the affordance task by creating low-control and high-control blocks. While both congruent and incongruent trials were impacted by this manipulation of cognitive control, neutral trials remained unaffected. These findings indicate that the affordances task involves conflicts at both the task level and the level of response, and can be used as a supplementary, non-linguistic measure of task conflict and the activation of task control.
Chapter
Full-text available
The classic problem of stimulus-response (S-R) compatibility (SRC) is addressed. A cognitive model is proposed that views the stimulus and response sets in S-R ensembles as categories with dimensions that may or may not overlap. If they do overlap, the task may be compatible or incompatible , depending on the assigned S-R mapping. If they do not overlap, the task is noncompatible regardless of the assigned mapping. The overlapping dimensions may be relevant or not. The model provides a systematic account of SRC effects, a taxonomy of simple performance tasks that were hitherto thought to be unrelated, and suggestive parallels between these tasks and the experimental paradigms that have traditionally been used to study attentional, controlled, and automatic processes. In this article, we address the classic problem of stimulus-response (S-R) compatibility (SRC). A model is proposed that attempts to provide a systematic account of performance in highly compatible, incompatible, and noncompatible tasks. At the core of our model is the idea that when a particular S-R ensemble produces either high or low compatibility effects, it is because the stimulus and response sets in the ensemble have properties in common, and elements in the stimulus set automatically activate corresponding elements in the response set. Noncompatible tasks are those in which the stimulus and response sets have nothing in common. If the activated response is the required one, it will be executed rapidly and correctly; if it is not, then it will be relatively slow and error prone. Whether a particular S-R ensemble will produce compatibility effects is often quite easy to determine because of the relationship between the stimulus and response sets. In the part of the model that treats the representational aspects of the problem, we postulate that this relationship is based on the commonality, simi-or GBDV @umichum.bitnet. larity, or correspondence of the sets in the ensemble. We call this the dimensional overlap of the ensemble. The automatic response activation mechanisms, as well as the response identification processes underlying SRC effects, are the processing aspect of the model. Even though the model is still in its qualitative development phase, it is able to make ordinal predictions concerning several different SRC effects, such as the effects of mapping, irrelevant dimensions, and number of alternatives. These effects had been viewed as unrelated empirical phenomena , and most of the work on SRC has dealt with them as manifestations of unique, nongeneralizable properties of particular dimensions or specific tasks. This approach has led to different accounts being proposed for "spatial," "symbolic," "sensori-motor," and "semantic" tasks. Our model attempts to break with such past approaches in a fundamental way by proposing to account for most major SRC effects in terms of common basic cognitive mechanisms. This unitary approach leads to a taxonomy of SRC tasks that reveals striking similarities between them and suggestive parallels with the experimental paradigms that have traditionally been used to investigate atten-tional, controlled, and automatic processes. Background on SRC
Article
Full-text available
This article describes a model of motion planning instantiated for grasping. According to the model, one of the most important aspects of motion planning is establishing a constraint hierarchy - a set of prioritized requirements defining the task to be performed. For grasping, constraints include avoiding collisions with to-be-grasped objects and minimizing movement-related effort. These and other constraints are combined with instance retrieval (recall of stored postures) and instance generation (generation of new postures and movements to them) to simulate flexible prehension. Dynamic deadline setting is used to regulate termination of instance generation, and performance of more than one movement at a time with a single effector is used to permit obstacle avoidance. Old and new data are accounted for with the model.
Article
Full-text available
The close integration between visual and motor processes suggests that some visuomotor transformations may proceed automatically and to an extent that permits observable effects on subsequent actions. A series of experiments investigated the effects of visual objects on motor responses during a categorisation task. In Experiment 1 participants responded according to an object's natural or manufactured category. The responses consisted in uni-manual precision or power grasps that could be compatible or incompatible with the viewed object. The data indicate that object grasp compatibility significantly affected participant response times and that this did not depend upon the object being viewed within the reaching space. The time course of this effect was investigated in Experiments 2-4b by using a go-nogo paradigm with responses cued by tones and go-nogo trials cued by object category. The compatibility effect was not present under advance response cueing and rapidly diminished following object extinction. A final experiment established that the compatibility effect did not depend on a within-hand response choice, but was at least as great with bi-manual responses where a full power grasp could be used. Distributional analyses suggest that the effect is not subject to rapid decay but increases linearly with RT whilst the object remains visible. The data are consistent with the view that components of the actions an object affords are integral to its representation.
Article
Full-text available
Perceptual accounts attribute the Simon effect to the correlation between relevant stimulus feature and response location (e.g., T. Hasbroucq & Y. Guiard, 1991). This account is questioned, and it is demonstrated that a Simon effect can be obtained without a stimulus–response correlation (Experiment 1). Experiments 2 and 3 replicated this finding and showed that the relative size of stimuli and response labels and temporal overlap of warning signal and stimulus cannot account for why the effect was absent in Hasbroucq and Guiard's study. However, no Simon effect occurred in a close replication of Hasbroucq and Guiard's original experiment (Experiment 4). Participants' reports suggested that this was due to a special stimulus-coding strategy, and in fact, the effect reappeared with a slightly modified stimulus display (Experiment 5). These results provide strong evidence against a perceptual approach to the Simon effect and support response-related views instead.
Article
Full-text available
Background: Theories of embodied knowledge argue that the representation and recruitment of motor processes may be important for deriving the meaning of many linguistic and perceptual elements. Aims: We examined the conditions under which gestural knowledge associated with manipulable objects is evoked. Methods & Procedures: A priming paradigm was used in which an object was presented in advance of a photograph of a hand gesture that participants were to mimic. On related trials, the target gesture was the same as the gesture typically used to interact with the object prime. On unrelated trials, the target gesture was not related to the object. In another set of experiments, a Stroop-like paradigm was used in which participants learned to produce manual responses to colour cues. After training, coloured photographs of manipulable objects were presented. The colour-cued gesture was either one typically used with the object or was unrelated to it. Outcomes & Results: In the priming experiments, response latencies were shorter in the related condition, but only when participants also made an identification response to the object prime. In the Stroop experiments, interference effects indicated that gestures to colour were affected by gestural knowledge associated with the object. Conclusions: These results indicate that conceptual representations of manipulable objects include specific forms of gestural knowledge that are automatically evoked when observers attend to an object.
Article
Full-text available
Action affordances can be activated by non-target objects in the visual field as well as by word labels attached to target objects. These activations have been manifested in interference effects of distractors and words on actions. We examined whether affordances could be activated implicitly by words representing graspable objects that were either large (e.g., APPLE) or small (e.g., GRAPE) relative to the target. Subjects first read a word and then grasped a wooden block. Interference effects of the words arose in the early portions of the grasping movements. Specifically, early in the movement, reading a word representing a large object led to a larger grip aperture than reading a word representing a small object. This difference diminished as the hand approached the target, suggesting on-line correction of the semantic effect. The semantic effect and its on-line correction are discussed in the context of ecological theories of visual perception, the distinction between movement planning and control, and the proximity of language and motor planning systems in the human brain.
Article
Full-text available
It has been suggested that representations for action, elicited by an object's visual affordance, serve to potentiate motor components (such as a specific hand) to respond to the most afforded action. In three experiments, participants performed speeded left–right button press responses to an imperative target superimposed onto a prime image of an object suggesting a visual affordance oriented to left or right visual space. The time course of response activation was measured by varying the onset time between the prime and target. Long-lasting and gradually developing correspondence effects were found between the suggested affordance of the prime and the side of response, with little effect of response modality (hands uncrossed, hands crossed, or foot response). We conclude that visual affordances can evoke an abstract spatial response code, potentiating a wide variety of lateralized responses corresponding with the affordance. Co-ordinated and effective behaviour requires that we perceive and assess the possibilities for action afforded by objects in the environment. Visual affordances of objects support or invite certain types of actions: In this sense, the appearance of a handle affords its grasping. Traditionally, the notion of affordances for action has been linked with the theory of "direct" perception offered by Gibson (1979). Gibson argued that affordances for action were based on intrinsic perceptual properties of objects, registered directly and without the need for intervening processes such as object recognition. Here we use the term "affordance" more broadly, de-coupled from a theory of direct perception. For our purposes, an affordance may best be described purely as a Please address all correspondenc e to J.
Article
Full-text available
It is suggested that seen objects potentiate a range of actions associated with them, irrespective of the intentions of the viewer. Evidence for this possibility is provided by the data from two experiments, both of which required a participant to make a binary motor response to an auditory stimulus. In the first experiment the response was a power or precision grip, which was performed whilst simultaneously viewing a real object which would normally be grasped using either a power or precision grip. A significant interaction of response and grip compatibility of the object was observed. Similar results were obtained in the second experiment when a wrist rotation of a given direction was used as a response, whilst viewing objects which would require wrist rotations if they were to be grasped. The effects of the seen objects on components of action are described as microaffordances which are said to be dispositonal states of the viewer's nervous system.
Article
This chapter is concerned with some of the issues involved in understanding how perception contributes to the control of actions. Roughly speaking, the term of action refers to any meaningful segment of an organisms intercourse with its environment. Two important features of this preliminary definition can be brought out more clearly when “actions” are contrasted with “responses” and “movements”. Unlike response-centered approaches to psychology, which consider the organisms activity more or less determined by the actual stimulus information, the action approach emphasizes intentional control as being simultaneous with (or even prior to) informational control of activity, assuming that intentional processes fix the rules for the selection and use of stimulus information (Heuer Prinz, 1987; Neumann Prinz, 1987). Unlike movement-centered approaches, which describe the organisms activity in terms of the dynamics of muscular contraction patterns and the kinematics of the resulting body movements, the action approach stresses the environmental consequences that go along with these bodily events, contending that meaningful interactions with the environment, rather than movements per se, should be considered the effective functional units of activity (Fowler Turvey, 1982; Neisser, 1985).