ArticlePDF Available

Drawing from Memory: Hand-Eye Coordination at Multiple Scales

PLOS
PLOS One
Authors:

Abstract and Figures

Eyes move to gather visual information for the purpose of guiding behavior. This guidance takes the form of perceptual-motor interactions on short timescales for behaviors like locomotion and hand-eye coordination. More complex behaviors require perceptual-motor interactions on longer timescales mediated by memory, such as navigation, or designing and building artifacts. In the present study, the task of sketching images of natural scenes from memory was used to examine and compare perceptual-motor interactions on shorter and longer timescales. Eye and pen trajectories were found to be coordinated in time on shorter timescales during drawing, and also on longer timescales spanning study and drawing periods. The latter type of coordination was found by developing a purely spatial analysis that yielded measures of similarity between images, eye trajectories, and pen trajectories. These results challenge the notion that coordination only unfolds on short timescales. Rather, the task of drawing from memory evokes perceptual-motor encodings of visual images that preserve coarse-grained spatial information over relatively long timescales as well.
Content may be subject to copyright.
Drawing from Memory: Hand-Eye Coordination at
Multiple Scales
Stephanie Huette
1
*
.
, Christopher T. Kello
1.
, Theo Rhodes
2.
, Michael J. Spivey
1
1Cognitive and Information Sciences, University of California Merced, Merced, California, United States of America, 2Department of Psychology, State University of New
York at Oswego, Oswego, New York, United States of America
Abstract
Eyes move to gather visual information for the purpose of guiding behavior. This guidance takes the form of perceptual-
motor interactions on short timescales for behaviors like locomotion and hand-eye coordination. More complex behaviors
require perceptual-motor interactions on longer timescales mediated by memory, such as navigation, or designing and
building artifacts. In the present study, the task of sketching images of natural scenes from memory was used to examine
and compare perceptual-motor interactions on shorter and longer timescales. Eye and pen trajectories were found to be
coordinated in time on shorter timescales during drawing, and also on longer timescales spanning study and drawing
periods. The latter type of coordination was found by developing a purely spatial analysis that yielded measures of similarity
between images, eye trajectories, and pen trajectories. These results challenge the notion that coordination only unfolds on
short timescales. Rather, the task of drawing from memory evokes perceptual-motor encodings of visual images that
preserve coarse-grained spatial information over relatively long timescales as well.
Citation: Huette S, Kello CT, Rhodes T, Spivey MJ (2013) Drawing from Memory: Hand-Eye Coordination at Multiple Scales. PLoS ONE 8(3): e58464. doi:10.1371/
journal.pone.0058464
Editor: Joy J. Geng, University of California, Davis, United States of America
Received August 23, 2012; Accepted February 4, 2013; Published March 15, 2013
Copyright: ß2013 Huette et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits
unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Funding: This work was supported by NSF award 1031903. The funders had no role in study design, data collection and analysis, decision to publish, or
preparation of the manuscript.
Competing Interests: The authors have declared that no competing interests exist.
* E-mail: shuette@ucmerced.edu
.These authors contributed equally to this work.
Introduction
An organism’s perceptual and motor systems are coordinated
via reciprocal interactions that constitute perception-action loops
[1]. These loops are most salient at millisecond to second
timescales, as in perceptual-motor interactions involved in
locomotion [2], but they also span longer timescales in support
of more complex behaviors. An illustrative example can be found
in the dance of a honey bee–the bee finds pollen and later enacts
its location for the hive [3]. Perception-action loops on short
timescales support the flight of the bee to pollen, and memory is
used to encode and express successful flight paths at later points in
time. Thus memory is used extend the perception-action loop over
the entire period of foraging and subsequent communication.
Another example can be found in tool use by crows [4]. Food can
be placed in a contraption such that crows must fashion hooks
from pieces of wire to get the food. To be successful, crows must
gather information about objects and constraints in their
environment via sensory explorations that unfold on shorter
timescales. Impressively, crows are able also to integrate and
process this information on longer timescales for the purpose of
tool construction and usage. Honey bee foraging and communi-
cation, and crow tool construction and usage, are examples of
highly intelligent skills that nonetheless appear grounded in more
basic perceptual-motor interactions.
Intelligent human behaviors may also be supported by
perceptual-motor interactions, even though the repertoire of
human goals and intentions is far richer than that exhibited by
other species. One case that is analogous to the honey bee and
crow examples, and the focus of the present study, is drawing
a visual scene from memory. Perceptual-motor interactions guide
eye movements during an initial study period, to gather visual
information for the purpose of drawing the scene afterwards.
Perceptual-motor interactions during study may be encoded to
guide movements again during drawing, which would carry
a tendency to reproduce whatever aspects of study movements are
encoded. This kind of memory is analogous to how bee
movements are memorized to locate and then communicate the
location of resources.
The present experiment and analyses were designed to examine
the role of memory in encoding and then rendering a visual scene.
Our central research question is whether drawing from memory
can be theorized and analyzed as a reenactment of visual
information gathering. Reenactment does not necessarily mean
that trajectories of eye movements during study are isomorphic
with eye and pen trajectories during drawing. Instead, reenact-
ment can be construed more generally, in that only some aspects
of the spatial and temporal extents of eye trajectories during study
may reproduced later during drawing, and some temporal and
spatial relations may undergo nonlinear transformations as
a function of memory. Evidence for reenactment via memory
would constitute perceptual-motor coordination of eye movements
during study with subsequent eye and pen movements during
drawing.
The primary alternative to perceptual-motor coordination is
that visual memory abstracts away from the specific perceptual-
PLOS ONE | www.plosone.org 1 March 2013 | Volume 8 | Issue 3 | e58464
motor interactions that guide eye movements [5]. Symbolic
representation is the most commonly hypothesized and examined
form of visual memory, which seems apt for memory tasks that
encourage symbolic representation. For instance, consider experi-
ments in which participants are tasked with providing verbal
descriptions of scenes after viewing them [6], or providing verbal
answers to questions about scenes [7]. Language use may
encourage symbolic or symbolic-like encoding in visual memory,
and there is abundant evidence that memory processes in general
are task-dependent [8]. Given this evidence, we are led to ask how
visual memory operates when the task does not seem symbolic, as
in the case of encoding and then rendering a visual scene from
memory.
Evidence from previous studies suggests that, in perceptual-
motor tasks like drawing, memory is based more in perceptual-
motor encodings than symbolic encodings. For example, in
a classic study by Ballard, Hayhoe and Pelz [9], participants’
eye movements were recorded while they performed a copying
task. A model comprised of a particular configuration of blocks
was displayed on a computer screen, and participants used a mouse
to drag and drop blocks from a resource pile to copy the model.
Analyses of eye movements showed that perceptual-motor
interactions were used to offload visual memory onto the visual
display itself. The evidence for offloading was that eye movements
were made back to the model throughout the dragging and dropping
of blocks, which indicated that participants were unwilling to
symbolically encode the color and position of each block. Instead,
eye movements back to the model served as an external memory of
sorts. Tasks such as jigsaw puzzle completion and copying a picture
have yielded similar findings showing that perceptual-motor
interactions can serve memory aids [10].
Drawing from memory is different than the aforementioned
tasks because the model is not visually available at the time of
drawing. Therefore the environment cannot directly serve as an
external memory. Nonetheless, perceptual-motor interactions may
still be integral to memory, in that direct correspondences may be
encoded between scene viewing actions and subsequent drawing
actions. It is possible that, when studying an image, the eyes trace
a trajectory that follows the lines, curves, and features to be drawn
later, in the same order, placement, and orientation. A related
hypothesis has been proposed for recalling and visualizing images
from memory, rather than drawing images from memory. The
hypothesis goes by the name of scanpath theory [11,12], and the basic
tenet is that eye trajectories used to encode a scene are ‘‘retraced’’
when the scene is recalled from memory. Retracing the eye
trajectory is hypothesized to engage visual imagery and reinstate
the memory. Evidence for scanpath theory is mixed, with earlier
studies failing to show direct support [13,14], although some
indirect support was found [15]. Subsequent studies employed
more sophisticated methods and found that eye trajectories while
viewing scenes were correlated with eye trajectories while
visualizing, thinking, or talking about those same scenes [16,17].
Scanpath theory continues to be debated [18], and drawing
from memory adds a new dimension to the debate. In drawing
from memory, eye trajectories during study and pen trajectories
during drawing can be framed by corresponding physical
dimensions, thereby providing an opportunity for the trajectories
themselves to fall into direct correspondence with each other. In
fact, eye and pen trajectories are directly coordinated during the
act of drawing, when memory is not needed to bridge the gap
between studying an image and then drawing it later. For instance,
previous studies of hand-eye coordination have found direct
correspondence between eye location and arm velocity when
reaching for targets [19]. When drawing simple shapes, the eyes
tend to track where the pen touches the drawing surface. The eyes
may alternately lead or follow the pen, with a general tendency to
be drawn towards minima of tangential arm velocity [20]. Eyes
also tend to lead and follow the hands in more complex tasks like
making a sandwich [21].
In drawing from memory, our hypothesis is that the potential
for direct correspondences between eye and pen trajectories will
evoke memory encodings that link eye trajectories during study
with eye and pen trajectories during drawing. Such a linkage
would be perceptual-motor in nature, rather than symbolic. It
would also be consistent with the basic premise of scanpath theory.
A test of our hypothesis requires two issues to be addressed in
designing an experiment and method of analysis. First, to rule out
purely symbolic or purely perceptual encoding hypotheses,
trajectories during study and drawing periods must contain
correspondences that are specific to a given person studying and
then drawing a given image. Otherwise, correspondences may
stem from symbolic or spatial properties of an image, or from
general tendencies in eye movement patterns, such as a pre-
dominance of horizontal movements or movements towards the
center of the visual field.
Second, while it is possible for correspondences between
trajectories to be expressed as spatiotemporal co-location, as
hypothesized in scanpath theory, one might instead expect purely
spatial correspondences when drawing from memory. This
expectation arises because, in its final form, a sketch is purely
spatial in nature. Thus memory encodings need only support
spatial correspondences between study and drawing trajectories.
Moreover, drawing from memory may only evoke correspon-
dences at coarse spatial scales, given that fine-grained spatial
information may not be preserved in rough sketches by untrained
artists. By contrast, the most literal interpretation of scanpath
theory would require that study and drawing trajectories visit the
same locations for the same durations in the same order.
Here we present an experiment and analyses designed to
compare eye and pen trajectories at varied temporal and spatial
scales, in order to test for perceptual-motor encodings of visual
images in drawing from memory. Such encodings would support
extensions of hand-eye coordination via memory, and provide
evidence for a generalized version of scanpath theory. Natural
scenes rich in content were chosen as stimuli to support relatively
long viewing periods to gather visual information, thereby
providing us with sufficiently long trajectories for analysis. Natural
scenes also ensured that images contained features across a wide
range of spatial scales, thereby providing an opportunity for
trajectories to reflect both coarse-grained and fine-grained aspects
of scenes.
Materials and Methods
Sixteen University of California Merced undergraduates par-
ticipated in the experiment for course credit. The University of
California, Merced IRB approved this study, and each participant
signed a written consent form. Four participants were excluded
due to inability to calibrate with the eye-tracker below an error
threshold of one degree of visual angle. One additional participant
was excluded for failing to perform the drawing task properly,
leaving data from eleven participants for all analyses. Participants
were 18–22 years old, and nine of them were female. Five of them
self-identified as Asian, three as White, two as African American,
one as Hispanic, and one as Other. Seven participants self-
identified as bilingual or trilingual (all spoke English as one of these
languages). None of the participants reported being expert artists.
Hand-Eye Coordination at Multiple Scales
PLOS ONE | www.plosone.org 2 March 2013 | Volume 8 | Issue 3 | e58464
Six images of natural scenes were selected from a collection of
National Geographic’s Photo of the Day website: a canal lined
with boats and buildings, a whale breaching with mountains in the
background, children in a field, a flock of birds on a large tree in
a lagoon, a carnivorous plant dotted with water droplets, and a sea
anemone against a black background (see Figures S1, S2, S3, S4,
S5, S6, S7, S8, S9, S10, S11). Each original image was cropped to
160061100 pixels in resolution, and then up-sampled to
192061200 using the Python image manipulation library, in
order to match screen resolution. The complexity of natural scenes
helped to ensure that participants needed a relatively long study
time to encode each image, thereby eliciting long eye movement
trajectories needed for analyses. The variety and novelty of natural
scenes helped to minimize the chance of practice effects and
familiarity effects. Given the complexity, variety, and novelty of
these scenes, and given that participants were not expert sketch
artists, the task of drawing them from memory was more
challenging than experiments using simple line drawings.
Each participant was fitted with the head-mounted eyetracker
so that it was snug on their head. After adjusting cameras and
focusing each camera, thresholds for detecting pupils were
automatically calibrated. Each participant then looked at each
corner of the screen according to instructions from the experi-
menter. This allowed the experimenter to see if the track was lost
in a given corner, and if so, to readjust the cameras. A nine-point
calibration was performed, followed by a nine-point validation.
Validation served to check for tracking errors as a function of
location on the screen. The experiment began only after validation
showed minimal errors across the screen, and drift was checked
and corrected if necessary between each trial. Each drift correction
was examined after data collection to ensure no major drift had
occurred during the experiment, and no large differences in error
were found.
Each participant was seated approximately 36’’ in front of a 24’’
flat panel LCD monitor (visual angle of 14 degrees). Participants
viewed each of the six images in random order for 30 seconds per
image. After each image, the screen was blanked and the
instruction ‘‘Prepare to Draw’’ appeared for 4 seconds, after
which the screen was blanked and participants were able to draw
in black and white for 90 seconds using a Wacom Graphire
digitizing pad (93 mm in height6127 mm in width, with accuracy
of 60.25 mm and an operating sampling rate of 60 Hz). The
viewing period of 30 sec was found through pilot work to be
adequate time for participants to choose and encode features of
each scene to be drawn. The 90 sec drawing period was found to
be ample time for completing a rough sketch of scene that
captured the basic features memorized. Line thickness of the
drawing was independent of pressure on the tablet, and lines could
not be erased once created. During both study and drawing
phases, monocular eye position was recorded at 500 Hz using an
Eye Link II head mounted eye tracker. Note that, unlike drawing
on paper or on a touch screen, the eyes tracked lines being drawn
on the screen, instead of the pen itself. The digitizing pad has the
advantage that the pen, hand, and arm do not occlude the image
being drawn.
The data for each trial consisted of three position time series, all
in the same XY coordinates: study eye position (XY
es
), drawing eye
position (XY
ed
), and drawing pen position (XY
pd
). Blinks and other
artifacts, such as off-screen eye positions, were removed from the
eye position series for both study and drawing phases. Mean
amount of data discarded during the study and drawing phases
was 4.0% and 8.2%, respectively. The pen position series included
only samples when the pen was touching the pad, i.e. when lines
were being drawn. The data thus offers three potential compar-
isons: XY
es
6XY
ed
,XY
es
6XY
pd
, and XY
ed
6XY
pd
. Eye positions
were sampled every 2 milliseconds at times t
es
and t
ed
during study
and drawing periods, respectively. Pen positions were sampled
every 16.6 milliseconds at times t
pd
. Panel A of Figure 1 shows an
example of the XY
es
series obtained from one trial overlaid on the
corresponding image, down-sampled to reduce visual clutter.
Panel B shows the subsequent XY
pd
series for this trial, rendered as
the original sketch image, with the corresponding XY
ed
series
overlaid and down-sampled.
Results
We first tested whether the present experiment replicated the
spatiotemporal co-location between eye and pen found in previous
studies of drawing, and more generally in previous studies of hand-
eye coordination. Spatiotemporal co-location was measured by
Euclidean distance between eye and pen positions as a function of
time, D[XY(t
ed
), XY(t
pd
)]. Thus a distance was computed for all
possible pairs of positions, creating a matrix D of dimensionality t
ed
6t
pd
for each trial. Each matrix was normalized by dividing each
distance by the mean distance over the whole matrix. Normalized
values were binned as a function of temporal lag L=t
ed
t
pd
, and
averaged within each bin. Hand-eye coordination is expressed
when the mean of D[XY(t
ed
), XY(t
pd
)] decreases as |L| approaches
zero.
Results replicated previous studies [20] showing that hands and
eyes tend to co-locate when engaged in tasks like drawing (Figure 2,
blue line). D[XY(t
ed
), XY(t
pd
)] was minimal when t
ed
,t
pd
, and
increased to an asymptote near chance co-location as t
ed
and t
pd
diverged in the range 210 sec,L,+10 sec. The symmetry of
approach towards baseline indicates that, on average, eye both led
and followed the pen in equal proportions as a function of distance
between them. This function showed the same symmetric
approach to a minimum near |L| = 0 for each individual
participant and image (see Figure S12).
Next we tested whether eye trajectories during study exhibit
spatiotemporal co-location with eye and pen trajectories pro-
duced during drawing. To align trajectories, the beginning of
each time period was set to time zero, and then XY
es
times were
multiplied by a factor of three to equate the lengths of
trajectories (study periods were 30 sec whereas drawing periods
were 90 sec). Dmatrices were computed as described above,
and Figure 2 shows the resulting averages as a function of L
(green and red lines; see Figures S1, S2, S3, S4, S5, S6, S7, S8,
S9, S10, S11 for individual participant and image results). Co-
location was not evident in comparisons between study and
drawing trajectories, in that mean spatial distance did not vary
significantly as a function of lag.
To summarize the first analysis, spatiotemporal co-location
yielded evidence for concurrent coordination between eye and
pen during drawing, but no such evidence was found for
coordination via memory between study and drawing periods.
In isolation, this null result may mean that perceptual-motor
encodings did not serve to link eye trajectories during study
with time-warped versions of these trajectories during drawing.
Alternatively, drawing trajectories may be linked to study
trajectories, but not as stretched out, temporally preserved
copies. Instead, perceptual-motor encodings of trajectories may
be purely spatial in nature, or if any temporal information is
preserved, it may be obscured by nonlinear transformations.
Whatever the case may be, results failed to provide evidence for
a simple application of scanpath theory to eye and pen
trajectories in drawing from memory.
Hand-Eye Coordination at Multiple Scales
PLOS ONE | www.plosone.org 3 March 2013 | Volume 8 | Issue 3 | e58464
Hand-Eye Coordination at Multiple Scales
PLOS ONE | www.plosone.org 4 March 2013 | Volume 8 | Issue 3 | e58464
Spatial Allan Factor Analysis
It is possible that more complex temporal transformations might
yield evidence in Dmatrices that eye trajectories during study were
temporally related to eye and pen trajectories during drawing.
However, the end product of a drawing is purely spatial in nature,
which leads us instead to focus on the spatial configurations of
trajectories. While eye and pen may not visit the same scene
features in corresponding temporal orders and durations between
study and drawing periods, trajectories may nonetheless concen-
trate on the same features in the same locales. Our rationale for
considering purely spatial co-location is that the task of drawing
may encourage spatial alignment between study and drawing
periods, rather than temporal alignment.
Temporal information can be removed directly from the
original co-location measure by calculating D[XY
es
,XY
ed/pd
] for
all pairwise points, regardless of their time stamps. However, this
simple formulation does not readily express co-location at varying
spatial scales. It is possible that spatial configurations of eye
trajectories during study are only coarsely reproduced during
drawing, because fine-grained spatial details are either forgotten,
or lost by lack of precision in drawing behaviors or measurements.
In practical terms, this means that rich scene information
hypothesized to drive eye movements during viewing is not
present or measureable during drawing. Therefore, a measurement
of co-location at varying spatial scales may be needed to reveal the
scales at which spatial correspondences become measureable in
eye and pen trajectories.
We created a multiscale measure of spatial correspondence by
adapting the Allan Factor (AF) method developed for analyzing
clustering in temporal point processes, such as neural spike trains
[22,23]. AF analysis was originally developed to distinguish time
series generated by random (Poisson) point processes from those
with fractal (i.e. multiscale) clustering. Fractal clustering is relevant
to our present aims for two reasons. First, images of natural scenes
have been shown to exhibit fractal variations in the spatial
distribution of luminance [24], so one might expect eye trajectories
to also exhibit fractal spatial variations. For instance, the dynamics
of eye movements have been reported to exhibit fractal variations
in time, in the form of long-range correlations known as ‘‘1/f
noise’’ [25,26]. However, to our knowledge, no one has reported
spatial fractal analyses of eye trajectories. The second reason why
fractal clustering is relevant is that fractal analyses like AF are
inherently multiscale, which provides us with a basis for extending
AF analysis to examine correspondences between point processes
at varying spatial scales.
First we describe AF analysis as originally formulated for
temporal point processes. Given a one-dimensional point process
spanning a given length of time T
total
, AF analysis begins by
dividing the series into adjacent windows of duration T, where T
varies from a minimum to maximum in powers of two, i.e. T
min
and a value less than T
total
, such as T
total
/4. The number of points
(i.e. events) is counted in each window, where N
k
is the number of
points in the kth window of size T. Differences between adjacent
counts are calculated as
d(T)~Njz1(T){Nj(T),
and the AF value for a given timescale Tis calculated as follows,
where E[] is expected value:
A(T)~
EEdTðÞ
2
hi
2
EEN TðÞ½
Poisson processes yield A(T),1 for all T, whereas fractal
processes yield A(T)!T. This formulation of AF is tailored for
temporal point processes, but we can extend it straightforwardly
for spatial point processes. We did this by partitioning image and
drawing spaces containing sets of XY points (Fig. 1B) into square
tiles of size S(i.e. area in pixels). Some number Nof XY points fell
within each tile, and tile size Swas varied analogous to window
size T. Tile counts were compared against adjacent tiles in the X
and Ydimensions, N
x
and N
y
, by computing differences analogous
to the one-dimensional temporal case (similar to Haar wavelets
[27]):
dx(S)~Nxz1(S){Nx(S)and dy(S)~Nyz1(S){Ny(S):
The two-dimensional AF function is then
A(S)~
EE½dx(S)2z
EE½dy(S)2
4
EE½N(S)
A(S) and A(T) have the same property whereby a Poisson process
will yield constant AF variance near unity, and fractal point
processes will yield functions that scale with Sand T, respectively.
To extend the AF method further for measuring correspon-
dences between two sets of XY points, aand b, the cosines of angles
between their respective d
x
(S) and d
y
(S) vectors were computed at
each spatial scale:
Figure 1. Example data from one participant studying one image (A) and then drawing that image (B). Eye trajectories were down-
sampled to 15 Hz for the figure to reduce visual clutter. Drawing overlay (blue) shows example tiles used for Allan Factor analyses.
doi:10.1371/journal.pone.0058464.g001
Figure 2. Results of co-location analysis plotted as a function of
temporal lag. Distances were normalized by the mean distance over
all pairwise comparisons.
doi:10.1371/journal.pone.0058464.g002
Hand-Eye Coordination at Multiple Scales
PLOS ONE | www.plosone.org 5 March 2013 | Volume 8 | Issue 3 | e58464
Ca,b(S)~P
MxSðÞ
x~1
dax(S)dbx (S)
2dax(S)
kk
dbx(S)
kk
zP
MySðÞ
y~1
day(S)dby (S)
2day(S)
dby(S)
,
where M
x
(S) and M
y
(S) were the numbers of horizontal and vertical
comparisons at each scale, respectively. Cosines were used because
they normalize for overall counts per tile and differences between
tiles. On this measure, there is greater correspondence between
two sets of XY points at each given scale Sto the extent that C
a,b
(S)
approaches one, where correspondence is measured in terms of co-
location in spatial configuration. XY configurations are measured
as being dissimilar as C
a,b
(S) approaches zero.
To test our hypothesis of temporally extended coordination, A(S)
functions need to be compared between study and drawing
periods. In addition, we are interested in testing whether AF
functions were anchored to the images being drawn. The task of
drawing from memory would seem to encourage eye movements
that follow the contours of visually salient features in natural
scenes. If so, the spatial AF analysis that we formulated for
comparing eye and pen trajectories should also work for
comparing trajectories with the spatial distributions of visual
quantities corresponding to salient features. It is likely that eye and
pen trajectories will also be guided by top-down factors related to
intentions, past experiences, and the like [28]. However, in this
light, the task of drawing is itself a top-down factor that should
draw attention to visually salient features of images to be drawn
[29]. To quantify these features, images were passed through
a model of visual saliency based on theories of low-level visual
processing [30]. The model takes a greyscale bitmap as input, and
produces a saliency map as output (see Figure S13). Maps were
converted to sets of XY image points, where numbers of points
were linearly related to saliency values, and set equal to numbers
of eye position samples collected per image in the drawing
condition.
A(S) functions were computed for XY points in eye trajectories
recorded during study and drawing conditions, for pen trajectories
during drawing, and for saliency maps of the six images of natural
scenes. Figure 3A shows that, on average, AF values increased
monotonically as a function of Sfor all four types of XY points (see
Materials S1 for individual participant and image results, Figure
S14). A(S) functions were linear in logarithmic coordinates for eye
configurations, with aexponents estimated near ,0.5 using linear
regression. This linear trend indicates fractal clustering of eye
configurations, which is consistent with clustering in the spatial
distribution of luminance values in images of natural scenes [24].
By contrast, A(S) functions for pen and saliency map configurations
were monotonically increasing but curvilinear, indicative of
clustering only at the larger spatial scales. This restricted scale of
clustering may be due to slower pen movements, reduced
resolution in pen recordings, and/or spatial smoothing in the
saliency model.
Spatial co-location was measured by computing C
a,b
(S) for all
possible pairwise comparisons between XY configurations.
Figure 3B shows that co-location increased with larger scales in
all cases, and as expected, co-location was greatest for concurrent
eye and pen trajectories during drawing (see Materials S1 for
individual participant results, Figure S15). These initial results
confirm that C
a,b
(S) functions capture hand-eye coordination as
originally measured by spatiotemporal co-location, i.e. D[XY(t
ed
),
XY(t
pd
)]. Results also confirm that coordination via memory is not
detectable at finer spatial scales, which may be due to memory
limits or measurement error. Results also provide initial evidence
that the spatial configurations of both eye and pen trajectories are
co-located with the visually salient features of scene images at
larger scales. This evidence is consistent with the expectation that
the task of drawing from memory anchors the eyes and pen to
visually salient features to be drawn.
Spatial similarity was evident for all comparisons, but compar-
isons with two different kinds of baselines are needed to determine
the sources of similarity. Our hypothesized source of similarity is
perceptual-motor encoding that supports the coordination of eye
and pen movements across study and drawing periods. However,
we must test this hypothesis against two alternative explanations.
One alternative is that trajectories are spatially similar merely
because participants produce characteristic patterns of eye move-
ments, regardless of whether they are studying or drawing scenes,
and regardless of which scene is being drawn. As noted earlier,
characteristic patterns may include general tendencies towards
horizontal or central eye movements. These patterns could be
generated without memory to carry information from the study to
test period. The other alternative is that memory is engaged, but in
the form of symbolic encodings instead of perceptual-motor
encodings. Instead of memory for eye positions during study,
images may be encoded in terms of symbolic representations that
can be expressed linguistically, such as ‘‘there is canal running
down the middle with buildings and boats lining either side’’.
The two kinds of C
a,b
(S) baseline functions are based on image
surrogates and participant surrogates, respectively. For image
surrogates, eye and pen trajectories were paired with trajectories
produced by the same participant, but for a different, randomly
chosen image. For instance, a given participant’s eye trajectory
while studying the canal scene might be compared with his/her
eye or pen trajectories while drawing the whale scene. If spatial
similarities found between study and drawing are due to general
tendencies in the shapes of trajectories, then C
a,b
(S) values for
image surrogates should be the same as for original comparisons.
For participant surrogates, trajectories for the same image were
paired, but produced by different participants paired at random. If
spatial similarities are due to symbolic or purely visual encodings
based solely on the scenes themselves, then C
a,b
(S) values for
participant surrogates should be the same as for original
comparisons.
Both original and surrogate baseline C
a,b
(S) functions were
computed for each trial, and the latter were subtracted from the
former for targeted comparisons. Differences were summed over S
for each comparison, and T-tests were used to determine whether
these sums were reliably greater than zero (means of these sums
are shown in Figure 4). Results of statistical tests (Table 1, see also
Table S1 in Materials S1) showed that all comparisons were
significantly different from baseline with the exception of Eye(-
Study) 6Pen(Draw). We conclude that each eye trajectory during
each study period was specifically reproduced in corresponding
eye and pen configurations while drawing, but only at larger
spatial scales. The finding that original C
a,b
(S) functions showed
greater similar than both image and participant surrogates is
evidence that memory encodings were at least partly perceptual-
motor in nature. This conclusion is not mutually exclusive with the
possibility that encodings were also symbolic and/or visual in
nature, or that similarities were partly driven by general patterns
in eye movements.
Discussion
The drawing experiment reported herein provides evidence that
memory can serve to coordinate perceptual-motor interactions
over longer timescales than those operative in more immediate
Hand-Eye Coordination at Multiple Scales
PLOS ONE | www.plosone.org 6 March 2013 | Volume 8 | Issue 3 | e58464
interactions, such as hand-eye coordinations. Drawing is a task
that evokes hand-eye coordination, as found in temporally aligned
co-locations of eye and pen trajectories produced while drawing.
Drawing from memory is a task that also evokes coordination
between study and drawing periods, but evidence of this co-
ordination was found only in terms of spatial co-location, without
temporal alignment, and only at the larger spatial scales. AF
analyses showed that the degree of coordination, as measured by
coarse-grained spatial overlap, varied as a function of condition
and measure. Temporal analyses were insensitive to these
variations.
The correspondences of drawing trajectories with study
trajectories can be interpreted as evidence for a version of
scanpath theory applied to the task of drawing visual images from
memory, rather than recalling them from memory. This version
would need to be generalized for spatial configurations of
trajectories, independent of their temporal extents. The temporal
extents of eye trajectories may be preserved in other task contexts,
such as those that emphasize the temporal ordering and/or
durations of fixations. The theory would have to explain how the
spatial and temporal properties of perceptual-motor encodings can
vary as a function of task demands and intentions. The theory
would stand in contrast to memory processes that operate in
purely visual or symbolic modes that are independent of task
context. Purely visual or symbolic representations appear to be
inadequate because surrogate baseline analyses showed that the
particularities of eye trajectories for a given study session were
reproduced during the subsequent drawing session.
It would be interesting to investigate whether current theories of
visual-motor processing might be construed to account for the
present results. For instance, Cagli and colleagues recently
reported a Dynamic Bayes Network (DBN) that simulates the
online interactions between eyes and hands of the course of
copying simple line drawings [29,31,32]. Models like these may
encode information gathered during study periods as priors on
perceptual-motor interactions that unfold during drawing. If one
views scanpath theory as a general hypothesis about the relation-
ship between memory encodings and subsequent actions, then
DBNs may be seen as computational models that capture the basic
tenets of scanpath theory, and thereby provide a means of
applying them to tasks like drawing from memory.
Finally, results suggest that perceptual-motor coordination at
multiple scales is supportive of intelligent behaviors like commu-
nication and artwork, in species ranging from honey bees to
humans. Hand-eye coordination is typically considered more
dexterous than intelligent, in that reciprocal interactions between
perceptual and motor systems are concurrent and based primarily
upon immediate timing and co-location. Behaviors become more
intelligent as memory, planning, and abstraction become more
involved, and coordination becomes more complex. In drawing
from memory, higher-order functions are modestly engaged in
a task that allows for direct comparisons between concurrent and
non-concurrent coordination. In this light, higher-order cognitive
Figure 3. Mean AF functions (left) and cosine similarities (right) plotted in logarithmic coordinates as a function of tile size, for
configuration of points from eye, pen, and image data.
doi:10.1371/journal.pone.0058464.g003
Table 1. Means of C
a,b
(S) functions minus their respective
baselines, for each of the conditions shown in Figure 4.
Mean Std Error t value p value
Image X
- Eye(Study) 0.258 0.047 5.486 0.000
- Eye(Draw) 0.104 0.044 2.356 0.022
- Pen(Draw) 0.133 0.053 2.503 0.015
Eye(Study) X Eye(Draw)
Baseline:
- Image 0.140 0.052 2.707 0.009
- Participant 0.267 0.059 4.529 0.000
Eye(Study) X Pen(Draw)
Baseline:
- Image 0.059 0.063 0.932 0.355
- Participant 0.212 0.063 3.366 0.001
Eye(Draw) X Pen(Draw)
Baseline:
- Image 0.833 0.114 7.289 0.000
- Participant 0.969 0.096 10.134 0.000
doi:10.1371/journal.pone.0058464.t001
Hand-Eye Coordination at Multiple Scales
PLOS ONE | www.plosone.org 7 March 2013 | Volume 8 | Issue 3 | e58464
functions may be viewed as multiscale extensions of more basic
perceptual-motor interactions.
Supporting Information
Figure S1 Individual trial examples with fixations. One
example image (A) and corresponding drawing (B) from each of
the 11 participants, with eye tracking positions down-sampled to
15 Hz to reduce visual clutter. Five of six images are shown twice,
and each image is shown at least once.
(TIF)
Figure S2 Individual trial examples with fixations. One
example image (A) and corresponding drawing (B) from each of
the 11 participants, with eye tracking positions down-sampled to
15 Hz to reduce visual clutter. Five of six images are shown twice,
and each image is shown at least once.
(TIF)
Figure S3 Individual trial examples with fixations. One
example image (A) and corresponding drawing (B) from each of
the 11 participants, with eye tracking positions down-sampled to
15 Hz to reduce visual clutter. Five of six images are shown twice,
and each image is shown at least once.
(TIF)
Figure S4 Individual trial examples with fixations. One
example image (A) and corresponding drawing (B) from each of
the 11 participants, with eye tracking positions down-sampled to
15 Hz to reduce visual clutter. Five of six images are shown twice,
and each image is shown at least once.
(TIF)
Figure S5 Individual trial examples with fixations. One
example image (A) and corresponding drawing (B) from each of
the 11 participants, with eye tracking positions down-sampled to
15 Hz to reduce visual clutter. Five of six images are shown twice,
and each image is shown at least once.
(TIF)
Figure S6 Individual trial examples with fixations. One
example image (A) and corresponding drawing (B) from each of
the 11 participants, with eye tracking positions down-sampled to
15 Hz to reduce visual clutter. Five of six images are shown twice,
and each image is shown at least once.
(TIF)
Figure S7 Individual trial examples with fixations. One
example image (A) and corresponding drawing (B) from each of
the 11 participants, with eye tracking positions down-sampled to
15 Hz to reduce visual clutter. Five of six images are shown twice,
and each image is shown at least once.
(TIF)
Figure S8 Individual trial examples with fixations. One
example image (A) and corresponding drawing (B) from each of
the 11 participants, with eye tracking positions down-sampled to
15 Hz to reduce visual clutter. Five of six images are shown twice,
and each image is shown at least once.
(TIF)
Figure S9 Individual trial examples with fixations. One
example image (A) and corresponding drawing (B) from each of
the 11 participants, with eye tracking positions down-sampled to
15 Hz to reduce visual clutter. Five of six images are shown twice,
and each image is shown at least once.
(TIF)
Figure S10 Individual trial examples with fixations.
One example image (A) and corresponding drawing (B) from each
of the 11 participants, with eye tracking positions down-sampled to
15 Hz to reduce visual clutter. Five of six images are shown twice,
and each image is shown at least once.
(TIF)
Figure S11 Individual trial examples with fixations.
One example image (A) and corresponding drawing (B) from each
of the 11 participants, with eye tracking positions down-sampled to
15 Hz to reduce visual clutter. Five of six images are shown twice,
and each image is shown at least once.
(TIF)
Figure S12 Comparison co-location plot. Plots of co-
location functions averaged for each participant (left column)
and each image (right column), separated into three comparison
conditions: XY
gd
6XY
pd
(top), XY
gs
6XY
gd
(middle), and XY
gs
6
Figure 4.
C
a,b
(
S
) functions summed over
S
, and subtracted from image (filled bars) and participant (open bars) surrogate baselines,
with standard error bars. Cosine similarities reliably above baseline denoted by an *.
doi:10.1371/journal.pone.0058464.g004
Hand-Eye Coordination at Multiple Scales
PLOS ONE | www.plosone.org 8 March 2013 | Volume 8 | Issue 3 | e58464
XY
pd
(bottom). The periodic pattern in some functions was likely
due to differences in sample rates.
(TIF)
Figure S13 Saliency maps of stimulus images. Saliency
heat maps for each of the six images, overlaid with example
samples from their corresponding probability distributions.
(TIF)
Figure S14 Allan Factor functions. Plots of Allan factor
functions averaged for each participant in the gaze-study (top-left),
gaze-draw (top-right), and pen-draw conditions (bottom-left), and
for each image (bottom-right).
(TIF)
Figure S15 Ca,b(S) functions. Plots of Ca,b(S) functions
averaged per participant for each of the series shown in Figure 3B
from main text.
(TIF)
Materials S1 Supplementary Materials and Methods.
File contains: Table S1 Means of Ca,b(S). Means of C
a,b
(S)
functions minus their respective baselines, for each of the
conditions shown in Figure 4 from the main text.
(DOCX)
Acknowledgments
We thank Bryan Kerster for his help with programming the eye tracker.
Author Contributions
Conceived and designed the experiments: SH CK TR MS. Performed the
experiments: SH TR. Analyzed the data: TR SH CK. Contributed
reagents/materials/analysis tools: MS CK TR SH. Wrote the paper: CK
SH TR MS.
References
1. Gibson JJ (1979) The ecological approach to visual perception. Boston:
Houghton Mifflin. xiv, 332 p.
2. Lee DN, Kalmus H (1980) The Optic Flow Field: The Foundation of Vision
[and Discussion]. Philosophical Transactions of the Royal Society of London B,
Biological Sciences 290: 169–179.
3. von Frisch K (1967) The dance language and orientation of bees. Cambridge,
MA: Harvard University Press.
4. Weir AAS, Chappell J, Kacelnik A (2002) Shaping of Hooks in New Caledonian
Crows. Science 297: 981.
5. Wood JN, Grafman J (2003) Human prefron tal cortex: processing and
representational perspectives. Nat Rev Neurosci 4: 139–147.
6. Griffin ZM, Bock K (2000) What the Eyes Say About Speaking. Psychological
Science 11: 274–279.
7. Richardson DC, Dale R (2005) Looking To Understand: The Coupling Between
Speakers’ and Listeners’ Eye Movements and Its Relationship to Discourse
Comprehension. Cognitive Science 29: 1045–1060.
8. Hayhoe M (2000) Vision Using Routines: A Functional Account of Vision.
Visual Cognition 7: 43–64.
9. Ballard DH, Hayhoe MM, Pelz JB (1995) Memory Representations in Natural
Tasks. Journal of Cognitive Neuroscience 7: 66–80.
10. Land M, Tatler B (2009) Looking and acting: vision and eye movements in
natural behaviour. New York: Oxford University Press.
11. Hebb DO (1968) Concerning imagery. Psychological Review 75: 466–477.
12. Noton D, S tark L (1971) Scanpaths in eye movements during pattern
perception. Science 171: 308–311.
13. Antrobus JS, Antrobus JS, Singer JL (1964) Eye movements accompanying
daydreaming, visual imagery, and thought suppression. The Journal of
Abnormal and Social Psychology 69: 244–252.
14. Zikmund V (1966) Oculomotor activity during visual imagery of a moving
stimulus pattern. Studia Psychologica 8: 254–274.
15. Finke RA, Pinker S (1983) Directional scanning of remembered visual patterns.
Journal of Experimental Psychology: Learning, Memory, and Cognition 9: 398–
410.
16. Richardson DC, Dale R, Kirkham NZ (2007) The Art of Conversation Is
Coordination. Psychological Science 18: 407–413.
17. Spivey MJ, Geng JJ (2001) Oculomotor mechanisms activated by imagery and
memory: eye movements to absent objects. Psychological Research 65: 235–241.
18. Foulsham T, Kingstone A (2012) Fixation-Dependent Memory for Natural
Scenes: An Experimental Test of Scanpath Theory. Journal of Experimental
Psychology: General.
19. Johansson RS, Westling G, Ba¨ckstro¨ m A, Flanagan JR (2001) Eye–Hand
Coordination in Object Manipulation. The Journal of Neuroscience 21: 6917–
6932.
20. Gowen E, Miall RC (2006) Eye–hand interactions in tracing and drawing tasks.
Human Movement Science 25: 568–585.
21. Hayhoe MM, Shrivastava A, Mruczek R, Pelz JB (2003) Visual memory and
motor planning in a natural task. Journal of Vision 3.
22. Allan DW (1966) Statistics of atomic frequency standards. Proceedings of the
IEEE 54: 221–230.
23. Teich MC, Heneghan C, Lowen SB, Ozaki T, Kaplan E (1997) Fractal
character of the neural spike train in the visual system of the cat. J Opt Soc Am A
14: 529–546.
24. Graham DJ, Field DJ (2007) Statistical regularities of art images and natural
scenes: Spectra, sparseness and nonlinearities. Spatial Vision 21: 149–164.
25. Aks DJ, Zelinsk y GJ, Sprott JC (2002) Memory across eye-movements: 1/f
dynamic in visual search. Nonlinear Dynamics, Psychology, & Life Sciences 6:
1–25.
26. Shelhamer M, Joiner WM (2003) Saccades exhibit abrupt transition between
reactive and predictive, predictive saccade sequences have long-term correla-
tions. Journal of Neurophysiology 90: 2763–2769.
27. Struzik Z, Siebes A (1999) The Haar Wavelet Transform in the Time Series
Similarity Paradigm. In: Zytkow J, Rauch J, editors. Principles of Data Mining
and Knowledge Discovery: Springer Berlin/Heidelberg. 12–22.
28. Tatler BW, Hayhoe MM, Land MF, Ballard DH (2011) Eye guidance in natural
vision: Reinterpreting salience. Journal of Vision 11.
29. Coen-Cagli R, Coraggio P, Napoletano P, Schwartz O, Ferraro M, et al. (2009)
Visuomotor characterization of eye movements in a drawing task. Vision
Research 49: 810–818.
30. Harel J, Koch C, Perona P (2007) Graph-based visual saliency. Advances in
neural information processing systems. Cambridge, MA: MIT Press. 545–552.
31. Coen Cagli R, Coraggio P, Napoletano P, Boccign one G (2007) The Bayesian
Draughtsman: A Model for Visuomotor Coordination in Drawing Advances in
Brain, Vision, and Artificial Intelligence. In: Mele F, Ramella G, Santillo S,
Ventriglia F, editors: Springer Berlin/Heidelberg. 161–170.
32. Cagli RC, Napoletano P, Coraggio P, Boccignone G, De Santis A. Sensorimotor
coupling via dynamic bayesian networks; 2008. IEEE. 1582–1587.
Hand-Eye Coordination at Multiple Scales
PLOS ONE | www.plosone.org 9 March 2013 | Volume 8 | Issue 3 | e58464

Supplementary resources (16)

... Scan path theory posits that eye trajectories that encode a scene are retraced during memory recall, influencing drawing performance. Additionally, memory encoding in perceptual-motor tasks like drawing is predominantly based on perceptual-motor encodings rather than symbolic encodings [3]. ...
... Active visual and spatial WM processes have been shown to play distinct roles in copying geometrical patterns and drawing from memory, with the application of continuous WM models aiding in understanding their specific contributions [1]. Moreover, findings suggest that drawing complex scenes presents greater challenges compared to simple line drawings, emphasizing the role of WM in managing task complexity [3]. ...
... Research on hand-eye coordination during drawing tasks reveals the coordination between study and drawing periods, underscoring the dynamic interplay between perceptual and motor systems [3]. ...
... There is exciting research identifying specialized neurons in the hippocampus that respond to specific places or locations, both in animals and in humans. 62 In addition, there appear to be "grid cells" that help animals navigate through their environments in a kind of coordinate system of equilateral triangles. 63 The science writer Nicholas Carr has noted that computer assisted navigation can circumvent or even hamper the brain's spatial representation system. ...
... When drawing, the brain brings experiential aspects of an environment into working memory in order to construct the image on the page. 62 By seeing in the mind's eye, the object is connected to its place, time, and circumstance. ...
... They were references, memory prompts, for use in the studio by practicing designers. [60,61,62] Organized by subject matter, the 80-page book does not have large text portions, but rather contains captioned illustrations of building plans, sections, elevations, and details. By comparing ancient examples with "modern" works by esteemed masters such as Bramante, Raphael, Sangallo, and their contemporaries the book was an ideal teaching tool for apprentices. ...
Book
Draw In Order to See is the first book to survey the history of architectural design using the latest research in cognitive science and embodied cognition. Beginning with a primer on visual perception, cognitive science, design thinking, and modes of conception used by groups of architects in their practices, Mark Alan Hewitt surveys a 12,000-year period for specific information about the cognitive schemata used by Homo sapiens to make their buildings and habitats. The resulting history divides these modes of thinking into three large cognitive arcs: crafting, depicting, and assembling, within specific temporal frames. His analysis borrows from Merlin Donald’s thesis about mimetic and symbolic cognition as critical to the emergence of the modern mind, and further employs theories of enactment and embodiment to clarify their relationship to architecture. Individual chapters treat the emergence of depiction during the Renaissance, the education of architects in the modern era, Baroque illusionism and scenography, the breakdown of artisanal literacy during the Enlightenment, and modern experiments with models, montage, and illusions of movement. The author concludes with a critique of contemporary design and education, and promotes design with embodiment as a tonic for a profession in crisis, facing the challenges of climate change, energy shortages, inequality, and housing a population of over seven billion in the coming decades. This groundbreaking and valuable study presents a clear view of current research in two related fields that have not heretofore been compared, and outlines a strategy for future research. An extensive bibliography offers readers an up-to-date reference to both the science and the architectural history behind the text.
... Intrinsically, drawing can be decomposed into a series of sensoryguided reaching movements. The shape and position of the figure are essentially determined by the location where the hand or pen reaches (Battaglia-Mayer et al., 2003;Huette et al., 2013). ...
... For blind drawing tests, inputs from the proprioception and the vestibular system instead guide hand movement. The PPC coordinates the eyes and hands to modulate reaching movement (Jackson and Husain, 2006;Huette et al., 2013). Specifically, the PPC directs hand placement, adjusts velocity, and amends bias along the trajectory to the targeted loci (Buneo and Andersen, 2006;Jackson and Husain, 2006;Averbeck et al., 2009;Archambault et al., 2011;Battaglia-Mayer et al., 2015). ...
Article
Full-text available
Drawing is a comprehensive skill that primarily involves visuospatial processing, eye-hand coordination, and other higher-order cognitive functions. Various drawing tasks are widely used to assess brain function. The neuropsychological basis of drawing is extremely sophisticated. Previous work has addressed the critical role of the posterior parietal cortex (PPC) in drawing, but the specific functions of the PPC in drawing remain unclear. Functional magnetic resonance imaging and electrophysiological studies found that drawing activates the PPC. Lesion-symptom mapping studies have shown an association between PPC injury and drawing deficits in patients with global and focal cerebral pathology. These findings depicted a core framework of the fronto-parietal network in drawing tasks. Here, we review neuroimaging and electrophysiological studies applying drawing paradigms and discuss the specific functions of the PPC in visuospatial and sensorimotor aspects. Ultimately, we proposed a hypothetical model based on the dorsal stream. It demonstrates the organization of a PPC-centered network for drawing and provides systematic insights into drawing for future neuropsychological research.
... With its own unique benefits, hand drawing and sketching are pivotal in developing an intuitive understanding of scale. It can help students pay attention to proportions, perspectives, and dimensions of spatial objects and accurately represent scale as well as improve coordination between the hand, eye, and brain by performing physical and cognitive tasks (Huette et al 2013). In addition, the tactile and kinesthetic experience of hand drawing strengthens the body-perception connection, reinforcing students' understanding of scale and spatial relationships. ...
Article
Full-text available
Although spatial perception abilities are essential for improving architectural and urban design competence, the lack of appropriate educational tools has prevented them from being taught in higher education settings; accordingly, many students struggle with identifying such spatial characteristics. Using a Solomon four-group design approach, this study aimed to verify the effectiveness of virtual reality (VR)-aided training for “sense of scale,” which despite being one of the most critical abilities in design competence, is difficult to teach both in the classroom and on-site. We recruited 80 first- and second-year university students majoring in physical planning, and conducted two street-level sense of scale training sessions, as well as pre- and post-tests. We measured the sense of scale accuracy (i.e., the difference between the actual and reported values) using mean absolute percentage error, and the effect of experimental treatment was analyzed. The results demonstrated that the training was significantly effective at estimating dimensions for street width and building height but not D/H ratio. This suggests that a sense of “absolute” scale can be acquired through two VR-aided training sessions. Additional post hoc questionnaire surveys also demonstrated that the participants felt that VR-aided training was more effective for scale practice than traditional learning methods. The findings demonstrate that for spatial perception abilities that are difficult to teach with conventional teaching methods, VR-aided training could be utilized as an effective education technique.
... multiscale) methods are useful for analyzing transitions and clustering in trajectories. For example, search trajectories such as eye movement, hand-eye coordination, and foraging can be analyzed by quantifying the spatial distribution or nesting of temporal point processes using spatial Allen Factor analysis (Huette et al., 2013;Kerster et al., 2016). ...
... In drawing tasks from memory (see Fig. 4), the patient may be asked to draw a known object (e.g., a daisy), or to insert the numbers corresponding to the hours on the clock in the correct position (Bisiach and Vallar, 2000;Ronchi et al., 2009). Drawing from memory is a more complex task and requires a greater cognitive load, including the retrieval of the visual-spatial representation of the object from long-term memory (Bainbridge et al., 2019;Huette et al., 2013), and its reproduction on the sheet, making a continuous on-line comparison between the mental representation and the actual production of the intended object. ...
Chapter
Full-text available
Unilateral spatial neglect is a neuropsychological syndrome, more frequent after damage to the right cerebral hemisphere, featuring an impaired ability to orient toward the contralateral left side of space and the body, and the inability to detect and report events from that side, of which patients are not aware. In addition to these defective manifestations, patients may show gratuitous behaviors, which include additional productions, unrelated to the task's demand, such as drawings and repeated marks (perseveration) in exploratory target cancellation tasks. Neglect is a multi-component deficit, whose diverse manifestations share a divide between the attended ipsilateral and the neglected contralateral sides of space. Different pathological mechanisms contribute to neglect, including perceptual and premotor impairments, that may be specific to sensory modality, sector of space, and mental images. Neglect may manifest in different coordinate frames with reference to the body and body parts, and objects (egocentric and allocentric neglect). Spatial neglect is a severely disabling disorder, which interferes with functional recovery from stroke. The neural correlates of neglect include cerebral cortical regions surrounding the sylvian fissure: the posterior inferior-parietal cortex, the temporo-parietal junction, the posterior-superior temporal and the frontal premotor cortices, subcortical gray nuclei and white matter fiber tracts connecting these regions. Neglect is independent of, and cannot be traced back to, sensory and motor deficits, and may be interpreted as a derangement of multiple networks, primarily based in the right hemisphere, supporting spatial attention and representation of space and objects in it.
... When drawing from a target image or object, eye movements tend to be saccadic between fixation clusters (Reina & Schwartz, 2003) as the eyes tend to saccade to a target and stay there until the hand reaches the same point (Gowen & Miall, 2006;Reina & Schwartz, 2003). A recent study on sketching images from memory was used to compare perceptual-motor interactions on shorter and longer timescales (Huette, Kello, Rhodes, & Spivey, 2013). Eye and pen trajectories were found to be coordinated in time on shorter timescales during free drawing conditions, yet were associated with longer timescales where subjects drew from memory, demonstrating that the task of drawing from memory evokes perceptual-motor encodings of visual images that preserve spatial information over relatively long timescales. ...
Article
Handwriting, a complex motor process involves the coordination of both the upper limb and visual system. The gaze behavior that occurs during the handwriting process is an area that has been little studied. This study investigated the eye-movements of adults during writing and reading tasks. Eye and handwriting movements were recorded for six different words over three different tasks. The results compared reading and handwriting the same words, a between condition comparison and a comparison between the two handwriting tasks. Compared to reading, participants produced more fixations during handwriting tasks and the average fixation durations were longer. When reading fixations were found to be mostly around the center of word, whereas fixations when writing appear to be made for each letter in a written word and were located around the base of letters and flowed in a left to right direction. Between the two writing tasks more fixations were made when words were written individually compared to within sentences, yet fixation durations were no different. Correlation of the number of fixations made to kinematic variables revealed that horizontal size and road length held a strong correlation with the number of fixations made by participants. Copyright © 2015 Elsevier B.V. All rights reserved.
Chapter
This article discusses the conventional function of drawing as a tool to generate ideas and solve problems in Western art and in art and design higher education. The problem of design ideation is also examined regarding Generation Z fashion students and their seeming inability to express ideas through drawing. In our view, this is not because they don’t have graphic skills but because they exclude from consciousness much of the clothes details, a circumstance that is inversely proportional to their access to fashion images via digital devices as well as clothes via fast fashion stores. Our aim is not to discredit fashion education’s orthodox approach altogether, but to contribute to its redesign toward a more hands-on methodology. Thus, we present six exercises that use drawing in combination with other exploratory techniques, as their effectiveness proved their right to become part of the curriculum of our BA in Fashion Design.
Thesis
Full-text available
Drawing accuracy has been extensively studied in children, but very little is known of what would make some adults more accurate in copying objects or scenes than many others. One factor may simply be practice: artists have often spent thousands of hours making drawings. The focus of this thesis has been to explore how this intensive practice has affected visual and memory processes. In a series of studies, we first demonstrated that drawing expertise does not relate to a more veridical perception of the world: professional artists and art students were no better than novices at seeing scenes accurately – at undoing the automatic perceptual mechanisms that ordinarily correct for visual context like shadows and depth (visual constancies). This suggests that intensive training in drawing may not affect already well-established perceptual mechanisms, but might affect higher-order processes such as visual analysis of object structure. In a number of studies, we next investigated how trained draftspersons visually encode and integrate structural information when analyzing an object. First, to test whether artists had a more advanced ability to represent complex shapes, we designed a gaze-contingent moving window task where participants had to classify an object as structurally possible or impossible, while only being able to see a portion of the object centered on the gaze position. Experts were able to perform this task with smaller samples of the object. This result suggests that skill in drawing relates to the ability to integrate local samples from each fixation into a more robust internal representation. We then asked whether drawing accuracy could also be related to the encoding efficiency of structural information from a single fixation (no eye movements allowed), with the test object centered at fixation or located in peripheral vision. In this case, we found that experts could discriminate possible vs impossible objects with shorter presentation durations and this was true whether the object was presented at fixation or in the periphery. Finally, we investigated the role of visual memory during the drawing process and whether more skilled participants have a better representation of feature locations. To do so, we designed an interactive pen tablet experiment coupled with a change detection task where participants had to copy a figure on a pen tablet. Throughout the copying process, changes could occur in both the original figure and the copy and participants had to correct any changes they noticed (the figure and the drawing were visible in alternation). We found that all participants detected changes better when they occurred in the original than in their own drawing. Moreover, experts were better at detecting changes, but only when drawing was involved (contrasted with a simple change detection task without drawing). Taken together these results demonstrate that intensive training in drawing affects higher-order perceptual and visual memory mechanisms but not basic perceptual mechanisms that already well-grounded on the life-long perceptual experiences that we all share.
Chapter
Visually Situated Language Comprehension has been compiled as a state-of the-art introduction to real-time language processing in visually-situated contexts. It covers the history of this emergent field, explains key methodological developments and discusses the insights these methods have enabled into how language processing interacts with our knowledge and perception of the immediate environment. Scientists interested in how language users integrate what they know with their perception of objects and events will find the book a rewarding read. The book further covers lexical, sentence, and discourse level processes, as well as active visual context effects in both non-interactive and interactive tasks and thus present a well-balanced view of the field. It is aimed at experienced researchers and students alike in the hopes of attracting new talent to the field. Thanks to its in-depth methodological introduction and broad coverage it constitutes an excellent course book.
Article
Full-text available
The very limited capacity of short-term or working memory is one of the most prominent features of human cognition. Most studies have stressed delimiting the upper bounds of this memory in memorization tasks rather than the performance of everyday tasks. We designed a series of experiments to test the use of short-term memory in the course of a natural hand-eye task where subjects have the freedom to choose their own task parameters. In this case subjects choose not to operate at the maximum capacity of short-term memory but instead seek to minimize its use. In particular, reducing the instantaneous memory required to perform the task can be done by serializing the task with eye movements. These eye movements allow subjects to postpone the gathering of task-relevant information until just before it is required. The reluctance to use short-term memory can be explained if such memory is expensive to use with respect to the cost of the serializing strategy.
Chapter
Papers from the 2006 flagship meeting on neural computation, with contributions from physicists, neuroscientists, mathematicians, statisticians, and computer scientists. The annual Neural Information Processing Systems (NIPS) conference is the flagship meeting on neural computation and machine learning. It draws a diverse group of attendees—physicists, neuroscientists, mathematicians, statisticians, and computer scientists—interested in theoretical and applied aspects of modeling, simulating, and building neural-like or intelligent systems. The presentations are interdisciplinary, with contributions in algorithms, learning theory, cognitive science, neuroscience, brain imaging, vision, speech and signal processing, reinforcement learning, and applications. Only twenty-five percent of the papers submitted are accepted for presentation at NIPS, so the quality is exceptionally high. This volume contains the papers presented at the December 2006 meeting, held in Vancouver. Bradford Books imprint
Article
The cooperative action of different regions of human brains gives an amazing capacity to perform activities as diverse as playing the piano and hitting a tennis ball. Somehow, without conscious effort, eyes find the information that is needed to operate successfully in this world. The development of head-mounted eye trackers over recent years has made it possible to record where we look during different active tasks, and so work out what information the eyes supply to the brain systems that control limbs. The strategies that the eye movement system uses in the initiation and guidance of action are explored. This book examines a wide range of visually guided behaviour, from sedentary tasks like reading and drawing, to dynamic activities such as driving and playing cricket. A central theme is that the eye movement system has its own knowledge about where to find the most appropriate information for guiding action - information not usually available to conscious scrutiny. Thus, each type of action has its own specific repertoire of linked eye movements, acquired in parallel with the motor skills themselves. Starting with a brief background to eye movement studies, the book then reviews a range of observations and analyses of different activities. It ends with discussions of the nature of visual representation, the neurophysiology of the systems involved, and the roles of attention and learning.
Article
The ubiquity of apparently random behavior in visual search (e.g., Horowitz & Wolfe, 1998) has led to our proposal that the human oculomotor system has subtle deterministic properties that underlie its complex behavior. We report the results of one subject's performance in a challenging search task in which 10,215 fixations were accumulated. A number of statistical and spectral tests revealed both fractal and 1/f structure. First, scaling properties emerged in differences across eye positions and their relative dispersion (SD/M)—both decreasing over time. Fractal microstructure also emerged in an iterated function systems test and delay plot. Power spectra obtained from the Fourier analysis of fixations produced brown (1/f^2) noise and the spectra of differences across eye positions showed 1/f (pink) noise. Thus, while the sequence of absolute eye positions resembles a random walk, the differences in fixations reflect a longer-term dynamic of 1/f pink noise. These results suggest that memory across eye-movements may serve to facilitate our ability to select out useful information from the environment. The 1/f patterns in relative eye positions together with models of complex systems (e.g., Bak, Tang & Wiesenfeld, 1987) suggest that our oculomotor system may produce a complex and self-organizing search pattern providing maximum coverage with minimal effort.
Article
This paper presents the case for a functional account of vision. A variety of studies have consistently revealed "change blindness" or insensitivity to changes in the visual scene during an eye movement. These studies indicate that only a small part of the information in the scene is represented in the brain from moment to moment. It is still unclear, however, exactly what is included in visual representations. This paper reviews experiments using an extended visuo-motor task, showing that display changes affect performance differently depending on the observer's place in the task. These effects are revealed by increases in fixation duration following a change. Different task-dependent increases suggest that the visual system represents only the information that is necessary for the immediate visual task. This allows a principled exploration of the stimulus properties that are included in the internal visual representation. The task specificity also has a more general implication that vision should be conceptualized as an active process executing special purpose "routines" that compute only the currently necessary information. Evidence for this view and its implications for visual representations are discussed. Comparison of the change blindness phenomenon and fixation durations shows that conscious report does not reveal the extent of the representations computed by the routines.
Article
NYSTAGMOID EYE MOVEMENT WAS MEASURED IN 52 MALE AND FEMALE SS APPROXIMATELY 22 YR. OLD DURING VISUAL IMAGERY OF CONTINUOUS UNIDIRECTIONALLY MOVING CONTRASTS (BLACK AND WHITE STRIPS) AS A SPECIFIC FEATURE ACCOMPANYING THIS MENTAL PROCESS. A SIGNIFICANT RELATIONSHIP WAS FOUND BETWEEN SUBJECTIVELY REPORTED IMPROVEMENTS IN VISUAL IMAGINATION AND INCREASED INCIDENCE OF NYSTAGMOID EYE MOVEMENTS DURING SIGNALS TO IMAGINE A MOVING STIMULUS PATTERN, AS USED FOLLOWING ADMINISTRATION OF 20 MG. OF DEXPHENMETRAZIN AS AGAINST A PLACEBO. VISUAL IMAGERY OF STIMULUS FOR THE ELABORATION OF CONDITIONED OPTOKINETIC NYSTAGMUS EMBODIED INTO INTERTRIAL INTERVALS ENHANCES THE CONDITIONED REFLEX RESPONSE FOLLOWING IMAGERY. (PsycINFO Database Record (c) 2012 APA, all rights reserved)
Article
A summary of 50 years' work on the biology and behavior of honeybees. Liberally illustrated. Harvard Book List (edited) 1971 #215 (PsycINFO Database Record (c) 2012 APA, all rights reserved)