ArticlePDF Available

Reconstruction of shape contours from V1 activity at high resolution

Authors:

Abstract and Figures

The role of primary visual cortex (V1) in encoding physical stimulus features is well known, while stimulus categorization is mainly attributed to higher visual areas. However, visual experience is not stripped down to invariant, categorical-only "labels." Rather, visual experiences are remarkably rich with details resulting in high-resolution perception of objects. If V1 is involved in this process, high-resolution readout of shape contours should be possible from V1 activity. To test this, we presented various shapes to awake, fixating monkeys while recording V1 population activity using voltage-sensitive dye imaging. A simplified bottom-up model was constructed based on known cortical properties and without an image prior. Contours were reconstructed from single trials, in sub-degree resolution by applying the inverse model to neuronal responses. These novel reconstruction results suggest V1 can be an important constituent in the detailed internal representation of visual experiences.
Typical stimuli, V1 VSD responses and forward model predictions. (a) Visual stimuli: (i) Circular ring, diameter (d) = 1.6°, contrast (C) = 38%, centered positioned at (P) = −1°,– 2.5° (left and below, respectively) relative to the fixation point (fp, small red circle). (ii) 1.0 × 1.0° square, C = 100%, P =-1°–2.25°. (iii) Circular ring, d = 0.8°, C = 38%, P =-1.2°–2.5°. (iv) Circular ring, d = 1.6°, C = 38%, P =-1°–2°. v 2.0 × 2.0° black square outline, C =-100%, P =-1.5°–2.5°. (b) Evoked responses corresponding a(i–vi), calculated by averaging VSD maps 150–200 ms after stimulus onset. Imaging window outlined in black. Upper border of imaged region marks the vertical meridian (see Materials and Methods). Major blood vessels marked in gray. Red arrows mark weakly activated regions corresponding to parts of the stimulus contour. Responses i, iii are from monkey S. Responses shown in ii, iv, and v are from monkey T. (c) Forward model illustration (right to left): (i) example stimulus presented atop gray screen at lower left quadrant. fp illustrated in red. (ii) Stimulus convolved with contrast PSF (see Materials and Methods). (iii) Convolved stimulus (b) after non-linearity (see Materials and Methods). Iso-eccentricity (concentric) and iso-polarity (straight) lines in blue. (iv) Projection of encoded image onto V1 surface following retinotopic transformation (imaging window outline illustrated in black). Inset: Imaging window illustrated on a schematic macaque brain (not to scale). (d) Response patterns predicted by the forward model for stimuli in (a). (e) Encoding performance: Pearson correlation coefficient (r) between observed and predicted patterns for each stimulus type (n = 27 sessions): circles (d = 0.6–1.6°), squares (1° × 1°–3 × 3°), and a two co-aligned bars (0.25° each, spaced 1° apart).
… 
Content may be subject to copyright.
Reconstruction of shape contours from V1 activity at high resolution
Guy Zurawel, Itay Shamir, Hamutal Slovin
The Gonda Multidisciplinary Brain Research Center, Bar-Ilan University, Ramat Gan, Israel
abstractarticle info
Article history:
Received 30 June 2015
Accepted 24 October 2015
Available online 27 October 2015
Keywords:
Monkey
Primary visual cortex
Image reconstruction
High-resolution visual perception
Decoding
Imaging
The role of primary visual cortex (V1) in encoding physical stimulus features is well known, while stimulus cat-
egorization is mainly attributed to highervisual areas. However,visual experience is not stripped down to invari-
ant, categorical-only labels.Rather, visual experiences are remarkably rich with details resulting in high-
resolution perception of objects. If V1 is involved in this process, high-resolution readout of shape contours
should be possible from V1 activity. To test this, we presented various shapes to awake, xating monkeys
while recording V1 population activity using voltage-sensitive dye imaging. A simplied bottom-up model was
constructed based on known cortical properties and without an image prior. Contours were reconstructed
from single trials, in sub-degree resolution by applying the inverse model toneuronal responses. These novel re-
construction results suggest V1 can be an important constituent in the detailed internal representation of visual
experiences.
© 2015 Elsevier Inc. All rights reserved.
Introduction
Visual experiences are remarkably detailed and quickly processed
under variable conditions. The role of primary visual cortex (V1) in
encoding stimulus attributes of visual input is crucial to this capability
and consists of various processing mechanisms: retinotopy (Tootell
et al., 1988), receptive elds (RFs, Hubel and Wiesel, 1968), contrast
coding (DeValois and DeValois, 1988), neuronal non-linearity
(Albrecht, 1995), and several others (orientations, color, surfaces,
etc.). In contrast, high-level processing of complex features like invari-
ant categorization and face processing are associated with RFs and stim-
ulus attributes' convergence to and processing by higher visual areas
(Freiwald and Tsao, 2010; Hung et al., 2005). However, visual experi-
ence is not stripped down to invariant, categorical-only labels.Rather,
detailed visual information is maintained, resulting in a detailed, high-
resolution perception of visualobjects.The presence of ne details of vi-
sual stimuli (those not fully captured by categorical labels) in V1 could
contribute to detailed visual perception (Gur, 2015; Hochstein and
Ahissar, 2002; Wokke et al., 2013), potentially via readout by higher vi-
sual areas. Consequently, visual features such as shape contours should
be reconstructible (decoded) from V1 neuronal activity.
Decoding visual contents from neuronal activity has been explored
recently using various approaches. Using multivariate classication
and other statistical tools, stimulus category or orientation can be
decoded among several possibilities from cortical activity (Haynes and
Rees, 2005; Hung et al., 2005). Supplementing the observed activation
with a basis of visual elements wasshown to improve decoding results,
facilitating reconstruction of edge-dened visual stimuli (Miyawaki
et al., 2008). Use of a prior bank of images facilitated the identication
of an image that resembled the original one, visually (Kay et al., 2008)
as well as semantically (Horikawa et al., 2013; Naselaris et al., 2009).
Despite the great advances, reconstructing visual stimuli in high de-
tail directly from cortical activity without an image prior has yet to be
accomplished. However, several important steps have already been
made in this direction; Gollisch and Meister (2008) reconstructed the
contours of natural images using retinal spike latencies, and Stanley
et al. (1999) reconstructed features of natural image sequences from
cat LGN spiking activity. Finally, Thirion et al. (2006) were able to de-
code pattern identity and the position of pattern elements based on
fMRI activity and retinotopy.
Here, we set out to reconstruct high-resolution contours from V1 ac-
tivity without prior knowledge regardingshape identity or position. We
presented achromatic shapes to awake, xating monkeys, while imag-
ing V1 activity using voltage-sensitive dye imaging (VSDI). Contours
were reconstructed by applying the inverse of a simplied forward
model to observed responses.
Materials and methods
V1 population responses were recorded using voltage-sensitive dye
imaging (VSDI) from 3 hemispheres of 3 awake, xating macaque mon-
keys, a total of 27 imaging sessions. All experimental procedures were
approved by the Animal Care and Use Guidelines Committee of Bar-
Ilan University, supervised by the Israeli authorities for animal experi-
ments and conformed to NIH guidelines.
NeuroImage 125 (2016) 10051012
Corresponding author at: Gonda Multidisciplinary Brain Research Center, Bar-Ilan
University, 52900, Ramat Gan, Israel.
E-mail addresses: Guy.zur@gmail.com (G. Zurawel), Itay.shamir@live.biu.ac.il
(I. Shamir), Hamutal.Slovin@biu.ac.il (H. Slovin).
http://dx.doi.org/10.1016/j.neuroimage.2015.10.072
1053-8119/© 2015 Elsevier Inc. All rights reserved.
Contents lists available at ScienceDirect
NeuroImage
journal homepage: www.elsevier.com/locate/ynimg
Animals
Three adult male monkeys (Macaca fascicularis, 12, 13, 8 kg) were
used for the current study. The surgical procedure has been reported
in detail elsewhere and is outlined brieybelow.
Surgical procedures and voltage-sensitive dye imaging
All methods were approved by the Animal Care and Use Guidelines
Committee of Bar-Ilan University, supervised by the Israeli authorities
for animal experiments and conformed to the NIH guidelines. The surgi-
cal, staining and VSD imaging procedure have been reported in detail
elsewhere (Slovin et al., 2002; Zurawel et al., 2014).
Behavioral paradigm
The monkeys were trained on a simple xation task. On each trial,
after a short xation period (34 s), monkeys were briey(200
300 ms) presented with a single visual stimulus as described below.
The monkeys were required to maintain tight xation throughout the
whole trial and were rewarded with a drop of juice for each correct
trial. Stimulated trials were interleaved with blank trials, in which the
monkeys xated but no visual stimulus appeared.
Experimental setup
The experimental setup has been described in detail elsewhere
(Ayzenshtat et al., 2010; Zurawel et al., 2014).
Visual stimuli
Visual stimuli consisted of achromatic rings (diameter 0.6°1.6°),
squares and square surfaces (1.0°3.0°) or bars (2 co-aligned bars,
0.25° each, spaced 1° apart). Stimulus contrast ranged 38%100%
(Weber), iso-luminant background luminance 17, 37 or 74 cd*mm
-2
,
and eccentricity ranged 1.3.1° (see Table S1). Contour width was
0.03°0.09°. In two sessions, 3° square stimuli were presented on the
left hemi-eld and positioned such that the right side crossed the verti-
cal meridian to the right hemi-eld (e.g., Fig. 3a 9th row).All other stim-
uli were presented entirely in a single hemi-eld.
Eye movements
Eye positions were monitored by a monocular infrared eye tracker
(Dr. Bouis, Karlsruhe, Germany), sampled at 1 kHz and recorded at
250 Hz. Only trials where the animals maintained a tight xation were
analyzed.
Data analysis
Typically, we analyzed 1530 correct trials for each visual stimulus in
a recording session (see Table S1). Matlab software (version 2011b, The
MathWorks, Inc.) was used for statistical analyses and calculations. The
basic VSDI analysis consisted of (a) dening region-of-interest (only
pixels with uorescence level 15% of maximal uorescence were ana-
lyzed) ; (b) normalizing to background uorescence; (c) average blank
subtraction for removal artifacts (e.g., heartbeat, see more details and
schematic illustration of the basic VSDI analysis in Supplementary
Fig. S12 of Ayzenshtat et al., 2010); and (d) removal of pixels located
on blood vessels. VSDI responses were then averaged over 150200 ms.
This time range was used to avoid the effect of fast eye movements, i.e.,
saccades and micro-saccades (saccade and micro-saccade suppression
occurs up to 150200 ms after stimulus onset; Engbert and Kliegl, 2003).
Motion artifacts were avoided, mostly due to the animals' training to
calmly maintain xation while xed to the head-post. Further screening
for very large saccades or irregular ECG activity, followed by
examination for irregular noise patterns associated with a motion arti-
fact eliminated introduction of noise due to motion.
Retinotopic mapping of V1
Retinotopic mapping of V1 and the V1/V2 border was obtained in a
separate set of imaging sessions using VSD and optical imagingof intrin-
sic signals and has been described elsewhere (Ayzenshtat et al., 2012).
Encoding/decoding model
A simplied bottom-up forward model was constructed based on
known cortical mechanisms: (1) population RFs (the collective RF of
all neurons imaged in a single pixel) and absolute contrast coding as
captured by cortical point spread, (2) non-linear spiking responses,
and (3) retinotopic mapping. Three corresponding model components
were constructed as detailed below. In the forward direction
(encoding), model components were applied to the stimulus in the
above sequence. To reconstruct stimuli (reverse direction), the inverse
form of each component was used, and components were applied in re-
verse order to cortical data.
To arrive at a reversible model with a limited parameter set, the
model was simplied extensively,accountingonly for major V1 mecha-
nisms necessary to reconstruct high-contrast stimuli (detailed above).
Other known coding mechanisms such as orientations, spatial frequen-
cies, phase, and the weaker surface coding (Zurawel et al., 2014) were
left out to maintain reversibility and minimum assumptions. Introduc-
ing these mechanisms would likely result in improved encoding perfor-
mance and enable reconstructions of more elaborate stimuli. However,
such a complex model would entail difcult parameter acquisition and
validation, requiring a large parameterizing data set. Importantly,
inverting such a model would require more assumptions thus making
it less naïve. We therefore restricted stimuli to high contrasts and a nar-
row band of frequencies and used the PSF to capture typical contrast re-
sponses, while neglecting contributions and other stimulus dimensions
not addressed in the model. Therefore, prior to encoding, the absolute,
non-normalized contrast of stimuli was taken. As a result, negative con-
trast (black), positive contrast and surface stimuli were all represented
as contour-only versions thereof (See Fig. S1b). Model building blocks
are detailed below.
Estimation of the point-spread function (PSF)
Presentation of a small contrast stimulus (point) elicits a large
spread of V1 activity (Van Essen et al., 1984) that results from (1) RFs
mapping: each contrast-responsive neuron whose RF falls within the
stimulus' position responds to some degree, and there are overlapping
RFs in V1. (2) The point-spread VSD activation reects neuronal popula-
tions' membrane potential, encompassing both supra- and sub-threshold
response. The model is simplied by representing the contribution of
contrast responses for all orientations and their population RF in a single
PSF.
PSF was obtained empirically as the average (150200 ms post stim-
ulus) VSD pattern evoked by a high-contrast point stimulus (0.1° × 0.1°
white square, see point-spread time course in Fig. S4). Patternswere ex-
pectedlyanisotropic (average anisotropy 1.6) and extended over sever-
al mm of cortex (Grinvald et al., 1994). This is expected from a PSF with
a sub-threshold component, which typically extends beyond the spik-
ing PSF (Das and Gilbert, 1995; Sincich and Blasdel, 2001; Xing et al.,
2009). The transition to spiking was expressed in the non-linear part
of the model (see Non-linearity section). In cases the observed PSF
was noisy or not available, a 2D well-t exponential model was used
(br
2
N=0.66, see Fig. S1a):
fR;θðÞ¼ae bRgθðÞðÞ ð1Þ
1006 G. Zurawel et al. / NeuroImage 125 (2016) 10051012
gθðÞ¼cþ1c
2

ðcos 2 θθ0
ðÞðÞþ1ðÞ ð2Þ
where Rand θare polar coordinates, ais a scaling coefcient, bis the
spatial decay rate at the widest direction (θ=θ
0
), and gis the anisotro-
py function, modulating bas a function of angle, wherein cdenes the
anisotropy ratio and θ
0
the orientation of the spread. We used an expo-
nential model rather than a Gaussian (same number of parameters) to
achieve a better spatial t, specically at outer regions of the patterns.
The PSF pattern was then transformed to the VF (see Retinotopic
transformation section), and diameter kept within 1° in accordance
with known summation eld sizes at the imaged eccentricity
(Angelucci et al., 2002). Finally, the PSF pattern was normalized to a
sum of 1 (PSF, Fig. S1a).
PSF convolution and de-convolution
The PSF was convolved with the stimulus (encoding direction) or
de-convolved to arrive at the stimulus (decoding model, reconstruc-
tion). De-convolution was performed using RichardsonLucy (RL) de-
convolution (Lucy, 1974; Richardson, 1972), an iterative procedure
that converges to the maximum-likelihood solution of a source image
given a known PSF. RL operates under the single assumption of
Poissonian noise and receives only a blurred image and a known PSF
as input, with no other parameters or assumptions regarding the source
image.
Non-linearity
Non-linearity is a well-known property of cortical neurons
(e.g., Albrecht, 1995) . In the context of VSDI, non-linearity may repre-
sent the thresholdingtransition between membrane potential
(reected in VSD signal) to spiking activity. We used the sigmoid func-
tion in the inverse (decoding) model:
SxðÞ¼ 1
1þeβxx0
ðÞ ð3Þ
where xis the input value, βis the slope, and x
0
is the value at which
S(x) = 0.5. Parameters were dependent on imaged pixel distributions:
X
0
was set as a percentile (65th90th), βwas inversely proportional
to a percentile (65th87th) of the distribution, scaled by a xed factor:
β¼13
percentile xðÞ ð4Þ
In the forward (encoding) model, the inverse form was used:
S1xðÞ¼ log xðÞlog 1xðÞ
βð5Þ
X
0
and βpercentiles were t per-session based on session noise levels,
and the constant scale factor was determined using data and recon-
struction from a single session and maintained as a constant for all data.
Retinotopic transformation from the visual eld to the cortical surface
Retinotopic mapping from cortical surface to the VF (bidirectional)
was performed using a log-polar model (Schira et al., 2010), as de-
scribed previously (Zurawel et al., 2014). Parameters were determined
for each animal using independent control points.
Reconstruction and discrimination performance evaluation
To evaluate reconstruction performance, we calculated pixel-wise
Pearson correlation coefcients (r) for each reconstructionstimulus
pair. To overcome edge-width discrepancies (see discussion), stimulus
representations were taken as a 9-pixel (0.25°) wide mask of the
stimulus contour. To allow for small retinotopic inaccuracies (arising
from small variance in animal's gaze position, retinotopic model t,
etc.), each correlation was performed with limited transformation free-
dom (±0.5° translation, ±15° rotation, ±15% scaling, not exceeding
the next smaller/larger stimulus). All correlations computed on a
3° × 5° region of the VF that covered the entire portion of the VF corre-
sponding to the imaged area. For some sessions, background noise near
the boundaries of the imaging region created a visible artifact of contrast
(e.g., Figs. 2c v). These boundary regionswere excluded from the single-
trial correlation calculation.
Discrimination among stimuli using reconstructions was per-
formed, for each reconstruction, as follows: First, rvalues (described
above) were computed against all candidate stimuli. For averages, d
prime (d) was computed between the correctrvalue (correlation
with the correct stimulus) and the distribution of remaining rvalues
(correlation with all other stimuli). For single trials, ROC (receiver oper-
ating characteristics) was computed between the distribution of single-
trial correctrvalues and the distribution of all other single-trial r
values (obtained for all other stimuli), followed by area under the
curve (AUC) calculation. To eliminate variation due to VF position, the
analysis was repeated with all stimuli co-located with (translated to)
the same VF position.
Results
Our goal was to evaluate V1 as the neural substrate for high-resolu-
tion visual details (e.g., contours). This was done by reconstructing
shape contours along with their precise visual eld (VF) position, in
high-resolution and without the use of prior image or shape examples.
Three xating monkeys were presented with various circle, square, or
bar stimuli for 200300 ms. V1 population responses were recorded
using voltage-sensitive dye imaging (VSDI, total of 27 imaging sessions
of all stimulus types, 3 hemispheres), reecting supra and sub-
threshold membrane-potentials from the underlying neuronal ele-
ments. To reconstruct stimuli, neuronal data was fed into the inverse
of a simplied bottom-up model, constructed from well-known V1 pro-
cessing mechanisms: receptive elds (RFs) and cortical response to
local spatial contrast that elicit a cortical point-spread function (PSF),
non-linearity and retinotopy.
Population responses to achromatic shapes
Monkeys were presented with 10 different stimuli in various VF lo-
cations (17 stimulus/location combinations total). VSD responses were
averaged 150200 ms (see Fig. S4) after stimulus onset, capturing late
cortical response that is more closely linked to perception (Gilad et al.,
2013; Lamme, 1995) and prior to onset of eye movements (Engbert
and Kliegl, 2003). Responses evoked by large shapes (N1.5°) showed
some resemblance to the stimulus: circle stimuli (Fig. 1a i, iv) elicited el-
liptical, annular patterns over theV1 imaged area (Fig. 1bi,ivtwodiffer-
ent animals), and similarly squares (Fig. 1a v) a rectangular pattern
(Fig. 1b v). Cortical patterns were skewed and blurred relative to the
stimuli, which can be explained by retinotopic mapping, cortical spread
and magnication in V1 (Tootell et al., 1988). Some stimuli's parts fell
very close to the Vertical Meridian (VM also the V1/V2 border on the
cortical surface). This occasionally resulted in weaker activation of the
corresponding V1 regions (Figs. 1abiv,ab v, red arrows; see
Discussion). Consequently, these stimulus parts would likely be more
difcult to reconstruct (see Results).
Responses to smaller stimuli showed less resemblance to the origi-
nal stimulus, possibly as a result of the cortical point-spread and the im-
aging resolution: smaller circles (0.8°, Fig. 1aiii)evokedalled
(Fig. 1b iii) pattern, and responses to smaller squares (Fig. 1aii)lacked
the clear sharp corners (Fig. 1b ii). Surface lledstimuli also yielded
contour-like activations, accompanied by a weaker center (surface) re-
lated activation (Fig. S1b), a phenomenon we addressed recently
1007G. Zurawel et al. / NeuroImage 125 (2016) 10051012
(Zurawel et al., 2014). For simplicity, here we focus on reconstructing
stimulus contours and neglect the surface-related contribution.
These responses hint to possible size and resolution bounds of a re-
construction procedure relying solely on V1 population activity imaged
with current spatial resolution (see Discussion). We next constructed a
simplied forward model that could explain the data.
Stimulus encodingforward model
A forward model (illustrated in Fig. 1c) that accounted for local con-
trast coding and RFs, non-linearity, and retinotopy was applied to stim-
uli, via the following steps, performed in sequence: Convolution of the
original stimulus (Fig. 1c i) with a contrast-evoked PSF (see Materials
Fig. 1. Typical stimuli, V1VSD responses and forward model predictions.(a) Visual stimuli: (i) Circular ring, diameter (d) = 1.6°, contrast (C) = 38%, centered positioned at (P) = 1°,
2.5° (leftand below, respectively) relative to thexation point (fp,small red circle). (ii) 1.0 × 1.0° square, C = 100%, P = -1°2.25°. (iii)Circular ring, d= 0.8°, C = 38%, P = -1.2°2.5°. (iv)
Circular ring, d= 1.6°, C = 38%, P = -1°2°. v 2.0 × 2.0° black square outline, C = -100%, P = -1.5°2.5°. (b) Evoked responses corresponding a(ivi), calculated by averaging VSD maps
150200ms after stimulus onset.Imaging window outlined in black.Upper border of imaged region marks the vertical meridian(see Materials andMethods). Major blood vessels marked
in gray. Red arrows mark weaklyactivated regions corresponding to parts of the stimulus contour. Responsesi, iii are from monkey S. Responses shown in ii,iv, and v are from monkey T.
(c) Forward model illustration (right to left): (i) example stimulus presented atop gray screen at lower left quadrant. fp illustrated in red. (ii) Stimulus convolved with contrast PSF (see
Materials and Methods). (iii) Convolved stimulus (b) after non-linearity (see Materials and Methods). Iso-eccentricity (concentric) and iso-polarity (straight) lines in blue.(iv) Projection
of encoded image onto V1 surface following retinotopic transformation (imaging window outline illustrated in black). Inset: Imaging window illustrated on a schematic macaque brain
(not to scale). (d) Response patterns predicted by the forward model for stimuli in (a). (e) Encoding performance: Pearson correlation coefcient (r) between observed and predicted
patterns for each stimulus type (n= 27 sessions): circles (d=0.61.6°), squares (1° × 1°3 × 3°), and a two co-aligned bars (0.25° each, spaced 1° apart).
1008 G. Zurawel et al. / NeuroImage 125 (2016) 10051012
and Methods,Eqs.(1) and (2), Fig. S1a) capturing the contrast RFs con-
tribution of neuronal populations in each pixel (Fig. 1c ii), a pixel-wise
non-linearity (Eqs. (4) and (5),Fig. 1c iii) and a retinotopic transforma-
tion from the cortical surface to visual space (Fig. 1civ,seeMaterials and
Methods and Schira et al., 2010). For simplication and reversibility
purposes, additional stimulus features (e.g., orientation, frequency and
surface processing) were not explicitly addressed in the model (see
Materials and Methods).
To validate the forward model, each stimulus (dened by shape, size
and VF position, see Table S1) was encoded, and resulting patterns
(Fig. 1d shows several examples) were compared to corresponding ob-
served patterns (Fig. 1b). Similarity in shape and cortical position was
evident. Moreover, some unique, size-dependent features were also
reproduced: (1) elliptical patterns for large circles (Fig. 1di,ivcompare
with observed patterns in Fig. 1b i, iv) in contrast to lledpatterns for
smaller circles (Figs. 1d iii, b iii); (2) apparent square-like patterns for
Fig. 2. (a) Reconstruction model (inverse model) illustration:(i) Example VSD response evoked by 1.6° ring, large blood vesselsmarked in pink. Inset: Imaging window illustrated on a
schematic macaque brain (not to scale). (ii) Computed retinotopic projection of response pattern (i) onto the VF (see Materials and Methods). (iii) Image from (ii) after applying non-
linearity (see Materials and Methods). (iv) Image from (iii) following contrast PSF de-convolution. (b, c) Reconstructions from average responses. (b, iv) Example stimuli (same as
Fig. 1a). Insets (above arrows): average evoked responses. Weakly activated regions marked with red arrows. (c) Reconstructions corresponding to b, obtained by applyi ng the inver se
model on average responses (b insets). Original stimuli outlined in dashed gray. An artifact arising from background noise near edges of the imaging region is visible in v (sporadic acti-
vations leftof and above the squarecontour), see Materials and Methods for more detail. (d, i) Reconstructionstimulus correlations (r) for all stimuli (dened by shape, size and VF po-
sition, see Table S1). (ii)Average correlation over time. Each value represents mean calculated for 9 time points.Data averaged over 20 ms. Shading denotes ±1 SEM (n= 27 sessions).
1009G. Zurawel et al. / NeuroImage 125 (2016) 10051012
large square stimuli (Figs. 1d v, b v) compared to smaller squares
(Figs. 1d ii, b ii); and (3) reduced activations for portions of the pattern
that were near or crossed the vertical meridian(Figs. 1bivv, d ivv, red
arrows).
Predicted and observed patterns were similar for all stimuli, indicat-
ed by Pearson correlation coefcient (r) values (Fig. 1e, r= 0.62 ± 0.02
mean ± SEM, n= 27 sessions, pb0.001). Similarity to single-trial re-
sponses was expectedly lower, but still consistently positive (0.35 ±
0.02, n= 703 trials, pb0.001). Overall, qualitative and quantitative ex-
amination of the encoded patterns attested to the validity of the model.
We next turned to reconstruct the presented shapes using the model.
Stimulus contour reconstruction
To reconstruct stimulus contours, the inverse form of the model was
applied to observed responses, in reverse order. First, measured VSDI
patterns (Fig. 2a i) were retinotopically transformed onto visual space
(Fig. 2a ii, see Materials and Methods). Next, non-linearity was applied
(see Materials and Methods and Eq. (3)) resulting in a thresholded
image (Fig. 2a iii), possibly isolating the supra-threshold component
of the VSD signal. Finally, the contrast PSF was de-convolved from the
image using the RichardsonLucy (RL) de-convolution (see Materials
and Methods), an iterative maximum-likelihood based algorithm
(Fig. 2a iv and inset). Input to the RL was the blurred (thresholded)
image from the previous step and the contrast PSF, with no additional
prior knowledge regarding stimulus shape, size or position.
Reconstruction successfully reproduced the contour characteristics
of presented stimuli; large circle reconstructions (Fig. 2c i,iv) resembled
the outline, size and VF position of original stimuli (dashed circles and
Fig. 2b i,iv). Similarly, reconstructed large squares (Fig. 2c v) matched
stimuli contours in apparent shape, size, and position (Fig. 2b v). Recon-
structions of smaller stimuli (Fig. 2b ii, iii) were accurately sized and po-
sitioned, but with contours that showed deviations from the original
(rounded corners in Fig. 2c ii, near-lled circle in Fig. 2ciii).Expectedly,
weak activations corresponding to stimulus regions that were near or
crossed the vertical meridian (described above, red arrows in Fig. 2c
iv, v insets, Fig. 1b iv, v) resulted in incomplete contours (Fig. 2civ, v).
The apparent size- and position-dependent reconstruction quality
bounds may originate from limitations imposed by corresponding
evoked patterns, discussed above. Additionally, reconstructed edges ap-
peared thicker than originals and with varying intensity, as further ad-
dressed in discussion. Pixel-wise correlations between reconstructions
and stimulus contours (r,seeMaterials and Methods) were positive
and relatively high (Fig. 2d i, 0.55 ± 0.03, mean ± SEM, n=27sessions,
pb0.0001 for all).
To examine how the choice of time window (150200 ms, Fig. S4)
affected performance, we re-calculated rat various times (Fig. 2d ii).
Correlations were low at early times, rose gradually and plateaued
(~100 ms) at a steady level.
Single-trial contour reconstruction
Averaging is efcient in overcoming noise, but natural processing
occurs without repetitions, thus reconstruction from single-trial data
is important. Most single-trial reconstructions (Fig. 3a iii, examples
from multiple sessions, different animals and VF positions in each
row; see Fig. S5 for single-trial reconstructions from the same session)
resembled the stimuli (Fig. 3a i, centered for illustration). Similar to av-
erages, contours of smaller stimuli were more difcult to reconstruct
(e.g., Fig. 3a iii rows 46 from top). Single-trial reconstructions bore sev-
eral differences from their trial-averaged counterparts (Fig. 3a ii, single-
session average examples). Some contours appeared weaker(e.g.,
Fig. 3a iii 1st column vs. averages in 3a ii). Additionally, background re-
gions contained more visible noise (e.g., trials marked with a solid black
contour) that arose from within the imaged area or its edges. In other
cases, reconstructions were slightly scaled (e.g., dashed black contour).
Finally, some had weaker or missing contour parts compared to aver-
ages (e.g., dotted black contours). Overall, single-trial reconstructions
were expectedly noisier, but in many cases, stimulus contours could
be detected. Consequently, rvalues were lower than those of averages
(Fig. 3b, 0.42 ± 0.005, n= 703) and were signicantly positive (N0)
for all stimuli (maximal pb0.0001, Wilcoxon signed rank computed
separately for each session's distribution of rvalues).
Discrimination performance
Discrimination was measured by computing rvalues for each recon-
struction against all candidate stimuli (Fig. 3d bottom, normalized for
each column for visualization. see Materials and Methods). High rbe-
tween reconstructed pattern and original stimulus (diagonal entry in
3d, bottom) indicates that stimulus identity could be decoded with
high certainty. Discrimination was quantied using dprime (d) for av-
erages (original stimulusrvs. distribution of all others) and ROC analy-
sis for single trials (distribution of original stimulusrs vs. all others, see
Materials and Methods). Average reconstructions showed high ddis-
crimination (Fig. 3c, 2.46 ± 0.57 mean ± SEM, n= 17 unique stimuli).
For single trials, ROC area under the curve measures (AUCs, ranging 1 to
0.5, perfect to chance-level) were mostly high (Fig. 3d top, 0.87 ± 0.03
mean AUC ± SEM, n= 17) and above chance (Wilcoxon signed rank,
pb0.0003). Discrimination among similar, closely positioned stimuli
(see Fig. S2a) was more difcult, especially for small-medium sized
shapes (e.g.,Fig. 3d bottom, sporadic high values). Consequently, AUCs
for these reconstructions were lower.
To isolate contributions of stimulus contour and size, we repeated
the analysis with all stimuli co-located (see Materials and Methods
and Fig. S2b). As expected, discrimination performance decreased, but
remained above chance for all stimuli (0.8 ± 0.05, pb0.002, Fig. S3).
We attribute this decrease to the previous contribution of VF position
(Table S1).
Discussion
To evaluate V1 as the neural substrate underlying representation of
ne visual details that can contribute to detailed visual perception and
be read out by higher visual areas, shape contours were reconstructed
from relatively late (150200 ms) V1 population activity. Reconstructed
shape, size, and location accuracy were in thesub-degree range and ac-
complished via a brain-inspired model, without the use of prior image
or shape examples.
Brain-inspired modeling enabled novel, high-resolution contour re-
construction that extends beyond course reconstruction of VF position
(Thirion et al., 2006). Reconstruction was accomplished by embedding
theoretical and empirical knowledge of brain processing into the
model. In contrast, other studies used a prior collection or training set
Fig. 3. Single-trialstimulus reconstruction. (a)Reconstructions.(i) All stimulus categories, dened by shape and size(see Table S1). (ii) Typicalsingle-session average reconstructionscor-
responding to (ii). (iii) Typical single-trialreconstruction examples from various imaging sessions, animals and stimulus VF position (hence the variability in reconstructedcontours and
their positions among trials). Same model parameters as usedfor the averages. Note that the3° square stimulus was positioned such thatthe right side crossed the vertical meridian (9th
row from the top). (b) Single-trial reconstructionstimulus correlations (r) averaged for each stimulus (error bars denote ±1sem, n= 703 trialstotal), dened by shape, size and VF po-
sition (seeTable S1). Same x-axisas c. (c) dfor average reconstructions.(d) Bottom: correlation matrixbetween reconstructions and stimuli. Eachcolumn holds averagervalues (single-
trial correlation) of one reconstruction(x-axis) with all stimuli (y-axis), dened asin b (shape, size, andVF position. X/Y-axisentries with same stimulus specication differ in VF position,
see Table S1). Values are normalized (r
n
) to a sum of 1 within each columnfor visualization purposes. Values along diagonal (solidgray contour) are r
n
sfor the actual (correct) stimulus.
Top: Single-trial discrimination: AUCs computed for each reconstruction (same x-axis as bottom panel) between distributions of rvalues for correctstimuli and of all other r's (for all
other stimuli).
1010 G. Zurawel et al. / NeuroImage 125 (2016) 10051012
1011G. Zurawel et al. / NeuroImage 125 (2016) 10051012
of images to reconstruct more complex stimulus features or natural
stimuli (Kay et al., 2008; Miyawaki et al., 2008) The use of high-
resolution VSDI enabled more rened reconstructions. Reconstruction
resolution (bounds) was directly affected by the imaging resolution
and per-pixel averaging but is also likely eccentricity dependent and
therefore affected by the relative location of the V1 imaging.
By constructing the model using separate encoding modules in se-
quence, representations of different encoding mechanisms were kept
separate, and model reversibility was easily maintained. Such modular
construction also enables future introduction of additional (excluded)
components like spatial interactions and surface processing. The
latter may facilitate readout of additional features, provided inter-
component interactions are modeled properly.
To avoid model over-tting to the reconstructed stimuli, retinotopy
and PSF parameters were estimated via independent experiments using
stimuli that had minimal to no overlap with the test stimuli. Non-
linearity parameters were estimated directly from observed activation
distributions. This eliminated the possibility of circular contribution
from the encoding to the reconstruction direction that could affect re-
construction performance.
Reconstruction evaluation had low sensitivity to contour width (see
Materials and Methods), and reconstructed contours were thicker than
actual ones, which likely contributed to the presence of some elevated
correlations to stimuli other than the original. This may be due to corti-
cal magnication, explicitly modeled in the retinotopic component but
also partly expressed in the PSF. PSF imperfections and de-convolution
performance under noisy conditions also may have contributed.
For some stimuli positioned close to the VM, evoked activations cor-
responding to near-VM contour parts (0.5° and closer) were weaker
(Fig. 1b iv, v). Consequently, reconstruction of these contour portions
was more difcult (Fig. 2c iv, v). There are few possible explanations
for this phenomenon: (1) VSD staining quality; (2) effect of small eye
movements/drift, which may cause the contour located close to the
VM to shift to the opposite hemi-eld; and (3) the simplied PSF repre-
sentation, which does not fully capture the special characteristics of V1
spread prole near the VM (where activation in V1 can show a steep
decay near the V1/V2 border, which is also the VM).
Edge intensity varied along reconstructed contours. This could be
explained by variable SNR within the imaged area, but also by the
model's over-simplications; PSFs were constant throughout the VF,
while both RF and PSF size vary with eccentricity (Angelucci et al.,
2002; Van Essen et al., 1984). Similarly, contrast gain is inuenced local-
ly, and non-linearity parameters were xed for the entire image. Both
simplications were important in constructing a simple, reversible
model.
Acknowledgments
We thank Inbal Ayzenshtat for valuable input concerning retinotopic
transformations and computational considerations. This work was
supported by the DFG: Program of German-Israeli Project cooperation
(DIP grant, ref: 185/1-1) and by theIsraeli Center of Research Excellence
(I-CORE) in Cognition (I-CORE Program 51/11).
Appendix A. Supplementary data
Supplementary data to this article can be found online at http://dx.
doi.org/10.1016/j.neuroimage.2015.10.072.
References
Albrecht, D.G., 1995. Visual cortex neurons in monkey and cat: effect of contrast on the
spatial and temporal phase transfer functions. Vis. Neurosci. 12, 11911210.
Angelucci, A., Levitt, J.B., Walton, E.J.S., Hupe, J.M., Bullier, J., Lund, J.S., 2002. Circuits for
local and global signal integration in primary visual cortex. J. Neurosci. 22 ,
86338646.
Ayzenshtat, I., Meirovithz, E., Edelman, H., Werner-reiss, U., Bienenstock, E., Abeles, M.,
Slovin, H., 2010. Precise spati otemporal patterns among visual cortical areas and
their relation to visual stimulus processing. J. Neurosci. 30, 1123211245.
Ayzenshtat, I., Gilad, A., Zurawel, G., Slovin, H., 2012. Population response to natural im-
ages in the primary visual cortex encodes local stimulus attributes and perceptual
processing. J. Neurosci. 32, 1397113986.
Das, A., Gilbert, C.D.,1995. Long-range horizontal connections and their role in corticalre-
organization revealed by optical recording of cat primary visual cortex. Nature 375,
780784.
DeValois, R.L., DeValois, K.K., 1988. Spatial Vision. Oxford University Press, USA.
Engbert, R., Klieg l, R., 2003. Microsaccades uncoverthe orientation of covertattention. Vis.
Res. 43, 10351045.
Freiwald, W.A., Tsao,D.Y., 2010. Functional compartmentalization and viewpoint general-
ization within the macaque face-processing system. Science 330, 845851.
Gilad, A., Meirovithz, E., Slovin, H., 2013. Population responses to contour integration:
early encoding of discrete eleme nts and late perceptual grouping. Neuron 78,
389402.
Gollisch, T., Meister, M., 2008. Rapid neural coding in the retina with relative spike laten-
cies. Science 319, 11081111.
Grinvald, A., Lieke, E.E., Frostig, R.D., Hildesheim, R., 1994. Cortical point-spread function
and long-range lateral interactions revealed by real-time optical imaging of macaque
monkey primary visual cortex. J. Neurosci. 14, 25452568.
Gur, M., 2015. Space reconstruction by primary visual cortex activity: a parallel, non-
computational mechanism of object representation. Trends Neurosci. 38, 207216.
Haynes, J.D., Rees, G., 2005. Predicting the orientation of invisible stimuli from activity in
human primary visual cortex. Nat. Neurosci. 8, 686691.
Hochstein, S., Ahissar, M., 2002. View from the top: hierarchies and reverse hierarchies.
Review 36, 791804.
Horikawa, T., Tamaki, M., Miyawaki, Y., Kamitani, Y., 2013. Neural decoding of visual im-
agery during sleep. Science 340, 639642.
Hubel, D.H., Wiesel, T.N., 1968. Receptive elds and functional architecture of monkey
striate cortex. J. Physiol. 195, 215243.
Hung, C.P., Kreiman, G., Poggio, T., DiCarlo, J.J., 2005. Fast readout of object identity from
macaque inferior temporal cortex. Science 310, 863866.
Kay, K.N., Naselaris, T., Prenger, R.J., Gallant, J.L., 2008. Identifying natural images from
human brain activity. Nature 452, 352355.
Lamme, A.F., 1995. The neurophysiology visual cortexgure-ground segregation primary.
J. Neurosci. 15, 16051615.
Lucy, L., 1974. An iterative technique for th e rectication of ob served distributions.
Astron. J. 79, 745.
Miyawaki, Y., Uchida, H., Yamashita, O., Sato, M., Morito, Y., Tanabe, H.C., Sad ato, N.,
Kamitani, Y., 2008. Visual image reconstruction from human brain activity using a
combination of multiscale local image decoders. Neuron 60, 915929.
Naselaris, T., Prenger, R.J., Kay, K.N., Oliver, M., Gallant, J.L., 2009. Bayesian reconstruction
of natural images from human brain activity. Neuron 63, 902915.
Richardson, W.H., 1972. Bayesian-based iterative method of image restoration. JOSA 62,
5559.
Schira, M.M., Tyler,C.W., Spehar, B., Breakspear,M., 2010. Modeling magnication and an-
isotropy in the primate foveal conuence. PLoS Comput. Biol. 6, e1000651.
Sincich, L.C., Blasdel, G.G., 2001. Oriented axon projections in primary visual cortex of the
monkey. J. Neurosci. 21, 44164426.
Slovin, H., Arieli, A., Hildesheim, R., Grinvald, A., 2002. Long-term voltage-sensitive dye
imaging reveals cortical dynamics in behaving monke ys. J. Neurophysiol. 88,
34213438.
Stanley, G.B., Li, F.F., Dan, Y., 1999. Reconstruction of natural scenes from ensemble re-
sponses in the lateral geniculate nucleus. J. Neurosci. 19, 80368042.
Thirion, B., Duchesnay, E., Hubbard, E., Dubois, J., Poline, J.-B., Lebihan, D., Dehaene, S.,
2006. Inverse retinotopy: inferring the visual content of images from brain activation
patterns. Neuroimage 33, 11041116.
Tootell, R.B., Silverman, M.S., Hamilton, S.L., De Valois, R.L., Switkes, E., 1988. Functional
anatomy of macaque striate cortex. III. Color. J. Neurosci. 8, 15691593.
Van Essen, D.C., Newsome, W.T., Maunsell, J.H., 1984. The visual eld representation in
striate cortex of the macaque monkey: asymmetries, anisotropies, and individual var-
iability. Vis. Res. 24, 429448.
Wokke, M.E.,Vandenbroucke, A.R.E., Scholte, H.S., Lamme, V.A.F., 2013. Confuse your illu-
sion: feedback to early visual cortex contributes to perceptual completion. Psychol.
Sci. 24, 6371.
Xing, D., Yeh, C.I., Shapley, R.M., 2009. Spatial spread of the local eld potential and its
laminar variation in visual cortex. J. Neurosci. 29, 1154011549.
Zurawel, G., Ayzenshtat, I., Zweig, S., Shapley, R., Slovin, H., 2014. A contrast and surface
code explains complex responses to black and white stimuli in V1. J. Neurosci. 34,
1438814402.
1012 G. Zurawel et al. / NeuroImage 125 (2016) 10051012
... The copyright holder for this preprint this version posted https://doi.org/10.1101https://doi.org/10. /2022 monkeys' fovea, the spatial patterns of neural activation evoked simple shape stimuli can show a topographic code, namely, they resembles roughly the shape of the original stimuli (Zurawel, Shamir and Slovin, 2016;Macknik et al., 2019). ...
... Because the shape stimuli were comprised of thin contours (width 5 deg; i.e. not a filled surfaces), one would expect a lower neural response at the center of the shape contour, as was previously shown in monkeys' V1 (Zurawel, Shamir and Slovin, 2016;Macknik et al., 2019). However, we found the opposite: the population response corresponding to the shape center was higher than the response amplitude at the regions corresponding to the (which was not certified by peer review) is the author/funder. ...
... /2022 population responses to the shape contours that appear in the monkey's fovea, show clear resemblance to the presented shape contour with large differences between the shapes (Fig. 6A left). We note that in a recent work, we showed that the VSD activation maps for the shape stimuli can be well predicted from a 2D analytical model of a retinotopic map convolved with a PSF (see Zurawel, Shamir and Slovin, 2016). In contrast, the evoked VSD response in mice V1 (Fig. 6A right) showed a large patch of activation w/o a clear similarity to the visual stimulus and high similarity among the stimuli. ...
Preprint
Full-text available
Recently, mice became a popular research model for the visual system including objects processing. However, while primates use their fovea to inspect objects at high resolution, mice have low visual acuity and lack a foveal region. Thus, how objects are encoded in the visual cortex of mice and whether monkeys and mice use similar neural mechanisms to discriminate between shapes, is unknown. Using voltage sensitive dye imaging, we investigated spatio-temporal patterns of population response in the primary visual cortex (V1) of mice and monkeys to shape contours. Neural responses in mice showed blurred spatial patterns, low correlation to the original stimuli and only subtle spatial differences between the shapes. In contrast, the neural responses in monkey’s V1 fovea, showed prominent differences between the shapes at sub-degree resolution with high resemblance to the original shapes, in agreement with a topographic code. Furthermore, the spatial patterns of neural synchronization in monkey’s V1 suggested contour binding by synchrony, but not in mice V1. Finally, monkey’s V1 responses at larger eccentricities, resembled more to those in mice. These results may suggest that to discriminate between shapes, mice rely on neural processing of low-level cues, whereas monkeys rely more on a topographic code.
... The application of these methods to in-depth analyses of neural circuits has required the field to overcome multiple technical obstacles, from achieving adequate expression of genetic calcium indicators in NHPs when previously tested rodent methods failed to translate easily to primates, to designing imaging implants suitable for maintaining long-term brain health, to minimizing motion arti- Figure 1. A,ExperimentalsetupforVSDIinNHPs.B (Zurawel et al., 2016); baseline activity was subtracted before reconstruction. facts during imaging of brains at a spatial scale that is relevant to humans. ...
... The Slovin laboratory has applied VSDI to measure cortical responses to visual stimuli and analyze neural circuits in V1. They presented visual objects defined by either luminance or color while recording with VSDI in V1 of NHPs to test the hypothesis that visual perception of uniform surfaces is mediated by an isomorphic, filled-in representation (Zurawel et al., 2014(Zurawel et al., , 2016Zweig et al., 2015). Contrary to this hypothesis, the experiments revealed that neural responses to both color and luminance surfaces were similarly edge-enhanced in that VSDI pixels at the spatial locations occupied by edges showed greater voltage changes than those responding to the middle regions of the surfaces, appearing rapidly after stimulus onset. ...
... Finally, to test whether V1 can be used for high-resolution readout of visual stimuli, Slovin group attempted to reconstruct, at sub-degree resolution, visual contours from the VSD signal that were evoked by simple shapes (Zurawel et al., 2016;Fig. 1B). ...
Article
Full-text available
Novel genetically encoded tools and advanced microscopy methods have revolutionized neural circuit analyses in insects and rodents over the last two decades. Whereas numerous technical hurdles originally barred these methodologies from success in nonhuman primates (NHPs), current research has started to overcome those barriers. In some cases, methodological advances developed with NHPs have even surpassed their precursors. One such advance includes new ultra-large imaging windows on NHP cortex, which are larger than the entire rodent brain and allow analysis unprecedented ultra-large-scale circuits. NHP imaging chambers now remain patent for periods longer than a mouse's lifespan, allowing for long-term all-optical interrogation of identified circuits and neurons over timeframes that are relevant to human cognitive development. Here we present some recent imaging advances brought forth by research teams using macaques and marmosets. These include technical developments in optogenetics; voltage-, calcium- and glutamate-sensitive dye imaging; two-photon and wide-field optical imaging; viral delivery; and genetic expression of indicators and light-activated proteins that result in the visualization of tens of thousands of identified cortical neurons in NHPs. We describe a subset of the many recent advances in circuit and cellular imaging tools in NHPs focusing here primarily on the research presented during the corresponding mini-symposium at the 2019 Society for Neuroscience annual meeting.
... Because the two animal models have a large difference in their visual acuity, the surface size was adjusted to each animal in a behaviorally relevant way for each species. Most mouse studies use large stimuli (>30 • ; Douglas et al. 2006;Jurjut et al. 2017;Kirchberger et al. 2021;Schnabel et al. 2018), whereas monkeys and humans tend to be presented with much smaller surfaces (Zweig et al. 2015;Zurawel et al. 2016;Yang et al. 2022). When adjusting the surface size for each animal, the ratio of the V1 RF size to the surface size was preserved between species. ...
... ecologically) relevant manner and used a similar ratio between the RF size and the surface size. Most studies in mice have used stimuli larger than 30 • (Douglas et al. 2006;Jurjut et al. 2017;Schnabel et al. 2018;Kirchberger et al. 2021), while for monkeys, studies have employed much smaller stimuli (<4 • ; Yang et al. 2022;Zurawel et al. 2016;Zweig et al. 2015;Xing et al. 2014). This difference is related to the unique foveal representation in monkeys where the RFs of the neurons are small (0.2 • to 0.4 • ; Tehovnik and Slocum 2007). ...
Article
Luminance and spatial contrast provide information on the surfaces and edges of objects. We investigated neural responses to black and white surfaces in the primary visual cortex (V1) of mice and monkeys. Unlike primates that use their fovea to inspect objects with high acuity, mice lack a fovea and have low visual acuity. It thus remains unclear whether monkeys and mice share similar neural mechanisms to process surfaces. The animals were presented with white or black surfaces and the population responses were measured at high spatial and temporal resolution using voltage-sensitive dye imaging. In mice, the population response to the surface was not edge-dominated with a tendency to center-dominance, whereas in monkeys the response was edge-dominated with a "hole" in the center of the surface. The population response to the surfaces in both species exhibited suppression relative to a grating stimulus. These results reveal the differences in spatial patterns to luminance surfaces in the V1 of mice and monkeys and provide evidence for a shared suppression process relative to grating.
... However, fine stimulus details also contribute to shape our visual perception, and it 14 appears plausible for this to be reflected in the response pattern of many brain regions, 15 especially at the early stages of the visual stream [13]. How much information about 16 these fine details can be recovered from neural activity in V1? ...
... Bialek et al. first estimated a one 35 dimensional stimulus from single neuron recording in the fly visual system [2]. 36 Subsequently, other noteworthy examples, in addition to the ones previously mentioned, 37 include the reconstruction of natural movies from cat LGN using liner spatio-temporal 38 filters [23], and geometrical shape estimation from monkey V1 using deconvolution with 39 a point spread function [13]. Thirion et al. extended the field to artificial visual scene 40 reconstruction from fMRI recording in humans [24]. ...
Preprint
Deciphering the neural code, that is interpreting the responses of sensory neurons from the perspective of a downstream population, is an important step towards understanding how the brain processes sensory stimulation. While previous work has focused on classification algorithms to identify the most likely stimulus label in a predefined set of categories, fewer studies have approached a full stimulus reconstruction task. Outstanding questions revolve around the type of algorithm that is most suited to decoding (i.e. full reconstruction, in the context of this study), especially in the presence of strong encoding non-linearities, and the possible role of pairwise correlations. We present, here, the first pixel-by-pixel reconstruction of a complex natural stimulus from 2-photon calcium imaging responses of mouse primary visual cortex (V1). We decoded the activity of approximately 100 neurons from layer 2/3 using an optimal linear estimator and an artificial neural network. We also investigated how much accuracy is lost in this decoding operation when ignoring pairwise neural correlations. We found that a simple linear estimator is sufficient to extract relevant stimulus features from the neural responses, and that it was not significantly outperformed by a non-linear decoding algorithm. The importance of pairwise correlations for reconstruction accuracy was also limited. The results of this study suggest that, conditional on the spatial and temporal limits of the recording technique, V1 neurons display linear readout properties, with low information content in the joint distribution of their activity.
... Conventional coding mechanisms often study the neuronal modulation properties under visual stimulation, but there is also an alternative approach, which is the decoding of sensory inputs from the recorded neuronal responses. The existing decoding method involves the use of neuronal spike firing [6,7] and MRI [8,9]. The method used for decoding includes three parts: feature extraction, decoding algorithm design, and decoding evaluation. ...
... Previous reports have mostly investigated the modulation of mammalian optic nerve signals to external stimuli [8,17,19,20], whereas there is a significant difference between the mammalian visual system and the bird's visual system. This research explores how the neural signals represent the luminance information in a developed visual system of birds. ...
Article
Full-text available
Important aspects of brain information processing can be understood by examining decoding of visual stimuli from neuronal response signals. In this research, the luminance information is decoded from the local field potential signal in the optic tectum region of the pigeon. We designed a luminance visual stimulus model with transient flicker characteristics, recorded multichannel local field potential (LFP) signals using a microelectrode array, extracted LFP Fourier transform energy and phase features, constructed a multivariate linear inverse filter luminance information decoding algorithm, and evaluated decoding effects using a cross-correlation method. We found that LFP signal phase decoding of luminance information yielded better effects than amplitude decoding of luminance information. In the case of optimal frequency band, channels, delay time, and other parameters, the results of phase and amplitude codecoding could reach 0.94±0.02. Comparing the differences between neuronal spike decoding and LFP decoding, we found that LFP signal phase and amplitude codecoding resulted in luminance closer to that of the actual stimulus and required fewer decoding electrode channels.
... This enables the use of epi-illumination, where the illuminating light is sent to the cortex through the last objective by reflecting from a dichroic mirror placed between the two lenses. This instrument is well mastered by experimenters, and has lead to many discoveries over the years [5] [7] [19], [27], [3], [16], [33], [9], [35]. Nevertheless, there are some fundamental limitations inherent to the macroscope's design. ...
Article
Full-text available
Objective: Cortical activity can be recorded using a variety of tools, ranging in scale from the single neuron (microscopic) to the whole brain (macroscopic). There is usually a trade-off between scale and resolution; optical imaging techniques, with their high spatio-temporal resolution and wide field of view, are best suited to study brain activity at the mesoscale. Optical imaging of cortical areas is however in practice limited by the curvature of the brain, which causes the image quality to deteriorate significantly away from the center of the image. Approach: To address this issue and harness the full potential of optical cortical imaging techniques, we developed a new wide-field optical imaging system adapted to the macaque brain. Our system is composed of a curved detector, an aspherical lens and a ring composed of LEDs providing uniform illumination at wavelengths relevant for the different optical imaging methods, including intrinsic and fluorescence imaging. Main results: The system was characterized and compared with the standard macroscope used for cortical imaging, and a 3-fold increase of the area in focus was measured as well as a 4-fold increase in the evenness of the optical quality in vivo. Significance: This new instrument, which is to the best of our knowledge the first use of a curved detector for cortical imaging, should facilitate the observation of wide mesoscale phenomena such as dynamic propagating waves within and between cortical maps, which are otherwise difficult to observe due to technical limitations of the currently available recording tools.
... Another research direction would be to generate inverse models of visual stimuli based on V1 representations obtained with fast optical imaging such as voltage sensitive dyes. 32 The next step would be to create complex visual percepts using an INS array to generate a stimulation pattern in V1. ...
Article
Full-text available
Significance: We present a new optical method for modulating cortical activity in multiple locations and across multiple time points with high spatial and temporal precision. Our method uses infrared light and does not require dyes or transgenic modifications. It is compatible with a number of other stimulation and recording techniques. Aim: Infrared neural stimulation (INS) has been largely confined to single point stimuli. In this study, we expand upon this approach and develop a rapidly switched fiber array capable of generation of stimulus patterns. Our prototype is capable of stimulating at nine separate locations but is easily scalable. Approach: Our device is made of commercially available components: a solid-state infrared laser, a piezoelectric fiber coupled optical switch, and 200 - μ m diameter optical fibers. We validate it using intrinsic optical signal imaging of INS responses in macaque and squirrel monkey sensory cortical areas. Results: We demonstrate that our switched array can consistently generate responses in primate cortex, consistent with earlier single channel INS investigations. Conclusions: Our device can successfully target the cortical surface, either at one specific region or multiple points spread out across different areas. It is compatible with a host of other imaging and stimulation modalities.
... Multivoxel pattern analysis (MVPA) techniques overcome this limit and have been instrumental in assessing how the cortex stores, encodes, and organizes information. Meso-scale responses to features such as color, orientation, and edges can be used to build multivoxel "decoders," which can even serve to estimate the experience of the subject (Naselaris et al. 2009;Zurawel et al. 2016). Spatial patterns can carry category (Haxby et al. 2001) or task-related information (Kriegeskorte et al. 2006), they can be robust even in the absence of conscious perception (Sterzer et al. 2008;Schurger et al. 2010), and these findings can generalize to nonvisual modalities (Abrams et al. 2011). ...
Article
Full-text available
A fundamental step to predicting brain activity in healthy and diseased populations is characterizing the common spatio-temporal response to a shared experience. Multivoxel pattern analysis allows us to investigate information encoding through these patterns; however, we have yet to explore local, stimulus-driven, patterns of cortical activity during naturalistic stimulation. We sought to examine these patterns with minimum interpolation—excluding functional alignment—to characterize the most basic degree of shared response between subjects. We used an unbiased analytic approach, combined with rich, naturalistic, and nonsemantic stimulation to estimate shared spatial patterns in functional magnetic resonance imaging responses across a large group. We found that meso-scale spatial patterns were shared nonuniformly across the visual cortex and represent information distinct from the shared temporal response. Shared spatial patterns were stimulus-driven, modulated by pattern size, and more sensitive to the contrast of 3D versus 2D stimulus differences than the temporal signals. Although the grand functional structure of the brain is understood to be common, these results suggest that even at a meso-scale, we share common spatial structures with anatomical alignment alone. The strength of this similarity varies across the cortex, suggesting some spatial structures are innately organized, whereas others are shaped by factors such as learning and plasticity.
... Novel brain imaging methods, in terms of both hardware (e.g. functional near-infrared spectroscopy, fNIR or fNIRS; Dresler et al. 2011), and software (statistical learning or machine learning algorithms; Huth et al. 2016;Wen et al. 2016, preprint;Zurawel et al. 2016) may enable a more objective readout of the dream. This is particularly important considering that (i) memory is reconstructive and there is a time lag between the actual dream and its report, (ii) the original dream content may be distorted due to interfering material of the waking environment, (iii) there are limitations inherent to verbal reports, and (iv) moral censorship (Hobson et al. 2000;Windt 2013). ...
Article
Full-text available
The science of dreaming constitutes a relevant topic in modern-day neuroscientific research and provides major insights into the study of human consciousness. Linking specific, universal, and regularly occurring stages of sleep with dreaming encourages the direct and systematic investigation of a topic that has fascinated humankind for centuries. In this review, we explore to what extent individuals dream during periods of rapid eye movement and non-rapid eye movement sleep, and we introduce research on lucid dreaming. We then discuss how dreaming during different stages of sleep varies in terms of phenomenological characteristics, and to what extent individuals are conscious throughout the sleep cycle. Finally, we provide a synopsis of the previous literature on brain activity during sleep, and we aim to clarify how the neurofunctional changes observed throughout sleep may lead to changes in phenomenological aspects of dreams, and in the domain of consciousness.
Article
Constructing computational decoding models to account for the cortical representation of semantic information plays a crucial role in understanding visual perception. The human visual system processes interactive relationships among different objects when perceiving the semantic contents of natural visions. However, the existing semantic decoding models commonly regard categories as completely separate and independent visually and semantically and rarely consider the relationships from prior information. In this work, a novel semantic graph learning model was proposed to decode multiple semantic categories of perceived natural images from brain activity. The proposed model was validated on the functional magnetic resonance imaging data collected from five normal subjects while viewing 2750 natural images comprising 52 semantic categories. The results showed that the Graph Neural Network-based decoding model achieved higher accuracies than other deep neural network models. Moreover, the co-occurrence probability among semantic categories showed a significant correlation with the decoding accuracy. Additionally, the results suggested that semantic content organized in a hierarchical way with higher visual areas was more closely related to the internal visual experience. Together, this study provides a superior computational framework for multi-semantic decoding that supports the visual integration mechanism of semantic processing.
Article
Full-text available
We investigated the cortical mechanisms underlying the visual perception of luminance-defined surfaces and the preference for black over white stimuli in the macaque primary visual cortex, V1. We measured V1 population responses with voltage-sensitive dye imaging in fixating monkeys that were presented with white or black squares of equal contrast around a mid-gray. Regions corresponding to the squares' edges exhibited higher activity than those corresponding to the center. Responses to black were higher than to white, surprisingly to a much greater extent in the representation of the square's center. Additionally, the square-evoked activation patterns exhibited spatial modulations along the edges and corners. A model comprised of neural mechanisms that compute local contrast, local luminance temporal modulations in the black and white directions, and cortical center-surround interactions, could explain the observed population activity patterns in detail. The model captured the weaker contribution of V1 neurons that respond to positive (white) and negative (black) luminance surfaces, and the stronger contribution of V1 neurons that respond to edge contrast. Also, the model demonstrated how the response preference for black could be explained in terms of stronger surface-related activation to negative luminance modulation. The spatial modulations along the edges were accounted for by surround suppression. Overall the results reveal the relative strength of edge contrast and surface signals in the V1 response to visual objects.
Article
Full-text available
The neuronal mechanisms underlying perceptual grouping of discrete, similarly oriented elements are not well understood. To investigate this, we measured neural population responses using voltage-sensitive dye imaging in V1 of monkeys trained on a contour-detection task. By mapping the contour and background elements onto V1, we could study their neural processing. Population response early in time showed activation patches corresponding to the contour/background individual elements. However, late increased activity in the contour elements, along with suppressed activity in the background elements, enabled us to visualize in single trials a salient continuous contour "popping out" from a suppressed background. This modulated activity in the contour and in background extended beyond the cortical representation of individual contour or background elements. Finally, the late modulation was correlated with behavioral performance of contour saliency and the monkeys' perceptual report. Thus, opposing responses in the contour and background may underlie perceptual grouping in V1.
Article
Full-text available
Reading Dreams How specific visual dream contents are represented by brain activity is unclear. Machine-learning–based analyses can decode the stimulus- and task-induced brain activity patterns that represent specific visual contents. Horikawa et al. (p. 639 , published online 4 April) examined patterns of brain activity during dreaming and compared these to waking responses to visual stimuli. The findings suggest that the visual content of dreams is represented by the same neural substrate as observed during awake perception.
Article
Full-text available
A striking example of the constructive nature of visual perception is how the human visual system completes contours of occluded objects. To date, it is unclear whether perceptual completion emerges during early stages of visual processing or whether higher-level mechanisms are necessary. To answer this question, we used transcranial magnetic stimulation to disrupt signaling in V1/V2 and in the lateral occipital (LO) area at different moments in time while participants performed a discrimination task involving a Kanizsa-type illusory figure. Results show that both V1/V2 and higher-level visual area LO are critically involved in perceptual completion. However, these areas seem to be involved in an inverse hierarchical fashion, in which the critical time window for V1/V2 follows that for LO. These results are in line with the growing evidence that feedback to V1/V2 contributes to perceptual completion.
Article
Full-text available
The primary visual cortex (V1) is extensively studied with a large repertoire of stimuli, yet little is known about its encoding of natural images. Using voltage-sensitive dye imaging in behaving monkeys, we measured neural population response evoked in V1 by natural images presented during a face/scramble discrimination task. The population response showed two distinct phases of activity: an early phase that was spread over most of the imaged area, and a late phase that was spatially confined. To study the detailed relation between the stimulus and the population response, we used a simple encoding model to compute a continuous map of the expected neural response based on local attributes of the stimulus (luminance and contrast), followed by an analytical retinotopic transformation. Then, we computed the spatial correlation between the maps of the expected and observed response. We found that the early response was highly correlated with the local luminance of the stimulus and was sufficient to effectively discriminate between stimuli at the single trial level. The late response, on the other hand, showed a much lower correlation to the local luminance, was confined to central parts of the face images, and was highly correlated with the animal's perceptual report. Our study reveals a continuous spatial encoding of low- and high-level features of natural images in V1. The low level is directly linked to the stimulus basic local attributes and the high level is correlated with the perceptual outcome of the stimulus processing.
Article
It has been proposed that activity in human primary visual cortex (V1) is necessary but not sufficient for conscious awareness, but direct physiological evidence for the activation of human V1 in the absence of awareness has been elusive. Here, we used functional MRI to measure activity in V1 while participants viewed masked and invisible orientated gratings. Remarkably, when participants were unable to report the orientation of a masked grating they were viewing, we could nevertheless predict its orientation significantly better than chance, based only on a single brief measurement of activity from their primary visual cortex. This was achieved by using multivariate pattern-recognition to substantially improve the ability to detect orientation signals in V1 (see also Kamitani & Tong, VSS 2004). Thus, activity in primary visual cortex conveys orientation-specific information that is inaccessible to conscious report. Orientation can therefore be represented unconsciously in V1, suggesting that information must be relayed to another region of the brain to be represented in conscious experience.
Article
The current view posits that objects, despite changes in appearance, are uniquely encoded by 'expert' cells. This view is untenable. First, even if cell ensemble responses are invariant and unique, we are consciously aware of all of the objects' details. Second, in addition to detail preservation, data show that the current hypothesis fails to account for uniqueness and invariance. I present an alternative view whereby objects' representation and recognition are based on parallel representation of space by primary visual cortex (V1) responses. Information necessary for invariance and other attributes is handled in series by other cortical areas through integration, interpolation, and hierarchical convergence. The parallel and serial mechanisms combine to enable our flexible space perception. Only in this alternative view is conscious perception consistent with the underlying architecture. Copyright © 2015 Elsevier Ltd. All rights reserved.
Article
An iterative method of restoring degraded images was developed by treating images, point spread functions, and degraded images as probability-frequency functions and by applying Bayes’s theorem. The method functions effectively in the presence of noise and is adaptable to computer operation.
Article
The topographic organization of striate cortex in the macaque was studied using physiological recording techniques. Results were displayed on two-dimensional maps of the cortex, which facilitated the quantitative analysis of various features of the visual representation. The representation was found to be asymmetric, with more cortex devoted to lower than to upper fields. Over much of striate cortex the representation is anisotropic, in that the magnification factor depends upon the direction along which it is measured. There is considerable individual variability in these features as well as in the overall size of striate cortex. Outside the fovea, the cortical representation shows only modest deviations from a logarithmic conformai mapping, in which the magnification factor is proportional to the inverse of eccentricity in the visual field. Comparison of receptive field size with cortical magnification was used to estimate the “point image size” in the cortex (i.e. the extent of cortex concerned with processing inputs from any given point in the visual field). Our evidence supports a previous report that point-image size varies significantly with eccentricity. This is of interest in relation to anatomical evidence that the dimensions of columnar systems in striate cortex are largely independent of eccentricity.