Conference PaperPDF Available

Bridging artificial and natural language learning: Comparing processing- and reflection-based measures of learning

Authors:

Abstract and Figures

A common assumption in the cognitive sciences is that artificial and natural language learning rely on shared mechanisms. However, attempts to bridge the two have yielded ambiguous results. We suggest that an empirical disconnect between the computations employed during learning and the methods employed at test may explain these mixed results. Further, we propose statistically-based chunking as a potential computational link between artificial and natural language learning. We compare the acquisition of non-adjacent dependencies to that of natural language structure using two types of tasks: reflection-based 2AFC measures, and processing-based recall measures, the latter being more computationally analogous to the processes used during language acquisition. Our results demonstrate that task-type significantly influences the correlations observed between artificial and natural language acquisition, with reflection-based and processing-based measures correlating within – but not across – task-type. These findings have fundamental implications for artificial-to-natural language comparisons, both methodologically and theoretically.
Content may be subject to copyright.
Bridging artificial and natural language learning:
Comparing processing- and reflection-based measures of learning
Erin S. Isbilen (esi6@cornell.edu)
Cornell University, Department of Psychology, Ithaca, NY 14850 USA
Rebecca L.A. Frost (rebecca.frost@mpi.nl)
Max Planck Institute for Psycholinguistics, Language Development Department, Nijmegen, 6525 XD, Netherlands
Padraic Monaghan (p.monaghan@lancaster.ac.uk)
Lancaster University, Department of Psychology, Lancaster, LA1 4YF, UK
Morten H. Christiansen (christiansen@cornell.edu)
Cornell University, Department of Psychology, Ithaca, NY 14850 USA
Abstract
A common assumption in the cognitive sciences is that
artificial and natural language learning rely on shared
mechanisms. However, attempts to bridge the two have
yielded ambiguous results. We suggest that an empirical
disconnect between the computations employed during
learning and the methods employed at test may explain these
mixed results. Further, we propose statistically-based
chunking as a potential computational link between artificial
and natural language learning. We compare the acquisition of
non-adjacent dependencies to that of natural language
structure using two types of tasks: reflection-based 2AFC
measures, and processing-based recall measures, the latter
being more computationally analogous to the processes used
during language acquisition. Our results demonstrate that
task-type significantly influences the correlations observed
between artificial and natural language acquisition, with
reflection-based and processing-based measures correlating
within – but not across – task-type. These findings have
fundamental implications for artificial-to-natural language
comparisons, both methodologically and theoretically.
Keywords: statistical learning; chunking; language; artificial
language learning; cross-situational learning; non-adjacent
dependencies; learning; memory; serial recall; methodology
Introduction
Connecting individual differences in artificial and natural
language learning is an ongoing endeavor in the cognitive
sciences. These studies operate on the assumption that
artificial language learning tasks designed for use in the
laboratory draw on the same cognitive processes that
underpin language acquisition in the real world (e.g.,
Saffran, Aslin, & Newport, 1996). Yet, attempts to bridge
artificial and natural language learning have yielded mixed
results, often finding weaker correlations between language
measures that should in theory rely on shared computations
(Siegelman, Bogaerts, Christiansen & Frost, 2017). Part of
the problem may lie in the nature of the tests used to
evaluate learning; although artificial and natural language
learning may rely on the same computational processes,
different tests may tap into separate subcomponents of these
skills, making the relationship difficult to unpack.
Artificial language learning tasks are assumed to capture
key aspects of how learners acquire language in the real
world: by drawing on the distributional information
contained in speech. Through exposure to statistical
regularities in the input, the cognitive system picks up on
linguistic units without awareness by the learner (Saffran et
al., 1996). Yet, in adults, statistical learning is typically
tested using measures that require participants to reflect on
their knowledge and provide an overt judgment, such as in
the two-alternative forced-choice task (2AFC); a test that,
while potentially informative, only provides a meta-
cognitive measure of learning. Indeed, language learning
measures can be broadly divided into two categories:
reflection-based measures (e.g., 2AFC), which translate the
primary effects of learning into a secondary response, and
processing-based measures, which rely on the same
computations as the learning itself (Christiansen, 2018). In
psycholinguistic research, it is often the case that the
learning measures employed at test do not align with the
processes employed during learning. We propose that this
disconnect may have constrained prior observations of the
relationship between artificial and natural language skills.
We seek to resolve some of this ambiguity in the study at
hand.
In the current study, we assess the degree to which
statistical learning abilities map onto natural language
acquisition, and evaluate correlations within and between
reflection- and processing-based measures. For the purpose
of this paper, we characterize artificial language learning as
statistical learning in a highly constrained, simplified
context, using the Saffran et al. (1996) familiarization
method. We simulate natural language acquisition by
presenting participants with a more complex cross-
situational learning task that utilizes natural vocabulary and
grammar with corresponding referents. For each part of the
experiment, we included two types of tests: reflection-based
tasks (2AFC), and processing-based tasks (recall), to allow
for a comparison of learning between and within task types.
1856
For our processing-based measure, we employed a
chunking-based recall task, building on the suggestion that
chunking plays a key role in statistical learning and
language acquisition (see Christiansen, 2018, for a review).
In 2AFC tasks, participants are required to indicate their
preference for one stimulus over another, which is taken to
indicate learning. In recall a task which is thought to rely
on chunking – participants repeat syllable strings that are
either congruent or incongruent with the statistics of the
input, with recall errors acting as a window into learning.
That is, learning is indexed by better recall of consistent
items when controlling for baseline phonological working
memory (Conway, Bauernschmidt, Huang & Pisoni, 2010;
Isbilen, McCauley, Kidd & Christiansen, 2017). If chunking
occurs during language acquisition, chunking-based tasks
may yield a better measure of learning than reflection-based
tasks such as 2AFC.
In the first part of the experiment, participants engaged in
a statistical learning task adapted from Frost and Monaghan
(2016), to test segmentation and generalization of non-
adjacent dependencies (the artificial language task). In the
second part, participants learned a fragment of Japanese,
comprising a small vocabulary and simple grammar using a
cross-situational learning task adapted from Rebuschat,
Ferrimand, and Monaghan (2017) and Walker,
Schoetensack, Monaghan and Rebuschat (2017; the natural
language task). We hypothesized that the correlations
observed between artificial and natural language learning
would show a strong effect of task type: reflection-based
measures would be more likely to correlate with other
reflection-based measures, whereas processing-based
measures would be more likely to correlate with other
processing-based measures. Such a pattern would have
important implications for individual differences work, and
about the deductions that can be applied to natural language
acquisition from artificial language learning tasks.
Part 1: Non-adjacent dependency learning in
an artificial language
In Part 1, we tested adults’ learning of an artificial language
composed of non-adjacent dependencies, which are
relationships between linguistic units that occur across one
or more variable intervening units (e.g., in an AXC structure
where units A and C reliably co-occur, but X varies
independently). These dependencies are found at multiple
levels of linguistic abstraction, including morphology within
words and syntactic dependencies between words, thereby
providing a tightly-controlled artificial structure that shares
structural similarity with natural language.
We examined learners’ ability to segment these non-
adjacent dependency sequences from speech, and generalize
them to new instances - skills which are integral to natural
language learning. We tested both segmentation and
generalization with a reflection-based task (2AFC), and a
processing-based task, the statistically-induced chunking
recall task (SICR; Isbilen et al., 2017). In the SICR task,
participants are presented with 6-syllable-long strings, that
are either composed of two words from the input, or the
same syllables presented in a random order. If participants
have successfully chunked the items in the artificial
language during training, we expect that they should
perform significantly better on recalling the strings derived
from the statistics of the input language. While 2AFC is
scored as a correct-incorrect binary, SICR is scored syllable-
by-syllable, which we suggest may provide more in-depth
insights into segmentation and generalization skills.
Building on the results of Frost and Monaghan (2016), we
hypothesized that both tasks would yield evidence of
simultaneous segmentation and generalization. However,
due to the differences in task demands between reflection-
and processing-based tests, we expected to see limited
correlations between measurement types.
Method
Participants 49 Cornell University undergraduates (30
females; age: M=19.43, SD=1.30) participated for course
credit. All participants were native English speakers, with
no experience learning Japanese.
Materials The same language and stimuli as Frost and
Monaghan (2016) were used, derived from Peña, Bonatti,
Nespor and Mehler (2002). The language was composed of
9 consonant-vowel syllables (be, du, fo, ga, li, ki, pu, ra, ta),
arranged into three tri-syllabic non-adjacent dependencies
containing three varying middle syllables (A1X1–3C1,
A2X1–3C2, and A3X1–3C3; 9 words in total). Four different
versions of the language were created to control for
potential preferences for certain phoneme combinations.
Syllables used for the A and C items contained plosives (be,
du, ga, ki, pu, ta), while the X syllables contained
continuants (fo, li, ra). The resulting 9 items are referred to
as segmentation words, sequences that were presented
during training. Nine generalization words were also
created, and were only presented at test. The generalization
words contained trained non-adjacent dependencies, but
with novel intervening syllables (thi, ve, zo, e.g., A1Y1-3C1).
The generalization words measure participants’ ability to
extrapolate the knowledge of the non-adjacent dependencies
gained during training to novel, unheard items.
For the 2AFC test, 18 additional foil words were created,
which were paired with segmentation and generalization
words. Foils for the segmentation test comprised part-word
sequences that spanned word boundaries (e.g., CAX,
XCA). Foils for the generalization test were part-words but
with one syllable switched out and replaced with a novel
syllable, to prevent participants from responding based on
novelty alone (e.g., NCA, XNA, CAN, see Frost &
Monaghan, 2016). For the SICR test, 27 six-syllable strings
were created: 9 composed of two concatenated segmentation
words (e.g., A1X1C1A2X2C2), 9 composed of two
generalization words (e.g., A1Y1C1A2Y2C2), and 9 foils.
The foils used the same syllables as the experimental items
in a pseudorandomized order that avoided using any
1857
transitional probabilities or non-adjacent dependencies from
the experimental items.
All stimuli were created using the Festival speech
synthesizer (Black et al., 1990). Each AXC string lasted
~700 ms, and was presented using E-Prime 2.0.
Procedure For training, the 9 segmentation words were
concatenated into a continuous stream that participants
heard for 10.5 minutes. Participants were instructed to listen
carefully to the language and pay attention to the words it
might contain.
To test learning, two different tasks were used: the 2AFC
task and the SICR task (Isbilen et al., 2017). The order of
the two tests was counterbalanced to account for potential
order effects. In the 2AFC task, participants were presented
with 18 pairs of words: 9 segmentation pairs and 9
generalization pairs, with each pair featuring a target word
and corresponding foil. Segmentation and generalization
trials were randomized within the same block of testing.
Participants were instructed to carefully listen to each word
pair and indicate which of the two best matched the
language they heard during training. In the SICR task, 27
strings were randomly presented for recall: 9 segmentation
items, 9 generalization items, and 9 foils that served as a
baseline working memory measure. Participants were asked
to listen to each string and say the entire string out loud to
the best of their ability. Participants were not informed of
any underlying structure of the strings in either task.
Results and Discussion
First, we examined the data for task order effects (2AFC
first/SICR second versus SICR first/2AFC second), and
language effects (which of the four randomized languages
participants heard). A one-way ANOVA revealed a
significant effect of order on both SICR measures
(Segmentation: F(3,45)=-2.30, p=.026; Generalization:
F(3,45)=-3.30, p=.002), with participants who received
2AFC prior to SICR scoring significantly higher on these
two measures. Similarly, language significantly impacted
SICR generalization performance, F(3,45)=6.94, p=.0006,
suggesting that different syllable combinations may vary in
difficulty when being spoken aloud. All subsequent analyses
involving SICR in the remainder of the paper control for
order, and for SICR generalization, for both order and
language.
2AFC Performance Replicating the findings of Frost
and Monaghan (2016), participants showed simultaneous
segmentation and generalization of non-adjacent
dependencies, with performance on both tasks being
significantly above chance (Segmentation: M=.84, SD=.13;
t(48)= 18.44, p<.0001; Generalization: M=.70, SD=.21;
t(48)= 6.61, p<.0001). Performance was significantly more
accurate on segmentation than generalization trials,
t(48)=5.77, p<.0001, and segmentation and generalization
scores were highly correlated: r(47)=.53, p<.0001.
SICR Performance Participants’ verbal responses from
the SICR task were transcribed by two coders blind to the
study design. The transcriptions were subsequently scored
against the target test items to obtain measures of overall
accuracy (the total number of syllables recalled in the
correct serial position), and non-adjacent dependency
accuracy (the number of A-C pairings recalled from each
item, out of the two possible pairings: e.g., A1xC1A2xC2).
Replicating the results of Isbilen et al. (2017), participants
accurately recalled significantly more syllables in the
correct order for the experimental items than the random
items. These results held for both the segmentation items
(Experimental: M=34.84, SD=10.16; Random: M=13.55,
SD=6.04; t(48)= 23.11, p<.0001), and for the generalization
items (Experimental: M=27.71, SD=10.56; Random:
M=8.35, SD=4.37; t(48)= 15.98, p<.0001). Similarly, the
number of non-adjacent dependencies (syllables in the 1st &
3rd and/or the 4th & 6th serial positions) recalled for
experimental items was significantly higher than those
recalled for the random, both for the segmentation items
(Experimental: M=8.53, SD=4.52; Random: M=2.57,
SD=2.34; t(48)= 13.34, p<.0001), and for the generalization
items (Experimental: M=6.63, SD=4.42; Random: M=1.50,
SD=1.45; t(48)= 9.51, p<.0001). Unlike 2AFC, the SICR
results revealed no significant difference in performance
between the segmentation and generalization difference
scores (experimental minus random), although
generalization scores were slightly lower due to the
inclusion of unfamiliar syllables. These results held for both
overall recall, and for the total number of non-adjacent
dependencies recalled (p=.10 in both cases). This difference
between 2AFC and SICR may in part stem from differences
in task demands: differences in familiarity between
segmentation and generalization items may influence ratings
in 2AFC more, due to its meta-cognitive nature. SICR
generalization and segmentation performance was
significantly correlated: r(47)=.34, p=.02.
Correlations between 2AFC and SICR To evaluate the
relationship between reflection- and processing-based
measures, correlations were run between 2AFC and SICR
scores. The SICR values used for the correlations were the
total difference scores, to maximize the measures’
comparability to 2AFC (which is akin to a difference score),
while also controlling for baseline phonological working
memory (the subtraction of the random items from the
experimental items). The only significant correlation was
between 2AFC segmentation and SICR segmentation,
r(47)=.31, p=.04 (not correcting for multiple comparisons).
1
No other SICR and 2AFC measures were significantly
correlated (all p>.08). In line with our hypothesis, these
findings suggest that reflection- and processing-based
measures appear to capture largely different aspects of
statistical learning skills (Christiansen, 2018; Siegelman et
al., 2017).
The results of Part 1 replicate the results of Frost &
Monaghan (2016) of simultaneous segmentation and
generalization of non-adjacent dependencies using both
2AFC and SICR. Taken together, these findings suggest that
1
In a pilot version of this study, (N=61) no such correlation was
observed, potentially suggesting a type II error: r(59)=-.04, p=.76.
1858
statistical-chunking processes may be able to account for the
segmentation and generalization of non-adjacent
dependencies, as well as that of sequential dependencies.
Furthermore, we found that although reflection- and
processing-based measures showed evidence of learning,
performance across the two tasks was largely uncorrelated.
To test whether this pattern extends to natural language
acquisition, Part 2 of the experiment evaluated grammar and
vocabulary acquisition using patterns from natural language,
with a comparison of 2AFC and recall task types.
Part 2: Cross-situational language learning of
Japanese
Natural language acquisition involves a host of different
factors, including word segmentation, word-referent
mapping, discovery of sequential structure, and
generalization to novel instances. In the second part of this
experiment, we increased the complexity of the task to
explore the degree to which the learning taking place during
segmentation and the discovery of non-adjacent dependency
structure maps onto more naturalistic language acquisition.
A cross-situational language learning task based on Walker
et al. (2017) was administered, exposing participants to
Japanese sentences co-occurring with complex scenes.
Cross-situational learning simulates naturalistic language
learning in the lab by analogy to infants’ acquisition of
word-referent mappings non-ostensively, through hearing
instances of a word occurring with the same referent across
different contexts. Similar to Part 1, both reflection-based
and processing-based measures were used to evaluate
learning. 2AFC tests of noun, marker, and verb learning
were performed. Additionally, a combined forced-choice
and recall task was also administered to test syntax
acquisition: participants repeated whole sentences they
heard out loud, after which they rated the grammaticality of
each sentence. We hypothesized that vocabulary and
grammatical regularities would be acquired simultaneously,
similar to the concurrent segmentation and generalization in
the non-adjacent dependency task. Furthermore, we
anticipated that while all tests would show some evidence of
learning, only within task-type correlations would be
significantly related.
Method
Participants The same 49 participants from Part 1
partook in Part 2 immediately following the first task.
Materials A small lexicon of Japanese words was used
for this experiment, taken from Rebuschat et al. (2017). The
language consisted of 6 nouns (fukuoru, owl; kame, turtle;
niwatori, chicken; shimauma, zebra; ushi, cow; zou,
elephant), four verbs (kakusu, hide; mochiageru, lift; taosu,
push; tobikueru, jump), and two case markers (-ga and -o),
which were appended to the end of each noun to indicate
whether the noun was the subject (-ga) or the object (-o) of
the sentence. For instance, the sentence kamega
shimaumao taosu would indicate that the turtle (subject)
pushes the zebra (object). The language also used Japanese
syntax, with sentences having two possible grammatical
orders: subject-object-verb (SOV), and object-subject-verb
(OSV). For training, 192 sentences were generated. For test,
96 additional sentences were presented: 24 for each of the
marker, noun, and verb tests, and 24 for the combined
syntax and recall task. Of the syntax stimuli, 12 were
ungrammatical items that used word orderings that are
invalid according to Japanese syntax (OVS, VOS, VSO,
SVO). The frequency, order, and object-subject assignment
of each word were all balanced. All auditory training and
test stimuli were created by a native Japanese speaker.
With each sentence, complex scenes depicting cartoon
animals as the referents for the nouns were also presented,
engaging in the action indicated by the verb of each
sentence (hiding, lifting, pushing, or jumping). During
training, two such scenes were presented, one the target
scene and the other a distractor, to allow for the accrual of
word-referent mappings through the use of cross situational
statistics. During the syntax test, only the target scene was
presented. All stimuli were presented in E-Prime 2.0.
Procedure The experiment consisted of two training
blocks, and two test blocks. During training participants
heard a Japanese sentence while watching two scenes play
on the computer screen: one displaying the target, and the
other the foil. The foil scene varied from the target both in
terms of the nouns and actions depicted. Participants were
asked to judge to the best of their ability which scene the
sentence referred to. Unknown to the participant, the last
trials of training tested their knowledge of the nouns, verbs,
and markers of the language, using the same method by
varying the two scenes by just one property (e.g., only one
object was different, or only the action was different, or
only the subject/object roles were different). Following
training, 12 syntax test trials were administered, which
presented a sentence paired with a single scene. Participants
were told that the speaker of these sentences were learning
Japanese, and that their task was to repeat the speaker’s
sentence out loud (the recall measure), and then indicate
whether the speaker’s sentence sounded “good” or “funny”
(the forced-choice measure). While this task is slightly
different from the other 2AFC measures in this experiment,
in which participants choose between a target and a foil,
they both require participants to engage in reflection about
learned material.
After the conclusion of the first training block and syntax
test, the same procedure was completed once more, starting
with training and ending with another syntax test. Each
training block contained 4 marker test trials, 4 verb test
trials, and 6 noun test trials. Each syntax test block
contained 12 test trials: 6 grammatical, and 6 ungrammatical
sentences. No feedback about participants’ performance was
provided at any time during training or test.
Results and Discussion
2AFC results Following the methods employed by Walker
et al. (2017), data from both testing blocks were pooled for
the analyses. With the exception of noun learning (M=.54,
1859
SD=.20; t(48)=1.58, p=.12), scores on all 2AFC measures
were significantly above chance (Marker test: M=.57,
SD=.19; t(48)=2.80, p=.0075; Verb test: M=.60, SD=.14;
t(48)=5.52, p=.0075; Syntax test: M=.61, SD=.14;
t(48)=5.52, p=.0075). Thus, our results showed that
vocabulary and grammatical acquisition of natural language
structure can occur simultaneously. The lack of significant
noun learning, while inconsistent with the findings of
Walker et al. (2017), may be explained by the fact that the
nouns in this task were longer (containing more syllables),
which may have contributed to reduced learning. The only
2AFC tests that were significantly correlated with each
other were performance on the syntax and verb test,
r(47)=.39, p=.0065, which is consistent with the findings of
Walker et al. (2017).
Syntax recall results Participantsverbal responses were
transcribed by a coder blind to the study design, and were
scored against the targets, with a point given for each
syllable recalled in the correct serial position. Overall, no
effect of grammaticality was found on recall performance,
with participants recalling approximately equal numbers of
syllables in the correct order for both the grammatical
(M=73.08, SD=19.46) and ungrammatical items (M=72.14,
SD=18.57; t(48)= .60, p=.55). However, unlike the artificial
language stimuli, the natural language recall items vary in
the total number of syllables, ranging from 8 to 14 syllables.
A linear mixed effects model of the raw recall data revealed
that while grammaticality and string length had no effect
independently on recall performance, the interaction of
grammaticality and string length was significant,
t(1147)=2.78, p=.0055, with length relating to significant
detriments in recall scores for ungrammatical items, but not
grammatical ones. This suggests that learning the statistical
structure of the language stabilized recall of the grammatical
items independent of length, whereas the ungrammatical
items were more severely impacted by memory limitations.
Correlations between 2AFC and recall To investigate
the connection between the reflection- and processing-based
measures in the cross-situational learning task, correlations
were run between all 2AFC measures in Part 2 of the
experiment, and the recall difference scores. No significant
correlations were found between any of the 2AFC measures
and recall performance (all p>.14). These results may be
taken as further evidence that reflection- and processing-
based tests do not measure the same aspects of learning.
The relationship between artificial & natural
language learning
To determine the connection between non-adjacent
dependency learning in an artificial context to vocabulary
and grammar acquisition in a natural language context,
correlations were calculated between the data from Parts 1
and 2 of the experiment. These analyses were first
performed within task type, then between task types. We
predicted that the 2AFC measures from Part 1 would only
correlate with the 2AFC measures from Part 2, and that only
the recall measures from Parts 1 and 2 would be correlated.
Parts 1 & 2: Reflection-based measures
Correlations between all reflection-based (2AFC) measures
were performed. However, only two 2AFC measures were
correlated across the two parts of the experiment. First,
participants’ ability to segment words in the non-adjacent
dependency task positively correlated with their ability to
learn the nouns in the cross-situational learning task,
r(47)=.35, p=.0139. Second, generalization ability on the
non-adjacent dependency 2AFC task negatively correlated
with participants’ ability to pick up on the markers on the
cross-situational learning task, r(47)=-.28, p=.0496.
Parts 1 & 2: Processing-based measures
Partial correlations between the SICR and cross-situational
recall performance raw data, controlling for string length,
item type (experimental versus random for SICR, or
grammatical versus ungrammatical for the cross-situational
items) and repeated measures revealed that SICR and cross-
situational recall abilities significantly correlated with one
another. SICR segmentation (r(615)=.20, p<.0001)
demonstrated a stronger correlation to cross-situational
recall than did SICR generalization (r(561)=.14, p<.0009).
Parts 1 & 2: Between measure-types
Correlations between reflection- and processing-based tests
from Parts 1 and 2 of the experiment were performed. The
results revealed no significant correlations between task
types from the two parts of the experiment (Table 1).
Table 1: 2AFC and Recall Correlations
Marker
Test
Noun
Test
Syntax
Test
Verb
Test
2AFC
Seg.
2AFC
Gen.
SICR
Seg.
.02
.02
.12
.13
.31*
.26
SICR
Gen.
-.09
.08
-.07
.15
.12
.16
Cross-sit.
recall
-.22
.18
.09
-.02
.17
.15
General Discussion
Bridging artificial and natural language learning is an
important endeavor in the cognitive sciences. A first step to
stabilizing the link between in-lab observations and real-
world behavior may come from strengthening the
connection between the tasks used to test learning, and the
computations employed during learning. Here, we argue for
the role of statistically-based chunking as the computational
link between learning and testing.
Short-term memory recall is a robust indicator of long-
term learning (Jones & Macken, 2015; McCauley, Isbilen &
Christiansen, 2017), with the accrual of statistical
regularities over time aiding memory retention in the here-
and-now. The use of such recall tasks as a measure of in-lab
language learning is motivated by evidence supporting the
notion that chunking may play a key role in statistical
learning, and can account for language acquisition,
1860
processing, and production more broadly (Christiansen &
Chater, 2016). Our results strengthen this argument by
demonstrating that statistically-based chunking can also
account for the simultaneous learning and generalization of
non-adjacent dependencies: a complex, dynamic linguistic
structure. While our findings appear to support some degree
of separability between segmentation and generalization
skills (see also Frost & Monaghan, 2017), these abilities
also appear to inform one another, with segmentation
performance strongly predicting generalization ability.
While our study has important methodological and
theoretical implications for individual differences work, it
also has a number of limitations. First, although our cross-
situational paradigm simulates aspects of natural language
acquisition in the lab, it does not capture language
acquisition exactly as it occurs in the real world. Second,
while this language learning experiment implemented many
separate subcomponents of natural language by design, we
also acknowledge that these features changed the task
demands. Learning in the artificial language task involved
only segmentation and generalization of individual words,
whereas the cross-situational learning task incorporated
complex grammar, referents, and whole sentences. Stronger
correlations may have been observed if the structure of the
two different tasks were more similar (see Siegelman et al.,
2017, for discussion).
Methodologically, our results suggest that the empirical
disconnect between the learning targeted and the measures
used at test may influence the correlations observed between
artificial and natural language outcomes. While the
processes leveraged for both in-lab and real-world language
acquisition may be analogous, the similarity or dissimilarity
of the tasks used to measure learning – and the specific
computations each task relies on may obscure the
connection between the targeted cognitive processes.
Moreover, there appears to be substantial individual
variation in performance on these two kinds of tasks, as
evidence of learning on one measure does not necessarily
translate to high performance on the other. While both
reflection- and processing-based measures test learning, our
results suggest that they may test slightly different kinds of
knowledge: meta-cognitive reflections over what was
learned, versus processing-based facilitation from accrued
statistics.
Acknowledgments
We would like to thank Phoebe Ilevebare, Farrah Mawani,
Eleni Kohilakis, Olivia Wang, Dante Dahabreh, and Jake
Kolenda for their help collecting and coding data. This work
was in part supported by NSF GRFP (#DGE-1650441)
awarded to ESI, a Cornell Department of Psychology
Graduate Research Grant awarded to ESI, and by the
International Centre for Language and Communicative
Development (LuCiD) at Lancaster University, funded by
the Economic and Social Research Council (UK)
[ES/L008955/1].
References
Christiansen, M.H. (2018). Implicit-statistical learning: A tale
of two literatures. Topics in Cognitive Science.
Christiansen, M.H. & Chater, N. (2016). The Now-or-Never
bottleneck: A fundamental constraint on language.
Behavioral and Brain Sciences, 39, e62.
Conway, C.M., Bauernschmidt, A., Huang, S.S. & Pisoni,
D.B. (2010). Implicit statistical learning in language
processing: Word predictability is the key. Cognition, 114,
356-371.
Frost, R.L.A., & Monaghan, P. (2016). Simultaneous
segmentation and generalisation of non-adjacent
dependencies from continuous speech. Cognition, 147, 70-
74.
Frost, R.L.A., & Monaghan, P. (2017). Sleep-driven
computations in speech processing. PloS one, 12(1),
e0169538.
Jones, G. & Macken, B. (2015). Questioning short-term
memory and its measurement: Why digit span measures
long-term associative learning. Cognition, 144, 1-13.
Isbilen, E.S., McCauley, S.M., Kidd, E. & Christiansen, M.H.
(2017). Testing statistical learning implicitly: A novel
chunk-based measure of statistical learning. Proceedings of
the 39th Annual Conference of the Cognitive Science
Society. Austin, TX: Cognitive Science Society.
McCauley, S.M. & Christiansen, M.H. (2011). Learning
simple statistics for language comprehension and
production: The CAPPUCCINO model. Proceedings of the
33rd Annual Conference of the Cognitive Science. Austin,
TX: Cognitive Science Society.
McCauley, S.M., Isbilen, E.S. & Christiansen, M.H. (2017).
Chunking ability shapes sentence processing at multiple
levels of abstraction. Proceedings of the 39th Annual
Conference of the Cognitive Science Society. Austin, TX:
Cognitive Science Society.
Peña, M., Bonatti, L., Nespor, M., & Mehler, J. (2002).
Signal-driven computations in speech processing. Science,
298, 604-607.
Rebuschat, P., Ferrimand, H., & Monaghan, P. (2017). Age
effects in statistical learning of Japanese: Evidence from the
cross-situational learning paradigm. Talk presented at the
International Association for the Study of Child Language.
Saffran, J. R., Newport, E. L., & Aslin, R. N. (1996). Word
segmentation: The role of distributional cues. Journal of
memory and language, 35(4), 606-621.
Siegelman, N., Bogaerts, L., Christiansen, M. H. & Frost, R.
(2017). Towards a theory of individual differences in
statistical learning. Phil. Trans. R. Soc. B, 372(1711),
20160059.
Walker, N., Schoetensack, C., Monaghan, P., & Rebuschat, P.
Simultaneous acquisition of vocabulary and grammar in an
artificial language learning task. Proceedings of the 39th
Annual Conference of the Cognitive Science Society. Austin,
TX: Cognitive Science Society.
1861
... After the first two sets, recall performance became better in the experimental group than in the control group, indicating some degree of acquisition of the rules underlying the letter strings. More recently, Isbilen et al. (2018Isbilen et al. ( , 2020 used recall of rule-based strings as a measure of learning in a series of statistical learning studies. In their 2020 study, participants were first exposed to a (meaningless) artificial language composed of trisyllabic nonwords concatenated together in a continuous stream, following a study by Saffran et al. (1996). ...
... Participants were significantly more accurate in their recall of experimental items compared to random ones, indicating that they had acquired knowledge of the trisyllabic sequences in the language. Similar results were obtained by Isbilen et al. (2018) with an artificial language including nonadjacent dependencies. In the test phase of this study, Isbilen et al. also included generalization strings, that is, strings with trained dependencies but novel intervening syllables. ...
... At the same time, any test employed should avoid raising attention to the rule, for the purposes of demonstrating implicit learning. One way to do so would be to ask participants to recall new sentences where the rule is used in novel contexts, as previously done by Isbilen et al. (2018). Another way, which could provide even stronger evidence for rule knowledge in production, is to test participants by exploiting another aspect of human memory, namely, the false memory phenomenon (Deese, 1959;Roediger & McDermott, 1995). ...
Article
Full-text available
Second language implicit learning research has shown that a variety of linguistic features can be acquired without awareness. However, this research overwhelmingly uses comprehension tests to measure implicit learning. It remains unclear whether newly acquired implicit knowledge can also be recruited for production. To address this question, we developed a novel paradigm based on elicited recall and false memory that was used to both train participants and test their implicit knowledge in production, including generating new instances of the rule. Participants learned a semiartificial language containing a rule based on one in a natural language (the alternation between Czech spatial prepositions v and na). Participants who remained unaware of the rule, as assessed by verbal report, nevertheless were able to use it in a production test involving novel items, while believing that they were performing a cued recall test. Even without extensive training, newly acquired implicit knowledge can immediately be evident in production. A one‐page Accessible Summary of this article in non‐technical language is freely available in the Supporting Information online and at https://oasis‐database.org
... While such comparisons do permit assessment of preference for the overall structure, they require learners to use trained A and C items flexibly in a way that deviates from their knowledge of syllable position, which may affect performance. Indeed, using amended test items (trained dependencies with entirely novel intervening items), Frost and Monaghan (2016) demonstrated that adults can segment statistical nonadjacent dependencies and generalize them to novel grammatically consistent instances in the absence of additional information, such as pauses between words (see Isbilen, Frost, Monaghan, & Christiansen, 2018, for a replication of this effect). ...
... In the current study, these tasks were completed by all participants (within subjects). We show that the same learners can segment non-adjacencies from speech, and generalize them to new instances (see also Isbilen et al., 2018). In line with previous studies, performance on the segmentation task was higher than that seen for the generalization task (see Isbilen et al., 2018, for a comparable finding), and crucially performance on these tasks was significantly correlated for both language conditionsadding further support to the notion that they may be underpinned by similar mechanisms. ...
... We show that the same learners can segment non-adjacencies from speech, and generalize them to new instances (see also Isbilen et al., 2018). In line with previous studies, performance on the segmentation task was higher than that seen for the generalization task (see Isbilen et al., 2018, for a comparable finding), and crucially performance on these tasks was significantly correlated for both language conditionsadding further support to the notion that they may be underpinned by similar mechanisms. ...
Conference Paper
Full-text available
Achieving linguistic proficiency requires identifying words from speech, and discovering the constraints that govern the way those words are used. In a recent study of non-adjacent dependency learning, Frost and Monaghan (2016) demonstrated that learners may perform these tasks together, using similar statistical processes-contrary to prior suggestions. However, in their study, non-adjacent dependencies were marked by phonological cues (plosive-continuant-plosive structure), which may have influenced learning. Here, we test the necessity of these cues by comparing learning across three conditions; fixed phonology, which contains these cues, varied phonology, which omits them, and shapes, which uses visual shape sequences to assess the generality of statistical processing for these tasks. Participants segmented the sequences and generalized the structure in both auditory conditions, but learning was best when phonological cues were present. Learning was around chance on both tasks for the visual shapes group, indicating statistical processing may critically differ across domains.
... Although the fundamental logic of using memory tasks to detect learning can be traced to early work by Miller (1956) and Melton (1963), the paradigm termed statistically induced chunking recall has been developed in recent work by a group of researchers (Isbilen, Frost, Monaghan, & Christiansen, 2018;Isbilen, McCauley, Kidd, & Christiansen, 2017). Conway et al. (2010) similarly characterized such learning not only as implicit but also statistical. ...
... Encoding the input is a prerequisite for both implicit and explicit learning; we do not claim that PSTM itself is implicit, but rather that it underpins the inductive processes at work in statistical learning. Although baseline PSTM has itself been associated with learning vocabulary and grammar in children and adults in both L1 and L2 (N. C. Ellis & Sinclair, 1996;French & O'Brien, 2008;Gathercole, Service, Hitch, Adams, & Martin, 1999;Service, 1992;Verhagen & Leseman, 2016;Verhagen, Leseman, & Messer, 2015), performance on memory based ISL tasks has now also been linked both to native language processing (Conway et al., 2010;Misyak & Christiansen, 2012;Misyak, Christiansen, & Tomblin, 2010) and language learning itself (Isbilen et al., 2018;Isbilen et al., 2017;Kidd, 2012). ...
Article
Full-text available
We investigated the potential influence of implicit learning mechanisms on L2 morphosyntactic attainment by examining the relationship between age of onset (AoO), two cognitive abilities hypothesized to underlie implicit learning (phonological short‐term memory and implicit statistical learning), and performance on an auditory grammaticality judgment test (GJT). Participants were 71 Polish‐English long‐term bilinguals with a wide range of AoOs (1–35 years) who differed in their context of learning and use (immersed vs. instructed). In immersed learners, we observed a growing dissociation between performance on grammatical and ungrammatical sentences as AoO was delayed. This effect was attenuated in those with better phonological short‐term memory and statistical learning abilities and is consistent with a decline in the ability to learn from implicit negative evidence. In instructed learners, GJT performance was subject to additive effects of AoO and grammaticality and was not associated with either cognitive predictor, suggesting that implicit learning mechanisms were not involved.
... While some degree of self-correction was observed in the auditory version of the task, participants' verbal responses were not available for playback during the experiment, and they could only self-correct from memory alone. Thus, while some studies show that different measures of statistical learning are not always correlated with one another, even when they intend to measure learning of the same material (Arnon, 2019;Erickson, Kaschak, Thiessen, & Berry, 2016;Misyak & Christiansen, 2012;, the correlation observed between 2AFC and SICR may in part stem from this shared reflection-based component, as reflectionbased tasks tend to correlate more highly with other reflection-based tasks than with processing-based tasks (Isbilen, Frost, Monaghan, & Christiansen, 2018). ...
... This model also allows for the active prediction of upcoming elements using previously learned information, similar to natural language acquisition. Behaviorally, our results from using SICR to measure the statistical learning of non-adjacent dependencies demonstrate that chunking can also capture the learning and generalization of non-adjacent structures, and it is thus not limited to the acquisition of adjacent distributional patterns alone (Isbilen et al., 2018;E. S. Isbilen, R. L. A. Frost, P. Monaghan, & M. H. Christiansen, unpublished data). ...
Article
Full-text available
The computations involved in statistical learning have long been debated. Here, we build on work suggesting that a basic memory process, chunking, may account for the processing of statistical regularities into larger units. Drawing on methods from the memory literature, we developed a novel paradigm to test statistical learning by leveraging a robust phenomenon observed in serial recall tasks: that short‐term memory is fundamentally shaped by long‐term distributional learning. In the statistically induced chunking recall (SICR) task, participants are exposed to an artificial language, using a standard statistical learning exposure phase. Afterward, they recall strings of syllables that either follow the statistics of the artificial language or comprise the same syllables presented in a random order. We hypothesized that if individuals had chunked the artificial language into word‐like units, then the statistically structured items would be more accurately recalled relative to the random controls. Our results demonstrate that SICR effectively captures learning in both the auditory and visual modalities, with participants displaying significantly improved recall of the statistically structured items, and even recall specific trigram chunks from the input. SICR also exhibits greater test–retest reliability in the auditory modality and sensitivity to individual differences in both modalities than the standard two‐alternative forced‐choice task. These results thereby provide key empirical support to the chunking account of statistical learning and contribute a valuable new tool to the literature.
... For instance, in testing human adults and children there is frequently a distinction between explicit, reflection-based tasks for adult responses, such as alternative forced choice, or go/ no-go responses, and implicit, processing-based tasks such as head-turn preferences or looking times. These tasks may tap into different mechanisms, with processing-based tasks more effective for assessing processing-based learning, such as acquisition of grammatical structures (Christiansen, 2019;Frizelle, O'Neill, & Bishop, 2017;Isbilen et al., 2018). ...
... Further meta-analytical techniques can help determine the additional sources of information that might support such learning, such as use of reflection-versus processingbased test measures (Vuong et al., 2016). In order to measure the effect of learning on processing, rather than explicit decision-making based on the structures experienced by the learner, a task that probes processing is proposed to be more effective (Christiansen, 2019;Frizelle et al., 2017;Isbilen et al., 2018); however, in the present analysis there was no statistically reliable difference between the two. This may be a consequence of the comparatively large number of reflection-based effects (135) relative to processingbased effects (21) included in this analysis, or of the range of grammars that tend to be tested in AGL studies, a large number of studies use Reber-style (1967) grammars, where explicit testing may produce a similar magnitude of effects. ...
Article
Full-text available
Artificial grammar learning (AGL) has become an important tool used to understand aspects of human language learning and whether the abilities underlying learning may be unique to humans or found in other species. Successful learning is typically assumed when human or animal participants are able to distinguish stimuli generated by the grammar from those that are not at a level better than chance. However, the question remains as to what subjects actually learn in these experiments. Previous studies of AGL have frequently introduced multiple potential contributors to performance in the training and testing stimuli, but meta‐analysis techniques now enable us to consider these multiple information sources for their contribution to learning—enabling intended and unintended structures to be assessed simultaneously. We present a blueprint for meta‐analysis approaches to appraise the effect of learning in human and other animal studies for a series of artificial grammar learning experiments, focusing on studies that examine auditory and visual modalities. We identify a series of variables that differ across these studies, focusing on both structural and surface properties of the grammar, and characteristics of training and test regimes, and provide a first step in assessing the relative contribution of these design features of artificial grammars as well as species‐specific effects for learning.
... It could be used to complement existing offline measures and might help to overcome some of the psychometric shortcomings of offline measures, such as test interference. As a more implicit and direct measure (see also Isbilen et al., 2018), SOL might prove particularly useful for studying SL in young children (e.g., Arnon, 2019;Lammertink et al., 2019) or certain patient groups (e.g., amnesic patients; Schapiro et al., 2014) who have difficulty with the 2-AFC testing format. It also has a considerable advantage (over, for example, the RT-based measure in an SRT task) due to the many degrees of freedom in constructing the regular sequence (e.g., contrary to the SRT task, it is possible to have several sequences with transitional probabilities of 1). ...
Article
Regularity detection, or statistical learning, is regarded as a fundamental component of our cognitive system. To test the ability of human participants to detect regularity in a more ecological situation (i.e., mixed with random information), we used a simple letter-naming paradigm in which participants were instructed to name single letters presented one at a time on a computer screen. The regularity consisted of a triplet of letters that were systematically presented in that order. Participants were not told about the presence of this regularity. A variable number of random letters were presented between two repetitions of the regular triplet, making this paradigm similar to a Hebb repetition task. Hence, in this Hebb-naming task, we predicted that if any learning of the triplet occurred, naming times for the predictable letters in the triplet would decrease as the number of triplet repetitions increased. Surprisingly, across four experiments, detection of the regularity only occurred under very specific experimental conditions and was far from a trivial task. Our study provides new evidence regarding the limits of statistical learning and the critical role of contextual information in the detection (or not) of repeated patterns.
Article
Statistical learning, the ability to extract regularities from input (e.g., in language), is likely supported by learners’ prior expectations about how component units co‐occur. In this study, we investigated how adults’ prior experience with sublexical regularities in their native language influences performance on an empirical language learning task. Forty German‐speaking adults completed a speech repetition task in which they repeated eight‐syllable sequences from two experimental languages: one containing disyllabic words comprised of frequently occurring German syllable transitions (naturalistic words) and the other containing words made from unattested syllable transitions (non‐naturalistic words). The participants demonstrated learning from both naturalistic and non‐naturalistic stimuli. However, learning was superior for the naturalistic sequences, indicating that the participants had used their existing distributional knowledge of German to extract the naturalistic words faster and more accurately than the non‐naturalistic words. This finding supports theories of statistical learning as a form of chunking, whereby frequently co‐occurring units become entrenched in long‐term memory. A one‐page Accessible Summary of this article in non‐technical language is freely available in the Supporting Information online and at https://oasis‐database.org
Article
Regularity detection, or statistical learning, is regarded as a fundamental component of our cognitive system. To test the ability of human participants to detect regularity in a more ecological situation (i.e., mixed with random information), we used a simple letter-naming paradigm in which participants were instructed to name single letters presented one at a time on a computer screen. The regularity consisted of a triplet of letters that were systematically presented in that order. Participants were not told about the presence of this regularity. A variable number of random letters were presented between two repetitions of the regular triplet making this paradigm similar to a Hebb repetition task. Hence, in this Hebb-naming task, we predicted that if any learning of the triplet occurred, naming times for the predictable letters in the triplet would decrease as the number of triplet repetitions increased. Surprisingly, across four experiments, detection of the regularity only occurred under very specific experimental conditions and was far from a trivial task. Our study provides new evidence regarding the limits of statistical learning and the critical role of contextual information in the detection (or not) of repeated patterns.
Conference Paper
Full-text available
Attempts to connect individual differences in statistical learning with broader aspects of cognition have received considerable attention, but have yielded mixed results. A possible explanation is that statistical learning is typically tested using the two-alternative forced choice (2AFC) task. As a meta-cognitive task relying on explicit familiarity judgments, 2AFC may not accurately capture implicitly formed statistical computations. In this paper, we adapt the classic serial-recall memory paradigm to implicitly test statistical learning in a statistically-induced chunking recall (SICR) task. We hypothesized that artificial language exposure would lead subjects to chunk recurring statistical patterns, facilitating recall of words from the input. Experiment 1 demonstrates that SICR offers more fine-grained insights into individual differences in statistical learning than 2AFC. Experiment 2 shows that SICR has higher test-retest reliability than that reported for 2AFC. Thus, SICR offers a more sensitive measure of individual differences, suggesting that basic chunking abilities may explain statistical learning.
Conference Paper
Full-text available
Several recent empirical findings have reinforced the notion that a basic learning and memory skill—chunking—plays a fundamental role in language processing. Here, we provide evidence that chunking shapes sentence processing at multiple levels of linguistic abstraction, consistent with a recent theoretical proposal by Christiansen and Chater (2016). Individual differences in chunking ability at two different levels is shown to predict on-line sentence processing in separate ways: i) phonological chunking ability, as assessed by a variation on the non-word repetition task, predicts processing of complex sentences featuring phonological overlap; ii) multiword chunking ability, as assessed by a variation on the serial recall task, is shown to predict reading times for sentences featuring long-distance number agreement with locally distracting number-marked nouns. Together, our findings suggest that individual differences in chunking ability shape language processing at multiple levels of abstraction, consistent with the notion of language acquisition as learning to process.
Article
Full-text available
Acquiring language requires segmenting speech into individual words, and abstracting over those words to discover grammatical structure. However, these tasks can be conflicting—on the one hand requiring memorisation of precise sequences that occur in speech, and on the other requiring a flexible reconstruction of these sequences to determine the grammar. Here, we examine whether speech segmentation and generalisation of grammar can occur simultaneously—with the conflicting requirements for these tasks being over-come by sleep-related consolidation. After exposure to an artificial language comprising words containing non-adjacent dependencies, participants underwent periods of consolidation involving either sleep or wake. Participants who slept before testing demonstrated a sustained boost to word learning and a short-term improvement to grammatical generalisation of the non-adjacencies, with improvements after sleep outweighing gains seen after an equal period of wake. Thus, we propose that sleep may facilitate processing for these conflicting tasks in language acquisition, but with enhanced benefits for speech segmentation.
Article
Full-text available
In recent years, statistical learning (SL) research has seen a growing interest in tracking individual performance in SL tasks, mainly as a predictor of linguistic abilities. We review studies from this line of research and outline three presuppositions underlying the experimental approach they employ: (i) that SL is a unified theoretical construct; (ii) that current SL tasks are interchangeable, and equally valid for assessing SL ability; and (iii) that performance in the standard forced-choice test in the task is a good proxy of SL ability. We argue that these three critical presuppositions are subject to a number of theoretical and empirical issues. First, SL shows patterns of modality- and informational-specificity, suggesting that SL cannot be treated as a unified construct. Second, different SL tasks may tap into separate sub-components of SL that are not necessarily interchangeable. Third, the commonly used forced-choice tests in most SL tasks are subject to inherent limitations and confounds. As a first step, we offer a methodological approach that explicitly spells out a potential set of different SL dimensions, allowing for better transparency in choosing a specific SL task as a predictor of a given linguistic outcome. We then offer possible methodological solutions for better tracking and measuring SL ability. Taken together, these discussions provide a novel theoretical and methodological approach for assessing individual differences in SL, with clear testable predictions. This article is part of the themed issue ‘New frontiers for statistical learning in the cognitive sciences’.
Article
Full-text available
Language learning requires mastering multiple tasks, including segmenting speech to identify words, and learning the syntactic role of these words within sentences. A key question in language acquisition research is the extent to which these tasks are sequential or successive, and consequently whether they may be driven by distinct or similar computations. We explored a classic artificial language learning paradigm, where the language structure is defined in terms of non-adjacent dependencies. We show that participants are able to use the same statistical information at the same time to segment continuous speech to both identify words and to generalise over the structure, when the generalisations were over novel speech that the participants had not previously experienced. We suggest that, in the absence of evidence to the contrary, the most economical explanation for the effects is that speech segmentation and grammatical generalisation are dependent on similar statistical processing mechanisms.
Article
Full-text available
Traditional accounts of verbal short-term memory explain differences in performance for different types of verbal material by reference to inherent characteristics of the verbal items making up memory sequences. The role of previous experience with sequences of different types is ostensibly controlled for either by deliberate exclusion or by presenting multiple trials constructed from different random permutations. We cast doubt on this general approach in a detailed analysis of the basis for the robust finding that short-term memory for digit sequences is superior to that for other sequences of verbal material. Specifically, we show across four experiments that this advantage is not due to inherent characteristics of digits as verbal items, nor are individual digits within sequences better remembered than other types of individual verbal items. Rather, the advantage for digit sequences stems from the increased frequency, compared to other verbal material, with which digits appear in random sequences in natural language, and furthermore, relatively frequent digit sequences support better short-term serial recall than less frequent ones. We also provide corpus-based computational support for the argument that performance in a short-term memory setting is a function of basic associative learning processes operating on the linguistic experience of the rememberer. The experimental and computational results raise questions not only about the role played by measurement of digit span in cognition generally, but also about the way in which long-term memory processes impact on short-term memory functioning. Copyright © 2015 The Authors. Published by Elsevier B.V. All rights reserved.
Article
Full-text available
Memory is fleeting. New material rapidly obliterates previous material. How then can the brain deal successfully with the continual deluge of linguistic input? We argue that, to deal with this “Now-or-Never” bottleneck, the brain must compress and recode linguistic input as rapidly as possible. This observation has strong implications for the nature of language processing: (i) the language system must “eagerly” recode and compress linguistic input; (ii) as the bottleneck recurs at each new representational level, the language system must build a multi-level linguistic representation; and (iii) it must deploy all available information predictively to ensure that local linguistic ambiguities are dealt with “Right-First-Time;” once the original input is lost, there is no way for the language system to recover. This is “Chunk-and-Pass” processing. Similarly, language learning must also occur in the here-and-now. This implies that language acquisition is learning to process, rather than inducing a grammar. Moreover, this perspective provides a cognitive foundation for grammaticalization and other aspects of language change. Chunk-and-Pass processing also helps explain a variety of core properties of language, including its multi-level representational structure and duality of patterning. This approach promises to create a direct relationship between psycholinguistics and linguistic theory. More generally, we outline a framework within which to integrate often disconnected inquiries into language processing, language acquisition, and language change and evolution.
Conference Paper
Full-text available
Whether the input available to children is sufficient to explain their ability to use language has been the subject of much theoretical debate in cognitive science. Here, we present a simple, developmentally motivated computational model that learns to comprehend and produce language when exposed to child-directed speech. The model uses backward transitional probabilities to create an inventory of ‘chunks’ consisting of one or more words. Language comprehension is approximated in terms of shallow parsing of adult speech and production as the reconstruction of the child’s actual utterances. The model functions in a fully incremental, on-line fashion, has broad cross-linguistic coverage, and is able to fit child data from Saffran’s (2002) statistical learning study. Moreover, word-based distributional information is found to be more useful than statistics over word classes. Together, these results suggest that much of children’s early linguistic behavior can be accounted for in a usage-based manner using distributional statistics.
Article
Full-text available
One of the infant's first tasks in language acquisition is to discover the words embedded in a mostly continuous speech stream. This learning problem might be solved by using distributional cues to word boundaries—for example, by computing the transitional probabilities between sounds in the language input and using the relative strengths of these probabilities to hypothesize word boundaries. The learner might be further aided by language-specific prosodic cues correlated with word boundaries. As a first step in testing these hypotheses, we briefly exposed adults to an artificial language in which the only cues available for word segmentation were the transitional probabilities between syllables. Subjects were able to learn the words of this language. Furthermore, the addition of certain prosodic cues served to enhance performance. These results suggest that distributional cues may play an important role in the initial word segmentation of language learners.
Article
Implicit learning and statistical learning are two contemporary approaches to the long‐standing question in psychology and cognitive science of how organisms pick up on patterned regularities in their environment. Although both approaches focus on the learner's ability to use distributional properties to discover patterns in the input, the relevant research has largely been published in separate literatures and with surprisingly little cross‐pollination between them. This has resulted in apparently opposing perspectives on the computations involved in learning, pitting chunk‐based learning against probabilistic learning. In this paper, I trace the nearly century‐long historical pedigree of the two approaches to learning and argue for their integration under the heading of “implicit statistical learning.” Building on basic insights from the memory literature, I sketch a framework for statistically based chunking that aims to provide a unified basis for understanding implicit statistical learning.