Mark Andrews’s research while affiliated with Nottingham Trent University and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (16)


Reconciling Embodied and Distributional Accounts of Meaning in Language
  • Literature Review

June 2014

·

211 Reads

·

132 Citations

Topics in Cognitive Science

Mark Andrews

·

Stefan Frank

·

Over the past 15 years, there have been two increasingly popular approaches to the study of meaning in cognitive science. One, based on theories of embodied cognition, treats meaning as a simulation of perceptual and motor states. An alternative approach treats meaning as a consequence of the statistical distribution of words across spoken and written language. On the surface, these appear to be opposing scientific paradigms. In this review, we aim to show how recent cross-disciplinary developments have done much to reconcile these two approaches. The foundation to these developments has been the recognition that intralinguistic distributional and sensory-motor data are interdependent. We describe recent work in philosophy, psychology, cognitive neuroscience, and computational modeling that are all based on or consistent with this conclusion. We conclude by considering some possible directions for future research that arise as a consequence of these developments.


Figure 1. 
Figure 2. 
Figure 3. Neighborhood cliques of kill according to the experience, language and combined models
Figure 4. 
Toward a theory of semantic representation
  • Article
  • Full-text available

June 2014

·

5,643 Reads

·

328 Citations

Language and Cognition

We present an account of semantic representation that focuses on distinct types of information from which word meanings can be learned. In particu-lar, we argue that there are at least two major types of information from which we learn word meanings. The first is what we call experiential infor-mation. This is data derived both from our sensory-motor interactions with the outside world, as well as from our experience of own inner states, par-ticularly our emotions. The second type of information is language-based. In particular, it is derived from the general linguistic context in which words appear. The paper spells out this proposal, summarizes research supporting this view and presents new predictions emerging from this framework.

Download

The Representation of Abstract Words: What Matters? Reply to Comment on

February 2013

·

142 Reads

·

37 Citations

·

Stavroula Kousta

·

David Vinson

·

[...]

·

Elena Del Campo

In Kousta, Vigliocco, Vinson, Andrews, and Del Campo (2011), we presented an embodied theory of semantic representation, which crucially included abstract concepts as internally embodied via affective states. Paivio (2013) took issue with our treatment of dual coding theory, our reliance on data from lexical decision, and our theoretical proposal. Here, we address these different issues and clarify how our findings offer a way to move forward in the investigation of how abstract concepts are represented.


The Representation of Abstract Words: Why Emotion Matters

December 2010

·

4,616 Reads

·

760 Citations

Although much is known about the representation and processing of concrete concepts, knowledge of what abstract semantics might be is severely limited. In this article we first address the adequacy of the 2 dominant accounts (dual coding theory and the context availability model) put forward in order to explain representation and processing differences between concrete and abstract words. We find that neither proposal can account for experimental findings and that this is, at least partly, because abstract words are considered to be unrelated to experiential information in both of these accounts. We then address a particular type of experiential information, emotional content, and demonstrate that it plays a crucial role in the processing and representation of abstract concepts: Statistically, abstract words are more emotionally valenced than are concrete words, and this accounts for a residual latency advantage for abstract words, when variables such as imageability (a construct derived from dual coding theory) and rated context availability are held constant. We conclude with a discussion of our novel hypothesis for embodied abstract semantics.


The Hidden Markov Topic Model: A Probabilistic Model of Semantic Representation

January 2010

·

213 Reads

·

52 Citations

Topics in Cognitive Science

In this paper, we describe a model that learns semantic representations from the distributional statistics of language. This model, however, goes beyond the common bag-of-words paradigm, and infers semantic representations by taking into account the inherent sequential nature of linguistic data. The model we describe, which we refer to as a Hidden Markov Topics model, is a natural extension of the current state of the art in Bayesian bag-of-words models, that is, the Topics model of Griffiths, Steyvers, and Tenenbaum (2007), preserving its strengths while extending its scope to incorporate more fine-grained linguistic information.


Integrating Experiential and Distributional Data to Learn Semantic Representations

July 2009

·

1,261 Reads

·

431 Citations

The authors identify 2 major types of statistical data from which semantic representations can be learned. These are denoted as experiential data and distributional data. Experiential data are derived by way of experience with the physical world and comprise the sensory-motor data obtained through sense receptors. Distributional data, by contrast, describe the statistical distribution of words across spoken and written language. The authors claim that experiential and distributional data represent distinct data types and that each is a nontrivial source of semantic information. Their theoretical proposal is that human semantic representations are derived from an optimal statistical combination of these 2 data types. Using a Bayesian probabilistic model, they demonstrate how word meanings can be learned by treating experiential and distributional data as a single joint distribution and learning the statistical structure that underlies it. The semantic representations that are learned in this manner are measurably more realistic—as verified by comparison to a set of human-based measures of semantic representation—than those available from either data type individually or from both sources independently. This is not a result of merely using quantitatively more data, but rather it is because experiential and distributional data are qualitatively distinct, yet intercorrelated, types of data. The semantic representations that are learned are based on statistical structures that exist both within and between the experiential and distributional data types.


Integrating attributional and distributional information in a probabilistic model of meaning representation

January 2005

·

85 Reads

·

5 Citations

In this paper we present models of how meaning is represented in the brain/mind, based upon the assumption that children develop meaning representations for words using two main sources of information: information derived from their concrete experience with objects and events in the world (which we refer to as attributional information) and information implic-itly derived from exposure to language (which we refer to as distributional information). In the first part of the paper we present a model developed using self-organising maps (SOMs) starting from speaker-generated features (properties that speakers considered to be important in defining and describing the meaning of a word). This model captures meaning similarity between words based solely upon attributional information and has been shown to be success-ful in predicting a number of behavioural semantic effects. In the second part of the paper, we present a probabilistic model that goes beyond attributional information alone, integrating this information with distributional information derived from text corpora. The ability of this integrated model to learn semantic relationships is demonstrated with reference to comparable probabilistic models that use only attributional or distributional information.


The role of attributional and distributional information in semantic representation

January 2005

·

136 Reads

·

12 Citations

In recent studies of semantic representation, two distinct sources of information from which we can learn word meanings have been described. We refer to these as at-tributional and distributional information sources. At-tributional information describes the attributes or fea-tures associated with referents of words, and is acquired from our interactions with the world. Distributional in-formation describes the distribution of words across dif-ferent linguistic contexts, and is acquired through our use of language. While previous work has concentrated on the role of one source, to the exclusion of the other, in this paper we study the role of both sources in combi-nation. We describe a general framework based on prob-abilistic generative models for modelling both sources of information, and how they can be integrated to learn se-mantic representation. We provide examples comparing the learned structures of each of three models: attri-butional information alone, distributional information alone, and both sources combined.


Sexual and Asexual Paradigms in Evolution: The Implications for Genetic Algorithms

June 2004

·

16 Reads

Lecture Notes in Computer Science

In this paper, we generalize the models used by MacKay [1] in his analysis of evolutionary strategies that are based on sexual, rather than asexual, reproduction methods. This analysis can contribute to the understanding of the relative power of genetic algorithms over search methods based upon stochastic hill-climbing, e.g. [2], [3].


Learning and Inference in Hidden Stack Automata

11 Reads

In this paper, we describe the problems of learning and inference in Hidden Stack Au-tomata. These are generalizations of Hidden Markov models (HMM), whereby the finite-state HMM now has access to a pushdown stack memory. Providing learning and infer-ence algorithms is greatly facilitated by de-scribing the system in terms of a dynamical Bayesian network that couples a HMM and a piecewise-linear dynamical system. This en-semble exactly implements a pushdown au-tomaton, with the dynamical system act-ing as the unbounded stack memory for the HMM. The contents of the stack are coded numerically as points in the state-space of the dynamical system, while operations on the stack contents are performed by way of its dynamics. We provide a sequential Monte-Carlo method for the inference of the state-space trajectories given a sequence of observations. This provides us with other-wise intractable posterior estimates of the stack-contents. From this we can derive an expectation-maximization (EM) algorithm to estimate the parameters of the system given our training data. We provide examples of learning and inference in these Hidden Stack Automata, and compare their performance to HMMs.


Citations (12)


... Compared to concrete concepts (e.g., cup or mountain), which are experienced by exteroceptive modalities, abstract concepts (e.g., justice or democracy) are by definition intangible, and therefore could hardly be represented by the sensorimotor system (Barsalou, 2003a;Dove, 2009;Mahon & Caramazza, 2008). Yet, this critique failed to account for the diversity of experiences: Abstract concepts can be grounded in interoception and meta-cognitive states (e.g., the experience of recognizing something as true; Barsalou & Wiemer-Hastings, 2005;Connell et al., 2018;Prinz, 2012), as well as emotions (Kousta et al., 2011;Vigliocco et al., 2009). Later, the role of language as a means of representing concepts was also emphasized (Dove, 2011;Louwerse, 2011), as well as contributions from the social world (Borghi & Cimatti, 2009;Reinboth & Farkaš, 2022), and the role of predictive processing (Borghi et al., 2019;Van Elk & Bekkering, 2018). ...

Reference:

Issues in Grounded Cognition and How to Solve Them – the Minimalist Account
Toward a theory of semantic representation

Language and Cognition

... Initially, distributional semantics contrasted with embodied views (e.g., Glenberg & Robertson, 2000) which posited that all semantic content was fundamentally rooted in corporeal experiences. Still, in recent years, various models have been proposed, seeking to combine embodied and distributional approaches, i.e., treating language as a source where embodied experiences are encoded (Andrews et al., 2014;Günther et al., 2019;Louwerse, 2018). In this perspective, linguistic structures encapsulate embodied experiences accumulated through generational transmission, thereby providing an indirect foundation for semantic grounding. ...

Reconciling Embodied and Distributional Accounts of Meaning in Language
  • Citing Article
  • June 2014

Topics in Cognitive Science

... The concreteness effect, as mentioned in "Introduction" section, denotes an advantage that concrete words have over abstract words in various tasks, like reading, processing, recalling, naming, and recognizing (e.g., Borghi et al. 2017;Villani et al. 2019). Empirical evidence about the "concreteness effect" piled up in the past decades, while the origins of such effect and how it works in the human brain have been subjects of lively debates (e.g., Vigliocco et al. 2011). A component that muddies the waters of this theoretical debate is that empirical evidence supporting the concreteness effect is typically built on experiments based on stimuli that have been selected based on concreteness ratings, which are human judgments provided on Likert scales, about the perceived concreteness of given prompts (e.g., Brysbaert et al. 2014). ...

The Representation of Abstract Words: What Matters? Reply to Comment on

... This suggests that, to a large extent, meaning is encoded redundantly in both featural and distributional information, and the relationship between these two sources of data is complementary. Andrews et al. (2005, 2007, 2009) treated featural information and distributional information as joint distributions to be learned by a Bayesian model. They have demonstrated that the representations produced by this model are better able to reproduce behavioral data than are models that include only one of the two sources of information or treat the two data sources independently. ...

Integrating attributional and distributional information in a probabilistic model of meaning representation

... The idea of these models us to use hierarchical probabilistic graphical models to capture information about the topic of dierent documents. While the original LDA model uses a unigram model as the core language model, more recent implementation have tried to make use of more complex models such as hidden Markov models [3] or even syntactic parsing (a) unigram (b) mixture (c) pLSI trees [11]. In chapter 5 we present a hierarchical model for selecting the readout that also tries to capture high level features. ...

Learning semantic representations with hidden markov topics models
  • Citing Article

... The two theoretical approaches do not necessarily exclude one another since it is conceivable that our abstract knowledge exploits both sources mentioned above. According to this integrative point of view (Andrews et al., 2005Andrews et al., , 2007), both the attributive and distributive properties of words play an important role in symbol grounding. Attributive properties are non-linguistic physical attributes associated with a word, while distributive factors refer to common occurrences of a word with other linguistic elements. ...

Evaluating the contribution of intra-linguistic and extra-linguistic data to the structure of human semantic representations
  • Citing Article

... Such studies, however, have concentrated on one of these sources of information, often neglecting the other. More recently, evidence from machine learning has showed that models integrating both featural and distributional information can outperform featural-or distributional-only models (Andrews et al., 2005Andrews et al., , 2007Andrews et al., , 2009). For instance, Andrews et al. (2007) trained three Bayesian models using either a combination of both featural and distributional representations, or featural or distributional representations alone. ...

The role of attributional and distributional information in semantic representation

... Results from studies investigating the capacity for WA data to generate semantic similarity ratings in this manner (e.g. Steyvers, Shiffrin, & Nelson, 2004;Andrews, Vinson, & Vigliocco, 2008;Van Rensbergen, De Deyne, & Storms, 2016) have been impressive: such ratings have been shown to correlate with human judgements of semantic similarity, and to explain variance on psycholinguistic tasks such as lexical decision (Steyvers et al., 2004;De Deyne, Navarro, & Storms, 2012). These results suggest that networks built from WA data are not only suitable for use as simple estimates of semantic similarity for psycholinguistic experiments, but also that they capture important aspects of lexical processing. ...

Inferring a Probabilistic Model of Semantic Memory from Word Association Norms
  • Citing Article

... In this so-called bag-of-words assumption, the linguistic context of any given word is defined by which words co-occur with it and with what frequency. This assumption disregards sequential and syntactic information, but the sequential order in which words occur, the argument structure, and general syntactic relationships within sentences all provide important information about the meaning of the words, consequently limiting the extent to which semantic information is extracted from the text (Andrews & Vigliocco, 2010). The second approach is motivated by the so-called distributional hypothesis (Harris, 1954), which proposes that the meaning of a word can be derived from the linguistic contexts in which it occurs. ...

The Hidden Markov Topic Model: A Probabilistic Model of Semantic Representation
  • Citing Article
  • January 2010

Topics in Cognitive Science