Chris R. Sims’s research while affiliated with Rensselaer Polytechnic Institute and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (44)


Efficient Visual Representations for Learning and Decision Making
  • Article
  • Full-text available

September 2024

·

182 Reads

·

3 Citations

Psychological Review

·

Chris R. Sims

The efficient representation of visual information is essential for learning and decision making due to the complexity and uncertainty of the world, as well as inherent constraints on the capacity of cognitive systems. We hypothesize that biological agents learn to efficiently represent visual information in a manner that balances performance across multiple potentially competing objectives. In this article, we examine two such objectives: storing information in a manner that supports accurate recollection (maximizing veridicality) and in a manner that facilitates utility-based decision making (maximizing behavioral utility). That these two objectives may be in conflict is not immediately obvious. Our hypothesis suggests that neither behavior nor representation formation can be fully understood by studying either in isolation, with information processing constraints exerting an overarching influence. Alongside this hypothesis we develop a computational model of representation formation and behavior motivated by recent methods in machine learning and neuroscience. The resulting model explains both the beneficial aspects of human visual learning, such as fast acquisition and high generalization, as well as the biases that result from information constraints. To test this model, we developed two experimental paradigms, in decision making and learning, to evaluate how well the model’s predictions match human behavior. A key feature of the proposed model is that it predicts the occurrence of commonly found biases in human decision making, resulting from the desire to form efficient representations of visual information that are useful for behavioral goals in learning and decision making and optimized under an information processing constraint.

Download

Fig. 2 Example trajectories of exponential change in DR and RB, and their implications for observed behavior (i.e., Response Time and Accuracy). Characteristic effects can be observed (e.g., accuracy [b] increases with increasing DR [c] or increasing RB [d]; RT [a] decreases with increasing DR or decreasing RB), although the strengths of specific links may vary across the different levels of stimulus coherences. The values shown here were the fixed-effect posterior distribution estimated values and 95% CI from models reported in the Results, evaluated at the median stimulus coherence for the sole purpose of illustration (see also Supplementary Note section Best-fitting model summary output and Tables 1-5).
Bayes Factor (base-3 log) comparisons of models using 15 runs of bridge sampling, with the most equivocal being reported.
Multiple timescales of learning indicated by changes in evidence-accumulation processes during perceptual decision-making

June 2023

·

75 Reads

·

11 Citations

npj Science of Learning

Aaron Cochrane

·

Chris R Sims

·

Vikranth R Bejjanki

·

[...]

·

Evidence accumulation models have enabled strong advances in our understanding of decision-making, yet their application to examining learning has not been common. Using data from participants completing a dynamic random dot-motion direction discrimination task across four days, we characterized alterations in two components of perceptual decision-making (Drift Diffusion Model drift rate and response boundary). Continuous-time learning models were applied to characterize trajectories of performance change, with different models allowing for varying dynamics. The best-fitting model included drift rate changing as a continuous, exponential function of cumulative trial number. In contrast, response boundary changed within each daily session, but in an independent manner across daily sessions. Our results highlight two different processes underlying the pattern of behavior observed across the entire learning trajectory, one involving a continuous tuning of perceptual sensitivity, and another more variable process describing participants' threshold of when enough evidence is present to act.


Figure 1: Example of the RLβ-VAE model forming a reconstruction and predicted reward.
Figure 2: Left: Model pre-training reconstruction loss by training epoch, lower is better, color indicates latent dimension size. Middle: Contextual bandit training for 1000 runs of model accuracy by trail means (dots) are fit to a logarithmic function (lines). Right: Representation difference in mean-squared error between images containing hats, glasses, and both, compared to wearing neither. The left column of Figure 2 compares reconstruction loss by pre-training epoch. These results demonstrate a lower end of training reconstruction accuracy from models with smaller latent spaces. While these small latent dimensions are useful for quick hypothesis generation, they make accurate reconstruction of stimuli more difficult due to the tight information-bottleneck imposed on the model.
Learning in Factored Domains with Information-Constrained Visual Representations

March 2023

·

16 Reads

Humans learn quickly even in tasks that contain complex visual information. This is due in part to the efficient formation of compressed representations of visual information, allowing for better generalization and robustness. However, compressed representations alone are insufficient for explaining the high speed of human learning. Reinforcement learning (RL) models that seek to replicate this impressive efficiency may do so through the use of factored representations of tasks. These informationally simplistic representations of tasks are similarly motivated as the use of compressed representations of visual information. Recent studies have connected biological visual perception to disentangled and compressed representations. This raises the question of how humans learn to efficiently represent visual information in a manner useful for learning tasks. In this paper we present a model of human factored representation learning based on an altered form of a β\beta-Variational Auto-encoder used in a visual learning task. Modelling results demonstrate a trade-off in the informational complexity of model latent dimension spaces, between the speed of learning and the accuracy of reconstructions.



A Beta-Variational Auto-Encoder Model of Human Visual Representation Formation in Utility-Based Learning

December 2022

·

29 Reads

·

1 Citation

Journal of Vision

The human brain is capable of forming informationally constrained representations of complex visual stimuli in order to achieve its behavioural goals, such as utility-based learning. Recently, methods borrowed from machine learning have demonstrated a close connection between the mechanisms of visual representation formation in primate brains with the latent representations formed by Beta-Variational Auto-Encoders (Beta-VAEs). While auto-encoder models capture some aspects of visual representations, they fail to explain how visual representations are adapted in a task-directed manner. We developed a model of visual representation formation in learning environments based on a modified Beta-VAE model that simultaneously learns the task-specific utility of visual information. We hypothesized that humans update their visual representations as they learn which visual features are associated with utility in learning tasks. To test this hypothesis, we applied the proposed model onto the data from a visual contextual bandit learning task [Niv et al. 2015; J. Neuroscience]. The experiment involved humans (N=22) learning the utility associated with 9 possible visual features (3 colors, shapes or textures). Critically, our model takes in as input the same visual information that is presented to participants, instead of the hand-crafted features typically used to model human learning. A comparison of predictive accuracy between our proposed model and models using hand-crafted features demonstrated a similar correlation to human learning. These results show that representations formed by our Beta-VAE based model can predict human learning from complex visual information. Additionally, our proposed model makes predictions of how visual representations adapt during human learning in a utility-based task. Further, we performed a comparison of our proposed model across a range of parameters such as information-constraint, utility-weight, and number of training steps between predictions. Results from this comparison give insight into how the human brain adjusts its visual representation formation during learning.


Figure 2: Left: Model pre-training reconstruction loss by training epoch, lower is better, color indicates latent dimension size. Middle: Contextual bandit training for 1000 runs of model accuracy by trail means (dots) are fit to a logarithmic function (lines). Right: Representation difference in mean-squared error between images containing hats, glasses, and both, compared to wearing neither. The left column of Figure 2 compares reconstruction loss by pre-training epoch. These results demonstrate a lower end of training reconstruction accuracy from models with smaller latent spaces. While these small latent dimensions are useful for quick hypothesis generation, they make accurate reconstruction of stimuli more difficult due to the tight information-bottleneck imposed on the model.
Learning in Factored Domains with Information-Constrained Visual Representations

December 2022

·

57 Reads

·

6 Citations

Humans learn quickly even in tasks that contain complex visual information. This is due in part to the efficient formation of compressed representations of visual information, allowing for better generalization and robustness. However, compressed representations alone are insufficient for explaining the high speed of human learning. Reinforcement learning (RL) models that seek to replicate this impressive efficiency may do so through the use of factored representations of tasks. These informationally simplistic representations of tasks are similarly motivated as the use of compressed representations of visual information. Recent studies have connected biological visual perception to disentangled and compressed representations. This raises the question of how humans learn to efficiently represent visual information in a manner useful for learning tasks. In this paper we present a model of human factored representation learning based on an altered form of a β-Variational Auto-encoder used in a visual learning task. Modelling results demonstrate a trade-off in the informational complexity of model latent dimension spaces, between the speed of learning and the accuracy of reconstructions.


Figure 2
Multiple timescales of learning within a single task: Continuous-time changes in evidence-accumulation processes during perceptual decision-making

September 2022

·

63 Reads

Drift Diffusion Models have enabled strong advances in our understanding of decision-making, yet their application to examining learning has not been common. Using data from participants completing a dynamic random dot-motion direction discrimination task across four days, we characterized alterations in two components of perceptual decision-making (drift rate and response boundary). Continuous-time learning DDMs were applied to characterize trajectories of performance change, with different models allowing for varying dynamics. The best-fitting model included drift rate changing as a continuous, exponential function of cumulative trial number. In contrast, response boundary changed within each daily session, but in an independent manner across daily sessions. Our results highlight two different processes underlying the pattern of behavior observed across the entire learning trajectory, one involving a continuous tuning of perceptual sensitivity, and another more variable process describing participants’ threshold of when enough evidence is present to act.


Modeling Human Reinforcement Learning with Disentangled Visual Representations

June 2022

·

387 Reads

·

7 Citations

Humans are able to learn about the visual world with a remarkable degree of generality and robustness, in part due to attention mechanisms which focus limited resources onto relevant features. Deep learning models that seek to replicate this feature of human learning can do so by optimizing a so-called "disentanglement objective", which encourages representations that factorize stimuli into separable feature dimensions [4]. This objective is achieved by methods such as the β-Variational Autoencoder (β-VAE), which has demonstrated a strong correspondence to neural activity in biological visual representation formation [5]. However, in the β-VAE method, learned visual representations are not influenced by the utility of information, but are solely learned in an unsupervised fashion. In contrast to this, humans exhibit generalization of learning through acquired equivalence of visual stimuli associated with similar outcomes [7]. The question of how humans combine utility-based and unsupervised learning in the formation of visual representations is therefore unanswered. The current paper seeks to address this question by developing a modified β-VAE model which integrates both unsupervised learning and reinforcement learning. This model is trained to produce both psychological representations of visual information as well as predictions of utility based on these representations. The result is a model that predicts the impact of changing utility on visual representations. Our model demonstrates a high degree of predictive accuracy of human visual learning in a contextual multi-armed bandit learning task [8]. Importantly, our model takes as input the same complex visual information presented to participants, instead of relying on hand-crafted features. These results provide further support for disentanglement as a plausible learning objective for visual representation formation by demonstrating their usefulness in learning tasks that rely on attention mechanisms.


Conceptual knowledge shapes visual working memory for complex visual information

May 2022

·

189 Reads

·

3 Citations

Human visual working memory (VWM) is a memory store people use to maintain the visual features of objects and scenes. Although it is obvious that bottom-up information influences VWM, the extent to which top-down conceptual information influences VWM is largely unknown. We report an experiment in which groups of participants were trained in one of two different categories of geologic faults (left/right lateral, or normal/reverse faults), or received no category training. Following training, participants performed a visual change detection task in which category knowledge was irrelevant to the task. Participants were more likely to detect a change in geologic scenes when the changes crossed a trained categorical distinction (e.g., the left/right lateral fault boundary), compared to within-category changes. In addition, participants trained to distinguish left/right lateral faults were more likely to detect changes when the scenes were mirror images along the left/right dimension. Similarly, participants trained to distinguish normal/reverse faults were more likely to detect changes when scenes were mirror images along the normal/reverse dimension. Our results provide direct empirical evidence that conceptual knowledge influences VWM performance for complex visual information. An implication of our results is that cognitive scientists may need to reconceptualize VWM so that it is closer to “conceptual short-term memory”.


Fig. 2. Structure of a capacity-limited DAC model. Double arrows indicate the presence of an information-theoretic capacity constraint introduced through the training methods. Double arrows between and agent's decentralized qfunction and their policy (labelled #1) indicate the information constraint on agent's policy complexity described in Eq. 8 Double arrows between agent's policies and other agent's Q-functions (labelled #2) indicate the information constraint on policy inference described in Eq. 9.
Multiagent Particle Environments [3]. All environment names are taken from the codebase. Adversary: 2 Good agents and 1 adversary are rewarded
by closeness to a target, good agents must not reveal which object is the target by spreading to both the target and distraction. Crypto: 1 Good agent
communicates target landmark to another good agent over a public communication channel, 1 adversary attempts to decode communicated target. Push: 1
good agent moves towards target landmark while avoiding 1 adversary. Reference: 2 mobile good agents communicate to determine which landmark is the
target. Speaker: 1 static good agent communicates to 1 mobile good agent which landmark is the target. Spread: 3 good agents spread to cover all landmarks.
Tag: 1 good agent moves to distance itself from 3 adversaries using obstacles to slow their approach. World: 2 good agents move to gather food, hide within
trees, and avoid 4 adversaries, 1 adversary leader can observe good agents hiding in trees and communicate their location.
Final 1K training episode reward from all 8 environments for 10
CLDAC vs. MADDPG models. Left group is strictly cooperative environ-
ments, right is competitive and mixed environments. All environments have
βμ = 1e−3 and βπ optimized using 3 left out models, values listed in
appendix. Reference and World environments do not accommodate policy
inference and CL-π results are shown in place of CLDAC results.
Capacity-Limited Decentralized Actor-Critic for Multi-Agent Games

September 2021

·

325 Reads

·

7 Citations

This paper explores information-theoretic constraints on methods for multi-agent reinforcement learning (MARL) in mixed cooperative and competitive games. Within this domain, decentralized training has been employed to increase learning sample efficiency. However, these approaches do not explicitly discourage complex policies, which can lead to over-fitting. To address this, we apply an information theoretic constraint onto agents' policies that discourages overly complex behaviour when it is not associated with a significant increase in reward. A second challenge in MARL is the non-stationarity of the environment introduced by other agents' changing policies. Previous methods in MARL have sought to reduce the impact of non-stationarity by inferring other agents' policies, but this can lead to over-fitting to previously observed behaviour. To avoid this, a similar information-theoretic constraint is applied onto the inference of other agents' policies, resulting in a more robust estimate. We evaluate the effects of these information-theoretic constraints on a test suite of multi-agent games, and report an overall improvement in performance, with greater improvements found in competitive domains compared to cooperative games.


Citations (31)


... Recently, Generative AI models have been integrated with cognitive models by forming representations, of stimuli, such as textual information using LLM embeddings [30], [29]. This approach has demonstrated human-like abilities to recognize new stimuli based on past experiences, even when they are informationally complex [31]. We propose the use of LLM embeddings as attributes of a cognitive model to both predict participant learning and evaluate them under different experimental conditions. ...

Reference:

Training Users Against Human and GPT-4 Generated Social Engineering Attacks
Efficient Visual Representations for Learning and Decision Making

Psychological Review

... A notable exception is Cochrane, Cox, and Green (2023), who used a similar approach to what we advocate here to explore how performance on a classic test of latent bias (the Implicit Association Test) changes over the course of trials and blocks of trials. This approach uncovers differences between individuals that go beyond a simple effect size between conditions, but instead reveal more distinct and specified differences in start point, rate, and asymptote (see also Cochrane, Sims, et al., 2023). Our aim in the current article, therefore, was to advocate for the broader use of asymptotic regression for modeling time series effects and to provide some simple worked examples of the additional insight that can be gained from using it. ...

Multiple timescales of learning indicated by changes in evidence-accumulation processes during perceptual decision-making

npj Science of Learning

... This information-bottleneck motivation of these 145 models has been associated with cognitive limitations that impact decision making in humans, resulting in 146 suboptimal behavior (Bhui et al., 2021; Lai and Gershman, 2021). 147 These representations have been related to the processing of visual information from humans in learning 148 tasks (Malloy and Sims, 2022), as they excel in retaining key details associated with stimulus generation 149 factors (such as the shape of a ball or the age of a person's face) (Malloy et al., 2022b). Although we employ 150 β-VAEs in this work, there are many alternative visual GMs that are capable of forming representations 151 useful for decision making. ...

A Beta-Variational Auto-Encoder Model of Human Visual Representation Formation in Utility-Based Learning

Journal of Vision

... This information-bottleneck motivation of these 145 models has been associated with cognitive limitations that impact decision making in humans, resulting in 146 suboptimal behavior (Bhui et al., 2021; Lai and Gershman, 2021). 147 These representations have been related to the processing of visual information from humans in learning 148 tasks (Malloy and Sims, 2022), as they excel in retaining key details associated with stimulus generation 149 factors (such as the shape of a ball or the age of a person's face) (Malloy et al., 2022b). Although we employ 150 β-VAEs in this work, there are many alternative visual GMs that are capable of forming representations 151 useful for decision making. ...

Learning in Factored Domains with Information-Constrained Visual Representations

... Auto-Encoders (β-VAE) to additionally predict utility in a supervised fashion [9]. Disentangled representations have also been applied into improving zero-shot transfer learning in the DRL setting by using latent representations as input to a policy network [6]. ...

Modeling Human Reinforcement Learning with Disentangled Visual Representations

... Note that some previous studies already used training procedures to investigate learning-induced effects on VWM (Blalock, 2015;Chen et al., 2006;Oberauer et al., 2017;Sims et al., 2022;Zimmer et al., 2012). Typically, however, the training was merely designed to increase participants' familiarity with a certain class of stimuli (e.g., polygons, Chinese characters, colored objects), without providing specific semantic object associations. ...

Conceptual knowledge shapes visual working memory for complex visual information

... Using theory of mind as a means of overcoming bounded rationality is also of interest in cooperative games where parts of the environment are not observed by all players, such as the card game Hanabi [14]. Modeling teammate and opponent strategies as boundedly-rational has also demonstrated success in mixed cooperative and competitive environments [15]. ...

Capacity-Limited Decentralized Actor-Critic for Multi-Agent Games

... As β increases the information capacity of the latent representation decreases, while when set to 0, information is unconstrained. β-VAE models have previously been used to model human choice in a visual categorization task [2] and visual decision making [6]. Results from this experiment provide evidence that systematic biases of human categorization can be explained by the information-constrained representations formed by β-VAE models. ...

Modelling Visual Decision Making Using a Variational Autoencoder

... Using theory of mind as a means of overcoming bounded rationality is also of interest in cooperative games where parts of the environment are not observed by all players, such as the card game Hanabi [14]. Modeling teammate and opponent strategies as boundedly-rational has also demonstrated success in mixed cooperative and competitive environments [15]. ...

RL Generalization in a Theory of Mind Game Through a Sleep Metaphor (Student Abstract)

... The function and motivation behind β-VAE models shares a close connection with the information bottleneck approach Alemi, Fischer, Dillon, & Murphy, 2017). This method has been applied to modelling cognitive mechanisms that share similarities with economic decision making, such as predictive inference (Still, 2014) and information-constrained behaviour (Lai & Gershman, 2021;Malloy & Sims, 2020). One key feature of our proposed model is that it makes predictions on the formation of taskrelevant representations under information constraints. ...

Modelling Human Information Processing Limitations in Learning Tasks with Reinforcement Learning