Available via license: CC BY 4.0
Content may be subject to copyright.
Page 1/21
Nest-building using place cells for spatial
navigation in an articial neural network
Thomas Portegys ( portegys@gmail.com )
Dialectek
Research Article
Keywords: Articial animal intelligence, goal-seeking neural network, nest-building, place cells, spatial
navigation
Posted Date: August 31st, 2023
DOI: https://doi.org/10.21203/rs.3.rs-3301060/v1
License: This work is licensed under a Creative Commons Attribution 4.0 International License.
Read Full License
Page 2/21
Abstract
An animal behavior problem is presented in the form of a nest-building task that involves two cooperating
birds, a male and female. The female builds a nest into which she lays an egg. The male's job is to forage
in a forest for food for both himself and the female. In addition, the male must fetch stones from a
nearby desert for the female to use as nesting material. The task is completed when the nest is built and
an egg is laid in it. A goal-seeking neural network and a recurrent neural network were trained and tested
with little success. The goal-seeking network was then enhanced with “place cells”, allowing the birds to
spatially navigate the world, building the nest while keeping themselves fed. Place cells are neurons in the
hippocampus that map space.
Introduction
A task is presented to simulate two birds, a male and a female, that cooperate in navigation, foraging,
communication, and nest-building activities. These activities are commonly performed in many animal
species to ensure survival and reproduction. The female builds a nest into which she lays an egg,
completing the task. The male's job is to forage in a forest for food for both himself and the female. In
addition, the male must fetch stones from a nearby desert for the female to use as nesting material.
The nest-building task was recently proposed as a game providing an articial animal intelligence
challenge to evaluate machine learning systems (Portegys, 2022a). While the task itself is independent of
how it is tackled, here articial neural networks (ANNs) are chosen, as ANNs are capable of generalized
learning, and are intended to perform functions of the biological neural networks of animals.
The task was originally introduced in 2001 (Portegys, 2001), a solution for which was obtained using
Mona, a goal-seeking ANN that at the time employed domain-specic macro-responses, such as “Go to
mate”. The network was also manually coded instead of being learned.
In planning to attack the problem again, this time as a learning task with spatial enhancements (Portegys,
2022b) that would obviate the need for domain-specic responses, place cells seemed to be a good
choice. Place cells are neurons in the hippocampus that map space (Moser et al., 2015; Robinson et al.,
2020; Xu et al., 2019) allowing an animal to navigate its environment effectively.
There is a signicant body of work on using place cell inspired neural functionality in ANNs, much of it
involved with robotic navigation (Milford and Wyeth, 2010; Zhou et al., 2017). These systems are aimed
at solving specic tasks with models that mimic biological place cells. They are not intended to be
general-purpose ANNs, such as Mona, which are designed to learn arbitrary domain-independent tasks.
General-purpose ANNs borrow functionality from brains, such as neural connection updating, but are not
intended to be models of biological brains. Here place cells are incorporated into an ANN to allow it to
effectively operate in spatial environments. To my knowledge this is a novel development.
Page 3/21
As a comparison, an LSTM (Long short-term memory) recurrent neural network (RNN) (Hochreiter and
Schmidhuber, 1997) was also trained on the task, without spatial enhancement.
Historically, AI has mostly focused on human-like intelligence, for which there are now numerous success
stories: games, self-driving cars, stock market forecasting, medical diagnostics, language translation,
image recognition, etc. The impact of ChatGPT (OpenAI, 2023) as a generative language model is a
recent example. Yet the elusive goal of articial general intelligence (AGI) seems as far off as ever, likely
because these success stories lack the “general” property of AGI, operating as they do within narrow,
albeit deep, domains. A language translation application, for example, does just that and nothing else.
Anthony Zador (2019) expresses this succinctly: "We cannot build a machine capable of building a nest,
or stalking prey, or loading a dishwasher. In many ways, AI is far from achieving the intelligence of a dog
or a mouse, or even of a spider, and it does not appear that merely scaling up current approaches will
achieve these goals."
Mona is a goal-directed ANN, designed to control an autonomous articial organism. We can go back to
Braitenberg’s vehicles (1984) as a possible starting point for the study of how simple networks are
capable of controlling goal-directed behavior in automata. Dyer (1993) delineated the contrasting aims of
ANNs for animal behavior vs. traditional articial intelligence (AI) by framing the former as having
biological goals: “Survivability is the overriding task”. As an example of a simulated animal brain,
Coleman et al. (2005) developed an ANN-based cognitive-emotional forager that performed well in a task
that required not only foraging, but also the avoidance of predators.
Perhaps the best example of articial creatures controlled by ANNs is Yaeger’s Polyworld articial life
system (Lizier et al., 2009). Polyworld is an environment in which a population of agents search for food,
mate, have offspring, and prey on each other in a two dimensional world. An individual makes decisions
based on its neural network which is derived from its genome, which in turn is subject to evolution. To my
knowledge, however, training an ANN to build a nest is a novel undertaking.
In addition to simulating life forms, ANNs have been used as modeling and analysis tools for animal
behavior (Enquist and Ghirlanda, 2006; Wijeyakulasuriya et al., 2020).
Description
The code and instructions can be found at: https://github.com/morphognosis/NestingBirds
World
The world is a 21x21 two dimensional grid of cells. Each cell has a
locale
, and an
object
attribute.
Locale
describes the type of terrain:
plain
,
forest
, and
desert
. An
object
is an item to be found in the world:
mouse
(food), and
stone
(nest-building material). A forest exists in the upper left of the world, populated by mice,
which randomly move about, providing an element of surprise for a foraging bird. A desert is found in the
Page 4/21
lower right of the world, where stones are to be found at various locations. The birds are initially located
on the plain in the center of the world.
Birds
There is a male and a female bird. The female builds the nest and the male forages for mice and stones.
The nest is a stone ring in the center of the world in which the female lays her egg. The birds have four
components: senses, internal state, needs, and responses. These are sex-specic to suit the different roles
of the birds.
Male
Senses
locale, mouse-proximity, stone-proximity, mate-proximity, goal, has-object, ying, female-needs-mouse,
female-needs-stone.
Locale
pertains to the current location of the male and has a value of
plain
,
forest
, or
desert
.
The
proximity
sensor values are
present, left, right, forward
, or
unknown
. The
mouse-proximity
sensor
senses a mouse when in the forest, the
stone-proximity
sensor senses a stone when in the desert, and the
female-proximity
sensor senses the female within the bounds of the nest.
The
goal
sensor values are
eat-mouse, mouse-for-female, stone-for-female
, and
attend-female
.
The
has-object
sensor indicates an object carried by the bird and can
be mouse, stone
, or
no-object
.
The
ying sensor
is true when the male is in ight; otherwise false.
Female-needs-mouse
is sensed when the female expresses a corresponding response of
want-mouse
in
the presence of the male. This is the only time this sense is active; when not in the presence of the female
it is in the off state. A similar process exists for the
female-needs-stone
sense and
want-stone
female
response. Only one of the
female-needs/wants
is sensed/expressed at a time.
Internal state
food
.
Initialized to a parameterized value. When food reaches zero, the need for a mouse is surfaced as a goal.
Upon eating a mouse, food is increased by a random value.
Needs
mouse-need, female-mouse-need, female-stone-need, attend-female-need
.
Page 5/21
These correspond to the goals. Upon completion of an errand to satisfy a goal, signied by returning to
the female, the next goal is determined by current needs. As discussed,
mouse-need
is raised when food
reaches 0. The
female-mouse-need
and
female-stone-need
are signaled to the male by the female when
she desires a mouse to eat or a stone to place in the nest, respectively. If none of the above are raised, the
attend-female-need
is raised, causing the male to move to the female’s location.
Responses:
do-nothing
a no-op response.
move-forward
move forward in the orientation direction. Movement off the grid is a no-op.
turn-right/left
change orientation by 90 degrees.
eat-mouse
eat mouse if
has-object
is a mouse. If no mouse, this is a no-op.
get-object
if
has-object
is empty and an object in current cell, pick up the object and set it to
has-object
.
put-object
if
has-object
not empty and no object at current cell, put object in cell and clear
has-object
.
give-mouse
if
has-object
is mouse, and female present with empty
has-object
, transfer mouse to female.
give-stone
if
has-object
is stone, female present with empty
has-object
, transfer stone to female.
y
take ight. This activates a place motor neuron which will move to a specic location in the world. The
specic motor is determined by the current mediator neuron context (see Articial neural networks
section).
Page 6/21
alight
terminate ight after arriving at a location in the world determined by the active place motor neuron.
Female
Senses
object-sensors
,
orientation, goal, has-object
.
object-sensors
the female senses the object values in its Moore (3x3) neighborhood.
orientation
north, south, east and west.
goal
lay-egg, brood-egg, eat-mouse.
has-object
identical to male.
Internal state
food.
Identical to male.
Needs
lay-egg-need, brood-egg-need, mouse-need
.
The
mouse-need
need is raised by food level and sets the
eat-mouse
goal. It will cause the female to
express the
want-mouse
response. While building the nest and not hungry, the
lay-egg-need
is kept raised
with the associated
lay-egg
goal. The female asserts the
want-stone
response when she is located in a
cell where the nest requires a stone to be placed. After a stone is placed, the female proceeds to the next
location in the nest and repeats the process, continuously motivated by the
lay-egg-need
and
lay-egg
goal. When the nest is built, the female lays her egg in the center of the nest. After that
brood-egg
-
need
is
kept raised and the
brood-egg
goal keeps the female brooding on the egg.
Responses:
do-nothing
Page 7/21
a no-op response.
move-forward
move forward in the orientation direction. Movement off the grid is a no-op.
turn-right/left
change orientation by 90 degrees.
eat-mouse
eat mouse if
has-object
is a mouse. If no mouse, this is a no-op.
get-object
if has-object is empty and object in current cell, pick up the object and set it to
has-object
.
put-object
if
has-object
and no object in current cell, put object in cell and clear
has-object
.
want-mouse
when the
eat-mouse
goal is sensed, the
want-mouse
response signals the male to retrieve a mouse from
the forest for the female to eat.
want-stone
when the
lay-egg
goal is sensed, and the female is ready to place a stone in the nest, the
want-stone
response signals the male to retrieve a stone from the desert for the female to place in the nest.
lay-egg
when the female has completed the nest and has moved to its center, she lays the egg with this response.
Articial neural networks
Mona
Mona learns cause-and-effect chains and hierarchies of neurons that represent these relationships in the
environment. A detailed description of the architecture can be found in Portegys (2001). An overview is
provided here.
Three types of neurons are dened, as shown in Fig.1:
Receptor neuron: represents a sensory event.
Page 8/21
Motor neuron: expresses response to the environment.
Mediator neuron: represents a predictive relationship between neurons. The ring of its cause neuron
enables
its motor neuron to re. If the network res the motor neuron, the effect neuron will
probabilistically re. A mediator that mediates lower level mediators serves as a context that
recursively affects the probability of causation in its components.
Mona is a goal-seeking network, falling into the category of a model-based reinforcement learning system
(Moerland et al., 2023). Needs arising from internal and external sources are satised by the ring of
associated goal neurons. For example, a need for water is satised by ring neurons involved with
drinking water. Need drives backward through the network from a goal neuron as
motive
, following
enabled pathways to re a motor response that will navigate toward the goal.
There can be multiple needs vying for control. For example, food and thirst might both act on the network
simultaneously. The winner will be a function of the strength of the need and the enablement of the
network to achieve the goal.
In the nest-building task, the male learns a network consisting of chains of mediators that drives it to the
forest when it needs a mouse to eat, then returns it to the female for further goal setting. It also learns
mediators that orchestrate fetching mice and stones for the female. A powerful feature of Mona is that
mediators are modular, that is, they can take part in multiple goal activities. For example, the mediators
that y the male to the forest and hunt for a mouse are used when both the male and female require a
mouse.
The female’s need to lay an egg, in conjunction with the goal of sensing the egg in the nest, drives
through a chain of mediators that orchestrate a series of nest-building actions, each of which involves
expressing a want for a stone that signals the male to fetch a stone, placing the stone, and moving to the
next location until the nest is entirely constructed. The female then moves to the center of the nest and
lays her egg, satisfying the egg laying need. Then the need to brood the egg is raised, which keeps the
female sitting on the egg.
Bird responses are trained by overriding incorrect responses with correct ones. The correct responses are
incorporated into the network. During testing, responses are generated by the network.
Place motor neurons
To enhance Mona with place cell functionality, a place motor neuron was implemented. A biological place
neuron res when a specic location in the world is reached. Mona’s place neurons re similarly. However,
they also navigate to specic places in the world. A related capability seems to exist in the form of route
encoding in rats (Grieves et al., 2016).
In Mona, a place neuron is implemented as a motor neuron that records a specic location in the world
and will navigate to that location when motivated to respond. It res when the location is reached. For
Page 9/21
example, when the male’s goal is to fetch a mouse for the female, a mediator with a cause of sensing the
female’s want of a mouse, and an effect of sensing the proximity of a mouse in the forest will re its
place motor to enact a series of primitive movement responses that navigate to a prescribed location in
the forest.
Place motors can be learned while exploring the world by marking signicant locations. A special
responses initiates the commencement of a sequence of movements that terminate at some location, at
which time another special response marks the location, creating a new place motor neuron. In the
nesting birds, these two special responses are mapped to the male
y
and
alight
responses, respectively.
LSTM
The LSTM, introduced in 1997 (Hochreiter and Schmidhuber, 1997), is a recurrent neural network which
has established itself as a workhorse for sequential pattern recognition. LSTMs address a problem with
other recurrent neural networks in which corrective information vanishes as the time lag between the
output and the relevant input increases, leading to the inability to train long-term state information.
In the LSTM network, the hidden units of a neural network are replaced by memory blocks, each of which
contains one or more memory cells. A memory block is shown in Fig.2. The block can latch and store
state information indenitely, allowing long-term temporal computation. What information to store, and
when to output that information are part of the training process.
Scenario
This scenario is taken from the game proposal (Portegys, 2022a), which illustrates the task from a
general viewpoint. It shows intermittent states of the world, from initial state to egg-laying in the
completed nest. A video is available here: https://youtu.be/d13hxhltsGg
Results
Two ANNs were trained and tested on the task under varying conditions:
1. Mona version 6.0. Maximum of 500 mediator neurons. Maximum mediator level of 0, meaning
mediators mediated receptor and motor neurons exclusively; higher level mediators that mediate
lower level mediators were not needed.
2. An LSTM (Long-short term memory) recurrent neural network using the Keras 2.6.0 python package.
128 neurons in a hidden layer, and a mean squared error loss function. Input and output were one-hot
encoded. Training was conducted with 500 epochs.
Number of training datasets
Page 10/21
Performance was measured based on the number of training datasets, as shown in Figs.8 and 9. A
training dataset is generated by running the nesting birds with optimal bird responses until the nest is
completed while keeping the birds fed.
Each run is seeded with a different random number that controls the movements of the mice in the forest
and how long eating a mouse serves to satisfy a bird’s hunger. The testing accuracy was calculated as
the percent of correct responses out of 1000 steps. Measurements were based on the mean values of 20
trials. Very little variance in values was observed.
Both networks show excellent performance with only a single training dataset.
Dynamic testing
Mona performance is measured as it interacts with the world. That is, responses output to the world
cause changes in the world that are subsequently input to the bird. For example, if the bird turns to the
left, the world reacts by altering the sensory state of the bird accordingly. With place motor neurons, Mona
solves the task every time, but without place motors the male bird, which must perform complex
navigation to fetch mice and stones, repeatedly becomes lost, causing the entire task to fail.
When the same “dynamic” interface is applied to the RNN network the male bird repeatedly becomes lost
while fetching mice or stones. This means that even the few errors that the male makes are crucial,
preventing successful completion. If the male cannot return with a mouse for the female, for example, the
female cannot proceed with nest building.
RNN epoch testing
Mona trains in a single epoch, a skill frequently seen in human learning (Lee et al., 2015). The RNN
training is signicantly affected by the number of epochs of training, especially for the male, as shown in
Fig.10.
Conclusion
Enhanced with place motor neurons, Mona is capable of solving the nest-building task every time. The
RNN performs well with its typical usage, which is to predict upcoming responses, but fails as a navigator
for the male bird, causing nest-building to be unsuccessful. How place neuron functionality might be
incorporated into an RNN is an interesting topic.
Place motor neurons and goal-seeking causation learning are a powerful combination of capabilities for
the nest-building task which demands both spatial and sequential learning. This animal learning task
exposes shortcomings in a deep learning ANN that researchers interested in articial general intelligence
(AGI) should be aware of. I recommend that further research be conducted to (1) further simulate animal
behaviors, and (2) adopt mechanisms from neurobiology, such as place cells, that allow machines to
acquire animal-like capabilities. These I believe are essential to the achievement of AGI.
Page 11/21
References
1. Braitenberg, V. (1984). Vehicles: Experiments in synthetic psychology. Cambridge, MA: MIT Press.
"Vehicles - the MIT Press". Archived from the original on 2010-01-29. Retrieved 2012-06-18.
2. Coleman S.L., Brown V.R., Levine D.S., Mellgren R.L. (2005). A neural network model of foraging
decisions made under predation risk. Cogn Affect Behav Neurosci. 2005 Dec;5(4):434-51. doi:
10.3758/cabn.5.4.434. PMID: 16541813.
3. Dyer, M.G. (1993). Toward Synthesizing Articial Neural Networks that Exhibit Cooperative Intelligent
Behavior: Some Open Issues in Articial Life.
Articial Life
, vol. 1, no. 1_2, pp. 111-134, Oct. 1993, doi:
10.1162/artl.1993.1.1_2.111.
4. Enquist, M. and Ghirlanda, S. (2006). Neural Networks and Animal Behavior. Volume 33 in the series
Monographs in Behavior and Ecology Published by Princeton University Press.
https://doi.org/10.1515/9781400850785
5. Grieves R. M., Wood E. R., Dudchenko P. A. (2016). Place cells on a maze encode routes rather than
destinations. Elife. Jun 10;5:e15986. doi: 10.7554/eLife.15986.
. Hochreiter, S., Schmidhuber, J. (1997). Long short-term memory.
Neural Computation
, 9(8), 1735-
1780.
7. Lee, S.W., O’Doherty, J.P., Shimojo, S., (2015). Neural computations mediating
one-shot learning in the human brain. PLoS biology 13.
. Lizier, J.T., Piraveenan, M., Pradhana, D., Prokopenko, M., Yaeger, L.S. Functional and Structural
Topologies in Evolved Neural Networks. ECAL 2009.
9. Milford, M. and Wyeth, G., (2010). Persistent navigation and mapping using a biologically inspired
SLAM system,
Int. J. Robot. Res.
29:1131–1153.
10. Moerland, T.M., Broekens, J., Plaat, A., Jonker, C.M. (2023). Model-based Reinforcement Learning: A
Survey. Foundations and Trends in Machine Learning Series. Now Publishers.
https://books.google.com/books?id=FimgzwEACAAJ
11. Moser M.B., Rowland D.C., Moser E.I. Place cells, grid cells, and memory. (2015). Cold Spring Harbor
Perspectives in Biology. Feb 2;7(2):a021808. doi: 10.1101/cshperspect.a021808. PMID: 25646382;
PMCID: PMC4315928.
12. OpenAI. (2023).
ChatGPT
(Mar 14 version) [Large language model]. https://chat.openai.com/chat
13. Portegys, T.E. (2001). Goal-Seeking Behavior in a Connectionist Model.
Articial Intelligence Review
16, 225–253 (2001). https://doi.org/10.1023/A:1011970925799
14. Portegys, T.E. (2022a). I want to play a game.
https://www.researchgate.net/publication/364089546_I_WANT_TO_PLAY_A_GAME
15. Portegys, T.E. (2022b). Morphognostic honey bees communicating nectar location through dance
movements, Nature and Biologically Inspired Computing (NaBIC 2022).
1. Robinson, N. T. M., Descamps, L. A. L., Russell, L. E., Buchholz, M. O., Bicknell, B. A., Antonov, G. K.,
Lau, J. Y. N., Nutbrown, R., Schmidt-Hieber, C., Häusser, M. (2020). Targeted activation of
Page 12/21
hippocampal place cells drives memory-guided spatial behavior.
Cell
, 183, pp. 1586-1599.
17. Wijeyakulasuriya, D.A., Eisenhauer, E.W., Shaby, B.A., Hanks E.M. (2020). Machine learning for
modeling animal movement. PLoS ONE 15(7): e0235750.
https://doi.org/10.1371/journal.pone.0235750
1. Xu, H., Baracskay, P., O’Neill, J., and Csicsvari, J. (2019). Assembly responses of hippocampal CA1
place cells predict learned behavior in goal-directed spatial tasks on the radial eight-arm maze.
Neuron
101, 119–132.
19. Zador, A. (2019). A critique of pure learning and what articial neural networks can learn from animal
brains.
Nature Communications
volume 10, Article number: 3770.
https://www.nature.com/articles/s41467-019-11786-6
20. Zhou, X., Weber, C. and Wermter, S., (2017). Robot localization and orientation detection based on
place cells, Proc. ICANN 2017, Springer pp. 137–145.
Figures
Figure 1
A simple Mona network
Page 13/21
Figure 2
LSTM memory block
Page 14/21
Figure 3
Beginning state. Female is hungry (0 food), male has maximum food. Initial response for both is “do
nothing”. Both are located at center of world. Upper left is forest with mice (food). Lower right is desert
with stones for nest-building.
Page 15/21
Figure 4
While co-located, female signals to male with “want food” response. Male ies to forest and picks up a
mouse to feed to her.
Page 16/21
Figure 5
Female moves to location of rst nesting stone. Male follows her. She signals to male that she wants a
stone. Male ies to desert and picks up a stone.
Page 17/21
Figure 6
Male returns to female with stone. Discovers she is hungry. He ies to forest for mouse for her.
Page 18/21
Figure 7
Nest completed. Egg laid.
Page 19/21
Figure 8
Female test performance with number of training datasets
Page 20/21
Figure 9
Male test performance with number of training datasets.
Figure 10
Page 21/21
Training epochs testing