• Nem Talált Eredményt

PETER PAZMANY CATHOLIC UNIVERSITYConsortium members

N/A
N/A
Protected

Academic year: 2022

Ossza meg "PETER PAZMANY CATHOLIC UNIVERSITYConsortium members"

Copied!
43
0
0

Teljes szövegt

(1)

Development of Complex Curricula for Molecular Bionics and Infobionics Programs within a consortial* framework**

Consortium leader

PETER PAZMANY CATHOLIC UNIVERSITY

Consortium members

SEMMELWEIS UNIVERSITY, DIALOG CAMPUS PUBLISHER

The Project has been realised with the support of the European Union and has been co-financed by the European Social Fund ***

**Molekuláris bionika és Infobionika Szakok tananyagának komplex fejlesztése konzorciumi keretben

***A projekt az Európai Unió támogatásával, az Európai Szociális Alap társfinanszírozásával valósul meg.

PETER PAZMANY CATHOLIC UNIVERSITY

SEMMELWEIS UNIVERSITY

(2)

Peter Pazmany Catholic University Faculty of Information Technology

MODELLING NEURONS AND NETWORKS

Lecture 9

Models of memory

www.itk.ppke.hu

(Idegsejtek és neuronhálózatok modellezése)

(Memóriamodellek)

Szabolcs Káli

(3)

Overview

In this lesson you will learn about the structure and workings of the

neural networks responsible for storing and recalling memory traces. We will discuss several theories that try to explain how short- and long-term memory work.

Lesson topics:

Short-term memory

Recurrent network models of short-term memory

Model for working memory and decision making

Long-term memory: the hippocampus as a long-term memory store.

An attractor network model of spatial representations in area CA3

Generating non-directional place fields

The effects of manipulating the environment

Spatial representations, special cases: very different and very similar

environments

(4)

Short-term memory

• Can store only a few items at the same time.

• Can only store items for a limited amount of time. The cause of the decay is debated: Some suggest that its contents spontaneously degrade over time. Others explain the degradation by a competitive model: When many items are stored in memory, these elements compete with each other, so the more active content pushes out the less active ones.

• Persistent neural activity is required for the storage – excitatory synaptic loops could be responsible for it.

• Synaptic inhibition is also likely to be critical for shaping the selectivity of mnemonic neural activity in a working memory circuit.

• Studies have shown that when there is a delay between the

stimulus and the response, neurons show a selectively

enhanced activity throughout the delay period in the prefrontal

cortex (PFC) and in the inferotemporal cortex (ITC).

(5)

Recurrent network models of short-term (working) memory

Brunel and Wang, 2001: Model of object working memory in a large cortical network that is dominated by feedback inhibition. The model represents a cortical module of an area receiving information about the identity of objects.

Main features of the model:

• Persistent activity is generated within a local cortical circuit.

• Recurrent synaptic excitation is largely mediated by NMDA receptors.

• Recurrent network interactions are dominated by synaptic inhibition.

The network is composed of two types of cells: excitatory pyramidal cells (80% of the cells) and inhibitory interneurons (20%). Both are represented by integrate-and-fire models (see Lecture 7).

Both pyramidal cells and interneurons also receive excitatory connections

from outside the network. These connections send to the network all the

information (stimuli) received from the outside world, as well as

background noise due to spontaneous activity outside the module.

(6)

Recurrent network models of short-term (working) memory

Figure: Schematics of the memory model (explanation on next page).

(7)

Recurrent network models of short-term (working) memory

• Pyramidal cells send connections to other pyramidal cells and interneurons through AMPA and NMDA synapses.

• Interneurons send GABAergic connections to pyramidal cells and other interneurons.

• Both receive excitatory connections from other cortical areas.

• Pyramidal cells can be functionally divided in several groups

according to their selectivity properties. Group #1 is selective to

object #1, etc. Cells within a group have relatively stronger

connections (modulated by w+ > 1), while connections between

different groups are relatively weaker (modulated by w− < 1).

(8)

Delayed-match-to-sample task protocol

1. There is a 1 second startup period, in which the network exhibits spontaneous activity.

2. Stimulus presentation (sample) consists of a transient input (lasting for 500msec) to those cells selective to the shown stimulus. Other cells are unaffected.

3. After the external stimulus is removed, there is a delay period of 4 s.

4. Finally, a match stimulus is presented for 500ms. In the last

400ms of the match presentation, the external frequency to all

cells is multiplied by 1.5 to account for an increase in afferent

inputs due to the behavioral response/reward signal.

(9)

Delayed-match-to-sample task analysis

Figure: Network activity during the delayed- match-to-sample task.

A: Average activity of different types of cells.

Red: cells selective for the shown stimulus.

Green, yellow, blue, and brown: cells selective for other stimuli.

Cyan: cells not selective to any of the stimuli.

Black: Inhibitory cells.

The stimulus triggers persistent activity in a selective cell population (red) at about 25 Hz.

Delay period activity is switched off by a transient excitatory input generating a brief surge of activity in all neurons.

B: Interspike interval histograms of cells selective to the input sample during spontaneous and persistent activity.

(10)

Memory preservation in the presence of distractors

A distractor is a nonmatch stimulus different from the initial sample. It is applied to those neurons which are sensitive to this stimulus and with the same frequency as the sample and match stimuli.

The effect of distractors on network activity depends mainly on the stimulus amplitude (see figure in next page):

• At very weak stimulus intensity, presentation of the sample stimulus was unable to elicit any elevated delay activity.

• For intermediate values of stimulus amplitude the stimulus caused persistent activity that was resistant to distractors.

• For high values of stimulus amplitude, cue-related persistent activity (short- term memory) is disrupted by the nonmatch stimuli and the network is

distracted. In this case, a nonmatch stimulus is sufficiently powerful to bring its corresponding neuronal assembly (whose best stimulus is the distractor) from a low firing state to its high firing state. The initial match neural

population’s persistent activity is disrupted. These cells are turned off as a dynamical consequence of the rise of the new neuronal subpopulation, which leads to a transient surge of feedback inhibition to the rest of the network.

(11)

Results on the delayed-match-to-sample task

Figure A: The network retains its ability to recall the sample with

distractor stimuli at immediate intensities.

Figure B: The persistent activity is disturbed if the distractor intensity is high enough, and the network changes into another persistent state.

Red: initial stimulus population activity.

Green, violet: distractor population activity.

(12)

Combined model of working memory and decision-making

Machens, Romo, and Brody (2005): Model of working memory and decision making. The model’s main feature is mutual inhibition between the network units, which is a useful design motif of networks with multiple behaviors.

Experimental setup:

Stimuli are frequencies of vibrations applied to the tip of a monkey’s finger, and the activity of prefrontal cortical (PFC) and somatosensory cortical (S2) neurons is measured during the experiment.

1.Subjects perceive a short stimulus, called f1, and maintain it in their working memory.

2.They compare it after a delay with a brief second stimulus, called f2.

3.They decide immediately after which stimulus was larger.

(13)

Experimental setup and results

Horizontal scale indicates time, the rainbow colored bar indicates frequency for f1. For each trial, monkeys made the correct response more than 91% of the time.

A: Experiment setup. Stimuli are named f1 and f2.

B-E: Neuronal responses during the experiment. Blue lines at the end are „YES” responses; brown lines are „NO” responses. Color codes indicate f1 frequency during measurement.

B,C: Firing rates of prefrontal cortical (PFC) plus and minus neurons. The signals are encoded in two complementary sets of roughly equal number of neurons: Plus neurons (fig. B) fire the most if the decision at the end is „YES”, minus neurons (fig. C) if the decision is „NO”. During f1 (memorizing phase) and in the maintenance phase there is a strong f1-dependent firing rate.

During the comparison phase(f2), neurons segregate quickly into YES/NO categories. When the f1 stimulus is high, plus neurons tend to fire more, because „YES” answer is more likely, and the opposite is true for „NO” neurons. Each of the two sets carry all information necessary, the existence of redundant sets is unexplained.

D,E: Neuronal response in area S2. After 2.8 seconds, color codes indicate f2 response.

(14)

Model for working memory and decision making

One interesting finding of the experiments is that the same neurons which maintain the working memory are responsible for the decision making too.

The algorithmic representation for modeling both of these tasks:

•Both memory and decision making are represented by the value of a single state variable.

(dots on the picture, horizontal position indicates state variable value.)

•The dynamical modes are represented by an energy function L, the shape of which does not depend of the state variable. (line on the picture, energy levels are represented by the vertical axis.)

The state variable always moves to a lower energy level.

During the loading (memorizing) stage the external stimulus creates a single minimum in L.

During the maintenance phase, there is no stimulus that determines the shape of L, thus L is flat.

During the decision phase, the f2 stimulus is mapped onto a peak of the L function. Now all state positions to the left of the peak caused by f2 represent memories of a smaller f1, and vice versa.

If the horizontal axis stands for firing rate that grows from left to right, the position of the state variable mimics the firing rate of the plus neurons measured during the experiment. They mimic the minus neurons if the horizontal axis grows from right to left.

(15)

The authors of the article propose that the frontal-lobe areas instantiate this algorithm. Clues about the underlying architecture come from the analysis of firing-rate covariations between pairs of PFC neurons, which reveal distinct plus and minus neuron populations.

In a simplified version of this circuit each unit represents a population of neurons (fig. B), and the output is the average activity of the population. The response strength of these neurons to different strengths of excitatory input can be seen in fig. C (I/O function).

The dynamics of the circuit can be seen in Fig. D: The black line shows the output of the plus node as a function of the inhibitory output from the minus node (the plus node has a constant excitatory input applied.). The minus node’s I/O function is plotted by exchanging the horizontal and vertical axes (brown line).

This phase-plane plot describes the dynamics of the system. Points where the two curves intersect are steady-states, which can be stable of unstable.

Neural representation of the algorithm

(16)

Neural representation of the algorithm

Figure: Detailed explanation of the algorithm during storage (loading), maintenance and decision phase.

(for explanation see next page.)

(17)

Neural representation of the algorithm

(explanation of the figures on the previous page)

S2 plus neurons project to the PFC plus neurons, and S2 minus neurons project to the PFC minus neurons. Stimulus frequency in this figure corresponds to the angle of θ.

During the first stimulus, the position of the single stable point is determined by f1 (fig. E). Blue/green/red lines show the effect of three stimuli.

During the maintenance period inputs from S2 are silent (fig. F). The value of E (external constant input) can be chosen such that the two I/O curves largely overlap. This creates a continuous line of stable points.

During the decision phase (fig. G) it is assumed that E is externally reduced, to signal that the system is in the decision phase. Because E is low, there are two stable fixed points. Trajectories start from endpoints in fig. F.

In the experimental data, frontal neurons switch their stimulus dependence, but S2 neurons do not. This is represented by the circuit shown in fig. H.

As a result of this sign-switching, increasing f2 moves the unstable fixed point from a lower to a higher θ, which matches the f1-dependence of the stable point during loading.

(18)

Associative (long-term) memory

As we have seen, recurrent networks often have activity patterns which behave as fixed point attractors. These may also be considered as stored memory traces, provided that we can specify the fixed points (preferably via a plausible activity-based learning rule).

Auto-associative function: the dynamics of the network reconstructs the original pattern based on a fragment or a noisy version.

(19)

(Binary) Hopfield network:

• binary (-1 / +1) units

• all-to-all symmetrical connections

• linear threshold activation

• asynchronous state updates

• Hebbian learning

• dynamics minimizes energy function

• learning creates minima in the energy landscape

• learning multiple items can lead to spurious attractors

• capacity is about 0.14 N (N is the number of units)

Recurrent network models of long-term memory

(20)

The hippocampus as a long-term memory store

Capacity with graded-response neurons and sparse connectivity (model of hippocampal CA3 network by Treves and Rolls):

(1)

C

RC

is the number of modifiable recurrent connections per cell, a is the

sparseness of individual patterns, k ~ 0.2 - 0.3.

(21)

If C

RC

=12000 and a=0.02, p

max

~36000.

More recent analysis with spiking neurons (Roudi and Latham, 2007):

P

max

is proportional to C

RC

but independent of a, p

max

~120.

Two distinct modes of operation are needed:

• learning: plastic synapses, feedforward inputs dominating activity

• retrieval: no plasticity, strong recurrent interactions

May be implemented through neuromodulatory inputs (Hasselmo).

Two separate input pathways may be required:

• strong, sparse input during learning (mossy fibers)

• many associatively modifiable inputs for retrieval (perforant path)

The hippocampus as a long-term memory store

(22)

Spatial learning and memory depend on the hippocampus.

Place cells provide a population code for location (and other things).

How is this representation formed?

Does the CA3 recurrent network function as an attractor network in this case?

Samsonovich & McNaughton (1997): Fixed, statistically independent sets of attractors (continuous attractors) can be encoded.

But: experience-dependent, partially overlapping representations exist.

New idea: attractors may be determined by experience (in analogy with the fixed point attractors underlying memory for discrete items, as described earlier).

Formation of spatial representations in the hippocampus

(23)

Place fields:

compact, regular,

stable after ~10 minutes of exploration,

non-directional in open field,

directional in restricted environments.

"Orthogonal" representations for very different environments.

Entire set of place fields can follow transformations of the environment

Representations can overlap partially

Place cell representation

(24)

Place cell representation

Figure: place fields of five cells (recorded simultaneously).

The rat travels in a triangular maze; the dots indicate the locations where

each cell fired.

(25)

Model architecture

The model gets input from the entorhinal cortex (EC), whose activity is determined by the environment. This representation is transformed by feedforward pathways (direct input to CA3 and pathways through the dentate gyrus (DG)). The CA3 pyramidal cells (black circles) have recurrent connections and connections with inhibitory cells (open circles).

Solid lines represent connections that are modeled explicitly, solid thick lines represent modifiable connections.

(26)

The change of the membrane potential over time is described by the following equations:

Model dynamics

Where

ui is the membrane potential of the ith pyramidal cell,

v is the membrane potential of the global inhibitory cell (all relative to their resting potentials), and are the membrane time constants for pyramidal neurons and the inhibitory cell, respectively,

Jij is the strength of the connection from pyramidal neuron j to pyramidal neuron i, h is the efficacy of inhibition,

w represents the strength of the excitatory connection from any one pyramidal cell onto the inhibitory cell,

IiPP and IiMF are the inputs to cell i through the perforant path and the mossy fibers (MF), respectively.

gu(u)=[u-]+ is the threshold linear activation function for the pyramidal cells, where [. . .]+ makes all negative arguments zero but leaves positive numbers unaffected, stands for the threshold, and is the slope of the activation function above the threshold.

(27)

Sources of spatial information: all modalities Let us concentrate on:

• Visual inputs

• Self-motion - path integration

Inputs are location and direction dependent:

Model inputs

Left: The net two-dimensional tuning of a sample EC neuron in a rectangular box of dimensions 2x1; the preferred location of the cell is (0.5, 0.4); the current heading of the rat at each location is the same as the preferred head direction of the cell.

Right: The activity of an EC neuron as a function of the difference between its preferred direction and the actual heading of the rat.

(28)

Two-mode hypothesis (Hasselmo)

On-line neuromodulation of synaptic efficacies in the hippocampus by septal inputs (ACh and GABA).

Learning

(high ACh, in new environment)

• mossy fiber inputs dominate CA3 activity

• recurrent connections suppressed

• inhibition reduced

• plasticity in perforant path and recurrent synapses Recall

(low ACh, in familiar environment)

• recurrent connections dominate

• feedback inhibition

• plasticity suppressed

(29)

Inputs are directional, place fields in open environments are not - Can we account for this?

Generating non-directionality

1. Learning phase - establish weights

Assume equal exposure to all possible locations and head directions

Feedforward weights (perforant path) learn to represent essentially the same mapping from entorhinal cortex to CA3 as the pathway through the dentate gyrus

Recurrent weights determined by spatial correlations in DG inputs 2. Recall phase - run until convergence

CA3 place cell activity determined by attractor dynamics

Feedforward inputs select from attractors

(30)

When all headings are equally likely at each location during learning, the resulting attractors in the model can be direction independent.

Generating non-directionality

Figure: Population plots. Cells with a given preferred direction (north and south in this case) are arranged on an imaginary plane according to their preferred locations. This describes the behavior of all the cells with the actual position and heading of the animal kept fixed.

Plots show the activities of cells in EC, the net PP inputs to CA3 neurons (IiPP), and the final activities of the same place cells (marked CA3), as a function of the preferred location of the neuron; the two rows are for cells with preferred head direction indicated by the arrow next to each row.

(31)

Directionality is task dependent:

Generating non-directionality

Figure: The task-dependence of directionality. The contour plots show the place field of a CA3 cell that prefers the left direction, when the rat faces in the

direction indicated by the arrows.

Left column: The model rat explored the environment randomly during the learning phase – place fields are non-directional.

Right column: Model animal that always ran in one of the directions parallel to the long walls of the box during learning. The top plot is empty because the cell does not fire at all in that direction in this case.

(32)

Manipulating the environment 1

What happens if attractors are recalled in an environment that is different from the one where they were initially learned?

Toy example: attractor-based remapping

(33)

Effect on place fields of stretching the environment:

Manipulating the environment 2

Figure a: The place fields of four selected cells in the original and the stretched environment; the firing rates shown are averages over all head directions.

Figure b: Directionality of the place field shown in the bottom right corner of a; the place field depends on the heading of the rat (indicated by the arrows). This dependence on head direction is induced by the transformation of the environment; place fields in the original environment are essentially non-directional.

(34)

Idealized learning

Learning rule: associative - weight decay

Modulation: binary, global

Exploration model: Assume equal exposure to all possible locations and head directions in environment

Result:

Correctly models many properties of place fields, e.g. task-dependent

directionality

(35)

Exploration and exposure

Does the learning procedure actually work in real time?

Exposure:

Sample path:

An example trajectory, showing the first 5 minutes of exploration in a common paradigm in which the rat chases food pellets thrown into random locations in the

environment.

Sampling density as a function of position obtained by

convolving the sample path

with a Gaussian function.

(36)

Place cell firing after naive on-line associative learning:

Spatial representation

Place fields do not form correctly.

(37)

Learning continuous attractors

Uneven exploration results in substantial deviation from "ideal" weight structure:

random noise - mostly harmless

systematic bias - destroys attractor structure

Problem cannot be solved by using a modified learning rule involving

the same quantities.

(38)

Familiarity-gated learning

Try to equalize the amount of learning at different locations by modulating the learning rate:

Use exposure measure

to define novelty:

Choose optimally -> regular weight structure after 5 min

(39)

Naive on-line associative learning: Familiarity-gated learning:

Spatial representations

Place cell firing patterns during recall, after using different learning procedures.

The figure shows the firing rate maps of 18 randomly selected CA3 place cells after the exploration using simple Hebbian learning (left) and using the

familiarity-based learning procedure to establish the weights (right).

(40)

Further Results – very different environments

Figure: The place fields of five selected place cells in a rectangular and a circular apparatus that have very different sensory features.

The top row (R1) shows the place fields after learning in the rectangular apparatus but before any experience in the circular one, and the middle and

bottom rows show the place fields in the circular and rectangular environments (C and R2, respectively) after the rat has become familiar with both.

There is no obvious relationship between place fields of the same cell in the two environments. The effect of encoding a second environment on the place cell

representation in the first environment can be assessed by comparing the top and bottom rows. Although there are some visible changes, these tend to be small and do not affect the general structure of the spatial representation.

(41)

Overlapping attractors can also be encoded:

Experiment by Skaggs and McNaughton, 1998: Animals explored an apparatus that consisted of two visually identical boxes connected by a corridor. Many place cells were found to have similar place fields in the two regions, whereas others had uncorrelated place fields.

Further Results – very similar environments

Figure: The place fields of five CA3 place cells in the two identical boxes in the model; activity in the corridor connecting the boxes was not simulated.

The model has no difficulty storing and recalling partially overlapping spatial representations, because the degree of overlap is determined by what proportion of granule cells distinguishes between the two boxes in DG; CA3 cells simply inherit the selectivity of their MF inputs as attractors are established during the learning phase.

(42)

Overview

In this lesson we analyzed two models of short-term memory, and models of long-term memory. We learned about:

The basic properties of short-term memory: can only store a few items at one time, requires persistent neural activity to maintain memory traces and can only store items for a limited amount of time.

We analyzed a model by Brunel et al., which suggests that persistent activity is generated within a local cortical circuit, recurrent synaptic excitation is largely mediated by NMDA receptors, and recurrent network interactions are dominated by synaptic inhibition. With these assumptions the model is able to maintain persistent activity even in the presence of distractors, if the intensity of the distractor is not too high.

Next we analyzed a model by Machens et al., who proposed a combined model of short- term memory and decision making. The model integrates short-term memory storage, maintenance, recall and comparison because its dynamical properties are easily controlled without changing its connectivity. The network exhibits mutual inhibition between nonlinear units, which makes it easily possible to construct networks with multiple behaviors.

Finally we learned about a CA3 attractor network model by Káli et al. In this model local, recurrent, excitatory and inhibitory interactions generate appropriate place cell representations from location- and direction-specific activity in the entorhinal cortex.

Familiarity with the environment influences the efficacy and plasticity of the recurrent and feedforward inputs to CA3. Depending on the training experience provided, the model generates place fields that are either directional or nondirectional and which change in accordance with experimental data when the environment undergoes simple geometric transformations.

(43)

Suggested reading

Books:

• Peter Dayan and L.F. Abbott: Theoretical Neuroscience, chapters 8 and 9

Articles:

• Nicolas Brunel and Xioa-Jing Wang (2001): Effects of

Neuromodulation in a Cortical Network Model of Object Working Memory Dominated by Recurrent Inhibition. Journal of

Computational Neuroscience 11, 63–85, 2001

• Christian K. Machens, Ranulfo Romo, Carlos D. Brody (2005):

Flexible Control of Mutual Inhibition: A Neural Model of Two-Interval Discrimination. Science Vol. 307. no. 5712, pp. 1121 – 1124

• Szabolcs Káli and Peter Dayan (2000): The Involvement of

Recurrent Connections in Area CA3 in Establishing the Properties of Place Fields: a Model. The Journal of Neuroscience, October 1,

2000, 20(19):7463–7477

Ábra

Figure A: The network  retains its ability to recall  the sample with
Figure a: The place fields of four selected cells in the original and the stretched  environment; the firing rates shown are averages over all head directions

Hivatkozások

KAPCSOLÓDÓ DOKUMENTUMOK

In this article, I discuss the need for curriculum changes in Finnish art education and how the new national cur- riculum for visual art education has tried to respond to

microstate with E subsystem is proportional to the number of microstates with energy E environment. which can be attained by

● In any reaction the rate can depend on the concentration of reactants. ● This relationship is mathematically described by the

● The other approach to get the value of the rate constant of a reaction is the transition state or activated complex theory. ● Transition state is the state with the highest

If the current injected into the cable is held constant, the membrane potential settles to a steady-state solution that is independent

• Finally we modeled synaptic conductances, paying extra attention to the NMDA channel: The conductance of this channel depends not only on the binding of the transmitter, but also

Minimum injected current that elicits an action potential as a function of pulse length (the threshold for constant injection is 0.018nA).. Solutions for current clamp

What is the minimal inhibitory weight needed to negate the excitatory input coming from dendrite 2 on the soma (plot the membrane potential of the soma)?. Isolate Dendrite 2