Literature DB >> 25960737

Memory dynamics in attractor networks.

Guoqi Li1, Kiruthika Ramanathan2, Ning Ning2, Luping Shi1, Changyun Wen3.   

Abstract

As can be represented by neurons and their synaptic connections, attractor networks are widely believed to underlie biological memory systems and have been used extensively in recent years to model the storage and retrieval process of memory. In this paper, we propose a new energy function, which is nonnegative and attains zero values only at the desired memory patterns. An attractor network is designed based on the proposed energy function. It is shown that the desired memory patterns are stored as the stable equilibrium points of the attractor network. To retrieve a memory pattern, an initial stimulus input is presented to the network, and its states converge to one of stable equilibrium points. Consequently, the existence of the spurious points, that is, local maxima, saddle points, or other local minima which are undesired memory patterns, can be avoided. The simulation results show the effectiveness of the proposed method.

Entities:  

Mesh:

Year:  2015        PMID: 25960737      PMCID: PMC4417571          DOI: 10.1155/2015/191745

Source DB:  PubMed          Journal:  Comput Intell Neurosci


1. Introduction

Memory is a fundamental component of our human brain; how to simulate the human memory process has attracted many scientists attention in the research of cognitive systems and architectures [1, 2]. Attractor networks [3-6] have been one of the most popular models for memory storage and retrieval in recent decades since the hypothesis of attractor dynamics is supported and observed in the neocortex and hippocanpus in various memory experiments [6-9]. In general, an attractor network is a network of recurrently connected nodes in a biological network, whose states may settle to some stable patterns. One distinguish advantage is that the network can be represented by neurons and their synaptic connections. The particular pattern of such a recurrent network, which is called its “attractor” [10-13], can be stationary, time varying, or even stochastic. In theoretical neuroscience, different kinds of attractor neural networks have been associated with different functions, such as memory, motor behavior, and classification. In this paper, we consider the patterns or the so-called attractors as the stationary memory patterns stored in the dynamic system, which allows us to employ methods in dynamical systems to quantitatively analyze the characteristics such as the stability and robustness of the network. Usually researchers design an attractor network and then propose its energy function, often called Lyapnov function, to analyze the network [14-16], since the energy function plays a very important role in analyzing the network stability and robustness. In this paper, we design the memory dynamics in an opposite direction. We first propose an energy function and then design the attractor network. The energy function actually contains the desired memory patterns we wish to store. Different from Hopfield networks [10, 14], we introduce multiplicative algebra into the attractor network. This is biologically possible since both addition algebra and multiplication algebra are the simplest and the most widespread of all operations in the nervous system [17]. In addition, the value of our proposed energy function is nonnegative and attains zero values only at the patterns stored in the network. This makes it easy to distinguish the desired memory patterns from some other possible undesired patterns which are called spurious points [18]. It is shown that the memory patterns are stored as the stable equilibrium points of the dynamical attractor networks, which are also the local minimum points of the energy function. Compared with existing results in attractor networks [7, 10, 14], the patterns are not necessarily binary and uncorrelated in this paper. Binary patterns simplify the network design significantly as seen in [14]. Also, the uncorrelated patterns give a minimum interactions of the interactions of the network, which makes its behavior analysis much easier as seen in [7, 19]. On the other hand, when a stimulus pattern is presented as an initial input of the dynamic system, the states of the system will converge to a particular stored attractor iteratively. This process is called associative memory retrieval. The “associative” means that the memory we retrieved was by its informational content rather than by names, addresses, or relative positions. One very touchy problem in the retrieval process is how to overcome the problem that the system states may converge to spurious points; see, for example, [18, 20]. Usually, there are two kinds of equilibrium points for a dynamic system, the stable ones and unstable ones. It is shown that only the stable equilibrium points, that is, the local minimum points, of the energy function exist in the proposed designed system. We also prove that those local minima are only the memory patterns stored in the network. Thus, the spurious points, that is, local maxima, saddle points, or other local minima which are the undesired memory patterns, can be avoided. The contributions of this paper are summarized as follows. Firstly, we have proposed a new energy function different from the energy function in Hopfield networks. This makes it easy to differentiate memory patterns from possible spurious points. Secondly, we have presented an attractor network design based on the proposed energy function. The patterns stored in the attractor network can be nonbinary and either correlated or uncorrelated. Finally, we have proven that, when an arbitrary input stimulus is presented to the designed attractor network, the states converge to one of the stored patterns. This implies that there are no spurious points in the designed dynamical systems. The rest of the paper is organized as follows. Some background knowledge is reviewed in Section 2. Section 3 introduces the main design method of the proposed attractor network. The convergence properties are analyzed in Section 4. The simulation examples are shown in Section 5. Finally, the paper is concluded in Section 6.

2. Background Knowledge

2.1. Multiplication Algebra in Nervous Systems

Addition algebra is both the simplest and one of the most widespread of all operations in nervous systems. However, as pointed in [17], a number of biological mechanisms could, in theory, implement a multiplication algebra. Actually multiplication can be implemented based on addition. For example, when we multiply two signals x and y, we can logarithmically transform the two, add the result,and then apply an exponential: Thus, later it can be seen that the memory dynamics in our proposed attractor network could be implemented in a network with neurons and synapses.

2.2. Notations and Definitions

Denote a nonlinear function f(x) and a pattern x = [x 1 ⋯ x ] ∈ R where m is the dimension. For a nonlinear dynamical system , we have the following definitions.

Definition 1 .

A pattern x 0 ∈ R is called an equilibrium point of , if f(x) is a zero vector at x = x 0, which is denoted as f(x 0) = 0.

Definition 2 .

A pattern x 0 ∈ R is called a stable equilibrium point of , if f(x) = 0 at x 0 and the Jacobian matrix at x 0 is a positive definite matrix.

Definition 3 .

A pattern x 0 ∈ R is called an unstable equilibrium point (saddle point) of , if f(x) = 0 at x 0 while the Jacobian matrix at x 0 is a negative semidefinite or an indefinite matrix.

3. A New Energy Function and the Attractor Network

A classical energy based attractor network is the Hopfield network invented by Hopfield [10, 14], which serves as content addressable memory systems with binary threshold nodes. Although Hopfield networks are guaranteed to converge to a local minimum, they may converge to a spurious point (undesired memory pattern) rather than a stored pattern (desired memory pattern). To solve this problem, we propose a new energy function, and an attractor network is then designed based on the proposed energy function, in which the patterns are not necessarily binary and uncorrelated. Finally it is concluded that the memory patterns are the stable equilibrium points of the dynamic system and spurious points can be avoided. Assume that x 1,…, x ,…, x with x = [x 1  ⋯ x ] ∈ R are n different stationary patterns that we wish to store. Our objective is to design an attractor network to store these patterns such that x for k = 1,…, n can be retrieved when an input stimulus is located around the neighborhood of x . Before presenting our proposed method, we would like to point out that an ideal attractor network should preserve the following two properties.

Property 1 .

x 1,…, x ,…, x are stable equilibrium points of the attractor network.

Property 2 .

x 1,…, x ,…, x are the only stable equilibrium points of the attractor network. The energy function is designed as the following form:where d (x) = (x − x )(x − x ) is an Euclidian squared distance between x and x . The energy function is the product of the distance d (x) for k = 1,…, n. For the energy function F(x), it can be checked that ∀x ∈ R , F(x) > 0, and F(x) = 0 if and only if x = x for k = 1,…, n.

Remark 4 .

As mentioned in the Introduction, compared with the energy function in the Hopfield network [10, 14], the value of the proposed energy function F(x) is nonnegative and attains zero value only on the memory patterns stored in the network. This makes it easy to distinguish the memory patterns from spurious points [18] which may exist in an attractor network. Note that the energy function attains its minimum only on the memory patterns x 1,…, x . This inspires us to design a dynamical system which can make the energy function decrease iteratively. The gradient ∇F(x) and the Hessian matrix ∇2 F(x) are given bywhere I is a m dimensional identity matrix. Let v = [v 1 ⋯ v ] be a random noise vector such thatwhere v max⁡ is a chosen small positive constant. Now an attractor network that decreases the energy function F(x) iteratively is represented aswhere the scalar constant κ is chosen asWe rewrote the differential equation for each neuron x as for i = 1,…, m with Equations (7) and (8) imply that ith neuron updates its state x according to the synaptic inputs collected from other neurons via synaptic connection strength ω , in the presence of a random noise v .

Remark 5 .

The Hopfield model has the advantage that it can be represented by neurons and their synaptic connections. Through carefully observing (5)–(8), we note that our proposed attractor network can also be represented by neurons and their synaptic connections by introducing the multiplication algebra in Section 2.1. In addition, the value of our proposed energy function is nonnegative and attains zero values only at the patterns stored in the network. This makes it easy to distinguish the memory patterns from some other possible undesired patterns which are called spurious points [18]. It is shown that the memory patterns are stored as the stable equilibrium points of the dynamical system, which are also the local minimum points of the energy function. Compared with existing results in attractor networks [7, 10, 14], the patterns are not necessarily binary and uncorrelated in this paper. Binary patterns simplify the network design significantly as seen in [14]. Also, the uncorrelated patterns give a minimum interaction of the interactions of the network, which makes its behavior analysis much easier as seen in [7, 19].

Remark 6 .

As mentioned earlier, there are two kinds of equilibrium points for a dynamic system: stable ones and unstable ones. In the next section, it will be shown that only the stable equilibrium points exist in the above dynamical system. For an arbitrary initial stimulus, the states of the dynamical system converge to one of its stable equilibrium points, that is, the local minimum points of the energy function, which cannot be its local maximum points or saddle points of the energy function. Thus, Property 1 can be achieved in the design. In the next section, it will also be proven that Property 2 can also be achieved in the design; that is, x 1,…, x are the only local minimum points of F(x).

4. Convergence Analysis

Lemma 7 .

For square matrix A ∈ R , assume that the eigenvalues of A are not all the same. If tr⁡(A) ≤ 0, A cannot be a positive semidefinite or positive definite matrix.

Proof

Denote the eigenvalues of A as λ 1,…, λ . Note that tr⁡(A) = ∑ λ . If tr⁡(A) = ∑ λ < 0, then ∃λ such that λ < 0. If tr⁡(A) = ∑ λ = 0, there also exists ∃λ such that λ < 0 since λ 1,…, λ cannot be identically equal. Then this lemma holds.

Lemma 8 .

For conformable matrices A, B, and C, tr⁡(ABC) = tr⁡(BCA) = tr⁡(CAB). Also, if matrices A and B are addable, tr⁡(A + B) = tr⁡(A) + tr⁡(B) [21]. To prove tr⁡(ABC) = tr⁡(BCA) = tr⁡(CAB), actually we only need to prove that tr⁡(AB) = tr⁡(BA). As tr⁡(AB) = ΣΣ A B = tr⁡(BA) and tr⁡(A + B) = Σ(A + B ) = Σ(A ) + Σ(B ) = tr⁡(A) + tr⁡(B), this lemma holds.

Theorem 9 .

The states of the attractor network in (5)–(8) converge to an stable equilibrium point x ∗ ∈ {x 1,…, x } such that ∇F(x ∗) = 0, F(x ∗) = 0, and ∇2 F(x ∗) is positive definite which is denoted as ∇2 F(x ∗) > 0. By Definition 1, the equilibrium points of system (7) are such thatAs v is a random vector and κ is chosen by (6), a point is an equilibrium point of system (7) if and only if F(x) = 0 gives us that all the possible equilibrium points of (7) are those points at which the values of energy function F(x) are zero; that is, x ∗ ∈ {x 1,…, x }. In addition, the Jacobian matrix at a point x is the Hessian matrix ∇2 F(x ). Then, it is easy to obtain thatSo x ∗ is a stable equilibrium point of the attractor network in (5) to (8) based on Definition 2.

Theorem 10 .

For the case that the dimension m ≤ 2, x 1,…, x are the only local minimum points of F(x). Obviously, x 1,…, x are local minimum points (and also global minimum points) of F(x). We now prove that F(x) has no other local minimum points. LetIt can be obtained that ∇2 F(x) = A + B. It is also known that the dimension of diagonal matrix A is m × m; thenAlso, Lemma 8 gives thatNow we assume that there is a point which is different from x 1,…, x but satisfies that ∇F(x ∗) = 0. This implies that (∇F(x ∗))∇F(x ∗) = 0. This is to say that where Thus, we have G(x ∗) = 0 if x ∗ is different from x 1,…, x . Let A ∗ and B ∗ be A and B at x = x ∗. Combining (13)-(14) and (16), it can be obtained thatSince x 1,…, x are all different from each other, the eigenvalues of ∇2 F(x ∗) cannot be all the same. When m = 1, we have G(x ∗) = tr⁡(2A ∗ + B ∗) = 0, which implies tr⁡(A ∗ + B ∗) < 0 since A ∗ is a positive definite matrix. Similarly, we have tr⁡(∇2 F(x ∗)) = 0 when m = 2. From Lemma 7, matrix ∇2 F(x ∗) cannot be a positive definite matrix. It is a seminegative definite or an indefinite matrix. This implies that x ∗ cannot be a local minimum point though ∇F(x ∗) = 0. It can be a local maximum point or a saddle point of F(x).

Remark 11 .

If the dimension m > 2, tr⁡((2/m) · A ∗ + B ∗) = 0 does not directly imply that ∇2 F(x ∗) is a negative semidefinite or an indefinite matrix. Let P ∈ R 2× be a full-row rank projection matrix, which projects a vector into a two-dimensional plane. The following theorem shows that x 1,…, x are still the only local minimum points of F(x).

Theorem 12 .

For the case that m > 2, x 1,…, x are also the only local minimum points of F(x). Let and be the corresponding projection of the local minimum points x 1,…, x in the two-dimensional plane. Then, F(x) and ∇F(x) become and , respectively, on this two-dimensional plane. It can be obtained thatIf ∇F(x) = 0 and ∇2 F(x) > 0, then and , which means that the local minimum points of F(x) in a higher dimensional space must be also the local minimum points of in the two-dimensional plane while the converse is not true. Obviously, x 1,…, x are the local minima of . Assume that there is a point x ∗ which is different from x 1,…, x but a local minimum of F(x). There exists a full-row rank projection matrix P ∗ ∈ R 2× such that . The full-row rank of P ∗ implies that is different from , where for i = 1,…, n are the local minimum points of . However, this is impossible by Theorem 10. Thus, there is no other local minimum point of F(x). If a point x ∗ satisfies that ∇F(x ∗) = 0 but it is different from x 1,…, x , it can be only a local maximum point or a saddle point. So this theorem holds.

5. Simulation Results

Example 1 .

Design a nonlinear dynamical system whose attractors are x 1,…, x ∈ R with n = 2, m = 1. This corresponds to the one-dimensional case. The energy function is constructed as F(x) = (x − x 1)2(x − x 2)2 from (2) with ∇F(X) and ∇2 F(X) being given byrespectively. The dynamic system can be then designed in (5)–(8). By solving ∇F(x ∗) = 0, we have x ∗ = x 1, x ∗ = x 2, or x ∗ = (x 1 + x 2)/2 as analyzed in Section 3. From x ∗ = (x 1 + x 2)/2 is a local maximum point of F(x). So x 1 and x 2 are the only two local minimum points of F(x). The energy function is shown in Figure 1 where “∗” denotes the local minimum point and “Δ” denotes the local maximum or the saddle point. From Figure 1, when the initial stimulus x < (x 1 + x 2)/2, the states converge to x 1; when x > (x 1 + x 2)/2, the states converge to x 2. If the initial stimulus x = (x 1 + x 2)/2, the states can converge to either x 1 or x 2, which depends on the random noise v 1 in (7).
Figure 1

Energy function for one-dimensional case.

Example 2 .

Example 1, analyze the dynamic system when n = 2, m = 2 and n = 2, m = 3, respectively.

Case 1 .

x 1 = [−0.5 −0.5]′ and x 2 = [0.5 0.5]′. Similar to Example 1, the energy function is constructed as F(x) = d 1(x)d 2(x) with ∇F(x) and ∇2 F(x) being given bySolving ∇F(x ∗) = 0 gives x ∗ = x 1, x ∗ = x 2, or X ∗ = (x 1 + x 2)/2. When X ∗ = (x 1 + x 2)/2, we haveThe two eigenvalues of ∇2 F(x ∗) are λ 1 = 2 and λ 2 = −2. This is consistent with Theorem 10, which gives tr⁡(∇2 F(x ∗)) = 0. Figure 2 shows the contour map of the energy function in a two-dimensional space. The contour lines around a saddle point look like a horse saddle. It can be concluded that x 1 and x 2 are two attractors while (x 1 + x 2)/2 is not an equilibrium point of system (5)–(8) by Theorem 9. The term κ · F(x)v guarantees that the attractor network (5)–(8) cannot stay at the point (x 1 + x 2)/2. But from Theorem 10, we know that (x 1 + x 2)/2 is a saddle point of F(x).
Figure 2

The contours of the energy function for Example 2.

Case 2 .

x 1 = [−0.5 −0.5 −0.5]′ and x 2 = [0.5 0.5 0.5]′. In this case, n = 2, m = 3. When x ∗ = (x 1 + x 2)/2, we haveThen x ∗ is still not an equilibrium point of system (5)–(8) but a saddle point of F(x) as the eigenvalue of ∇2 F(x ∗) is now λ 1 = −3, λ 2 = 3, and λ 3 = 3. But in this case, we have tr⁡(∇2 F(x ∗)) > 0. So we cannot determine whether x ∗ is a saddle point or a local minimum point of F(x) only based on Theorem 10. However, we can determine that it is also a saddle point by Theorem 12. This is consistent with what we observed in simulation, as the eigenvalues of above ∇2 F(x ∗) is 3,3, −3, and thus x ∗ is a saddle point and cannot be a local minimum point of F(x).

Example 3 .

Design a nonlinear dynamic system whose attractors are x 1 = [2 0]′, , and . We have n = 3 and m = 2 in this example. The energy function is constructed as F(x) = d 1(x)d 2(x)d 3(x) withwith its dynamical property described by the attractor network in (5)–(8). Theorem 9 tells us that x 1, x 2, and x 3 are the stable equilibrium points of F(x) and the attractor network in (5)–(8). However, F(x) has more saddle points but these saddle points are not equilibrium points (saddle points) of (5)–(8). To illustrate this, firstly, we find all the points such that ∇F(X ∗) = 0. If ∇F(X ∗) = 0, then (∇F(X ∗))∇F(X ∗) = 0, which meanswhere We know that (∇F(x ∗))∇F(x ∗) = 0 gives x ∗ = x 1, x ∗ = x 2, x ∗ = x 3 or G(x ∗) = 0. But G(x ∗) = 0 implies that Thus, x ∗ will be a saddle point of F(x) but not the attractor network in (5)–(8) if ∇F(x ∗) = 0 while x ∗ is different from x 1,…, x . As the attractor network in (5)–(8) does not have any saddle points from Theorem 9, usually, a saddle point of F(x) is located in between two local minimum points. As seen from Figure 3, two saddle points of F(x) are in between the three local minimum points on the plane. One is located about (−1.05,0.22) and the other is about (0.58,0.31).
Figure 3

The contours of the energy function for Example 3.

6. Conclusion

The contributions of this paper are summarized as follows.Our future work is to construct a biological plausible dynamical system in hardware (neuromorphic chip) which can stimulate the behavior of the designed network. This sheds new lights on the research towards the realization of artificial cognitive memory. We have proposed a new energy function which includes the information of the stored patterns, and it is different from the energy function in Hopfield network. The proposed energy function makes it easy to differentiate memory patterns from possible spurious points. We have presented an attractor network design based on the proposed energy function. The patterns stored in the attractor network can be nonbinary and either correlated or uncorrelated. The memory patterns are the attractors of the network and the equilibrium points of the dynamic system. When an arbitrary input stimulus is presented to the designed attractor network, it has been proved that the states converge to one of the stored patterns. There are no spurious states in the designed dynamic systems.
  18 in total

Review 1.  The role of single neurons in information processing.

Authors:  C Koch; I Segev
Journal:  Nat Neurosci       Date:  2000-11       Impact factor: 24.884

2.  Localist attractor networks.

Authors:  R S Zemel; M C Mozer
Journal:  Neural Comput       Date:  2001-05       Impact factor: 2.026

3.  Spike-driven synaptic dynamics generating working memory states.

Authors:  Daniel J Amit; Gianluigi Mongillo
Journal:  Neural Comput       Date:  2003-03       Impact factor: 2.026

4.  A robust method for distinguishing between learned and spurious attractors.

Authors:  Anthony V Robins; Simon J R McCallum
Journal:  Neural Netw       Date:  2004-04

5.  Memory dynamics in attractor networks with saliency weights.

Authors:  Huajin Tang; Haizhou Li; Rui Yan
Journal:  Neural Comput       Date:  2010-07       Impact factor: 2.026

6.  An energy function-based design method for discrete Hopfield associative memory with attractive fixed points.

Authors:  Mehmet Kerem Müezzinoglu; Cüneyt Güzeliş; Jacek M Zurada
Journal:  IEEE Trans Neural Netw       Date:  2005-03

7.  Attractor neural networks and spatial maps in hippocampus.

Authors:  Misha Tsodyks
Journal:  Neuron       Date:  2005-10-20       Impact factor: 17.173

8.  A controlled attractor network model of path integration in the rat.

Authors:  John Conklin; Chris Eliasmith
Journal:  J Comput Neurosci       Date:  2005 Mar-Apr       Impact factor: 1.621

9.  Neuroscience. Attractors in memory.

Authors:  Bruno Poucet; Etienne Save
Journal:  Science       Date:  2005-05-06       Impact factor: 47.728

10.  Neural networks and physical systems with emergent collective computational abilities.

Authors:  J J Hopfield
Journal:  Proc Natl Acad Sci U S A       Date:  1982-04       Impact factor: 11.205

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.