Daniel Miner1, Christian Tetzlaff1,2. 1. Third Institute of Physics, Georg-August-Universität Göttingen, Göttingen, Germany. 2. Bernstein Center for Computational Neuroscience, Georg-August-Universität Göttingen, Göttingen, Germany.
Abstract
In the course of everyday life, the brain must store and recall a huge variety of representations of stimuli which are presented in an ordered or sequential way. The processes by which the ordering of these various things is stored and recalled are moderately well understood. We use here a computational model of a cortex-like recurrent neural network adapted by a multitude of plasticity mechanisms. We first demonstrate the learning of a sequence. Then, we examine the influence of different types of distractors on the network dynamics during the recall of the encoded ordered information being ordered in a sequence. We are able to broadly arrive at two distinct effect-categories for distractors, arrive at a basic understanding of why this is so, and predict what distractors will fall into each category.
In the course of everyday life, the brain must store and recall a huge variety of representations of stimuli which are presented in an ordered or sequential way. The processes by which the ordering of these various things is stored and recalled are moderately well understood. We use here a computational model of a cortex-like recurrent neural network adapted by a multitude of plasticity mechanisms. We first demonstrate the learning of a sequence. Then, we examine the influence of different types of distractors on the network dynamics during the recall of the encoded ordered information being ordered in a sequence. We are able to broadly arrive at two distinct effect-categories for distractors, arrive at a basic understanding of why this is so, and predict what distractors will fall into each category.
In order to successfully plan actions, the brain must store the order in which various sub-actions need to be executed. Simple memory tasks as well, such as counting in common base-10, require memories of the ordering of concepts or learned stimuli. It is therefore clear that the storage and recall of ordered sequences is a critical and fundamental function of the brain. During rest and planning, sequences are observed to be rapidly recalled in the hippocampus [1, 2], and the cortex is tightly attuned to learned sequences as well [3, 4]The primary mechanism of learning in the brain is believed to be Hebbian synaptic plasticity, that is, changes in the strength (and potentially structure) of connections in the synaptic matrix as a function of the activity of the neurons on the pre- and post-synaptic side of connection. One of the most prominent models of this is spike timing-dependent plasticity (STDP), a form of learning specifically correlated with the time difference between pre- and post-synaptic firing [5, 6]. The exploitation of this time difference provides a natural and unsupervised candidate for a mechanism for storage of temporally ordered stimuli, symbols, or actions, as it will pick up on groups of neurons that are activated one right after the other [7, 8].Using this learning mechanism in a computational model, we are able to train the recurrent excitatory weight matrix of a neural network modeled after a small slice of cortex to store a sequence of symbols. The recall of this stored sequence, upon activation by a cue, is rapid and independent of the training speed, preserving only the order, as is observed in hippocampal replay [1, 2]. This builds off the previous work done with a similar network [9], which simulated a section of visual cortex to examine how exposure to stimulus motion in a particular direction effects subsequent response to such stimuli. As this is a simulation-based study, we have access to the entire synaptic matrix, and its examination suggests that, at least to the zeroth order, the sequence is stored in a representation that demonstrates properties and structures similar to both synfire chains and cell assemblies. The rapid replay indicates that the dynamics of the synfire element dominate the storage (as opposed to slow switching between recurrent assemblies, for example).We focus, in this study, on effects on what we refer to as the“primary representation.” By example, we mean the activity of the neurons which are intrinsically tuned to whatever stimulus we concern ourselves with (in this case, meaning they receive direct feedforward input from the stimulation source), and not neurons that are activated via circuitous loops or second-order effects.Our primary interest is the effect of distraction on this replay, something that is heretofore relatively unstudied in both simulation and experiment (though hints of some of the results we observe can be extracted from early network simulation studies [10]). We will, following training of model networks, apply several types of distraction to the sequence recall process and systematically observe the results. That is to say, in simpler terms—what happens if, once a replay or recall process has started, the brain area in question receives a strong new input or cue? How does this affect the recall process, and can we predict, from some properties of the new input (or distractor), what the effects will be?Following observations of recall under an array of distractor locations and timings, we conclude that we can broadly classify two distinct classes of distractor as a function of their effect, namely “relevant” (or highly disruptive) and “irrelevant” (minimally disruptive) distractors. We are then able to elucidate which class the distractor will fall into based on its spatio-temporal location relative to an instance of cued rapid recall. An interesting observation is that the distractors seem to fall strongly into one category or the other, rather than existing on a spectrum, and that the system is robust to a class of distractors even in the absense of any kind of top-down attention or guidance mechanism.
Materials and methods
Model basics and neuronal dynamics
Our basic model architecture consists of a recurrent excitatory and inhibitory reservoir with a separate input layer and specific plasticity mechanisms, and is shown in Fig 1.
Fig 1
Network architecture.
A symbolic diagram of the network architecture for the virtual experiment. The input layer, which provides learning signals, recall cues, and distractors in the form of spikes, and its (feedforward) connections to the main network are shown in blue. The excitatory portion of the network and its associated excitatory connections (both recurrent and feedforward) are shown in green. The inhibitory portion of the network and its associated inhibitory feedback connections are shown in red.
Network architecture.
A symbolic diagram of the network architecture for the virtual experiment. The input layer, which provides learning signals, recall cues, and distractors in the form of spikes, and its (feedforward) connections to the main network are shown in blue. The excitatory portion of the network and its associated excitatory connections (both recurrent and feedforward) are shown in green. The inhibitory portion of the network and its associated inhibitory feedback connections are shown in red.We base our model on the SORN family of neural networks, first introduced as a binary neuron model [11] and later elaborated upon as a spiking model [12]. We further update the model, upgrading the code to the Brian 2 simulator platform [13]. The neuronal dynamics are determined by a conductance-based leaky integrate-and-fire model, with an adaptive intrinsic firing threshold, as follows:Here, v is membrane voltage of the postsynaptic neuron j, gleak is the leak conductance, vrest is the resting potential, cmembrane is membrane capacitance, τx is the time constant for feature x, g[ampa,gaba] is the conductance for each neurotransmitter type, e[ampa,gaba] is the reversal potential for each neurotransmitter type, are the adaptation rate and increment, respectively, for the intrinsic firing threshold plasticity, is the firing threshold of neuron j, and w is the strength of the synaptic connection from presynaptic neuron i to postsynaptic neuron j. ξ is an Ornstein-Uhlenbeck noise generator, and σnoise is the noise variance. Iext is any external current applied to the neuron. w is the strength of any synaptic connection that might exist from neuron i to neuron j, and is equal to 1 at the moment neuron i spikes, and 0 otherwise. Of the given variables, those that function as fixed value parameters across all neurons (or all neurons of a certain type) have their values listed in Table 1. Parameter values were estimated from assorted surveys of local neocortical circuitry [14].
Table 1
Shared simulation parameters.
parameter
value
parameter
value
gleak
30 nS
vrest
-70 mV
cmembrane
300 pF
τmembrane
20 ms
τampa
2 ms
τgaba
5 ms
eampa
0 mV
egaba
-85 mV
ηdecayip
0.2 mV/second
ηspikeip
0.066 mV
σnoise
1 mV
Wtotal
20 nS
A+
0.05 nS
A−
0.05 nS
τ+
20 ms
τ−
20 ms
Simulation parameter values which are not listed in the text and which are shared across all trials.
Simulation parameter values which are not listed in the text and which are shared across all trials.
Plasticity mechanisms
It should be noted that already, the adaptive firing threshold in the neuron model (integrated into Eqs 5 and 6) can be considered an intrinsic or neuronal homeostatic plasticity mechanism.The synaptic weights w, when between two excitatory neurons, can be modified by two plasticity mechanisms. The first is spike timing-dependent plastcity (STDP) [5, 6, 15], a process which links weight changes to the temporal relationship between pre- and postsynaptic spikes. We define our STDP model as follows [16]:Here, i and j remain the pre- and postsynaptic indices, n and f are temporal indices of (temporally) nearby spikes, A+/− is learning amplitude, τ+/− is decay time constant, and + and − signify potentiation and depression, respectively. Despite the arbitrarily large sums, the time constants and average firing rates involved in our typical operating regime allow us to take a nearest-neighbor approximation in software implementation, speeding up simulation time [13].At the same time, we implement a synaptic normalization mechanism, executed upon each STDP-induced weight change, a mechanism inspired by biological observations of [11, 17, 18]. This is a form of homeostatic synaptic plasticity. Mathematically, it is modeled as follows:Here, W is the vector of incoming weights to neuron i, N is the length of that vector, and Wtotal is the target value for the total incoming weight (which is the same for all neurons of a given type).
Connectivity, setup, and shared model parameters
We initialize our model network with a population of n = 200 excitatory neurons with absolute refractory periods of and n = 40 inhibitory neurons with absolute refractory periods of . We recurrently connect the excitatory pool (without self-connections) and interconnect the inhibitory and excitatory pool (in both directions) with random sparse connections with probability pconnect = 0.2. Recurrent inhibitory connections are neglected. Recurrent excitatory connections are given an initial strength of w = 0.5 nS, and all other connections are given an initial strength of w = 1.0 nS. Again, parameter estimates are derived from various surveys of local cortical circuitry [14]. The rest of the shared simulation parameters are given in Table 1.
Learning and input protocol
The experimental protocol consists of of several steps, as follows, and is illustrated in Fig 2:
Fig 2
Experimental protocol.
A symbolic diagram and description of the experimental protocol.
Warm-up The network is activated for a period to allow all dynamical variables to converge to an equilibrium distribution.Training The network is trained on a repeating sequence.Relaxation Plasticity is turned off and the homeostatic mechanisms are allowed to restabilize.Testing The network is presented with recall cues.Experimental The network is presented with distractor signals as well and observed. b.Control The network is observed with only the recall cues.
Experimental protocol.
A symbolic diagram and description of the experimental protocol.
Warm-up
The warm-up phase is necessary because network initialization does not necessarily occur with all dynamical variables already in an equilibrium distribution. The membrane potentials, synaptic weights, and firing thresholds are initialized to random values within their standard operating range, and the network is then left to run for 50 seconds while convergence to equilibrium statistical distributions occurs. The basic model architecture for which the simulation is initialized is shown in Fig 1 and described in the “Connectivity, setup, and shared model parameters” subsection.
Training
Starting with the training process, the excitatory neurons in the network are randomly divided into 10 non-overlapping groups of 20 neurons each. The first 5 of these groups (which we label A, B, C, D, and E) recieve sequentially activated inputs. Each group is strongly connected (20 nS) to a 50 Hz Poisson spike source. For each training block (a one second, single-sequence training instance), each Poisson source is sequentially activated in order for a period of 100 ms, and then deactivated as the next one is activated. After all 5 sources have been activated in sequence, a 500 ms rest period occurs before the next training block (see Fig 3 for an example). This process is continued for 50 seconds (or blocks).
Fig 3
Training activity.
A spike raster plot of a section of training activity for one of the trials. Different clusters are highlighted by different colors, and training stimulation is represented by the wavy line overlay.
Training activity.
A spike raster plot of a section of training activity for one of the trials. Different clusters are highlighted by different colors, and training stimulation is represented by the wavy line overlay.
Relaxation
The relaxation period serves a similar purpose to the warm-up phase. Due to the elevated average input activity during the training phase, the adaptive firing thesholds reach a new equilibrium distribution. When this activity ceases, they need time to return to their resting equilibrium distribition. At the start of this phase, all Poisson sources and synaptic plasticity mechanisms are deactivated, and the trained network is then allowed to re-equilibrate for 50 seconds.
Testing
In the testing phase, a precisely timed instantaneous spike source (i.e. a source providing a synchronous burst, as opposed to the Poisson rate-based sources used in training) is connected to the neurons of the first element in the trained sequence. It is activated for a single burst and acts as a recall cue for the first element, triggering rapid replay of the learned sequence. With a variable time delay, an additional instantaneous spike source is activated as well, serving as a distractor signal. It may be connected to either the first sequence element group, the third (middle) sequence element group, the fifth (last) sequence element group, or the sixth group, which is not part of the trained sequence. It may be presented simultaneously with the recall cue, or with a delay of 1, 2, or 3 ms (the rapid replay of the entire trained 5 element sequence typically takes between 5 and 7 ms, so this is approximately 1/5, 2/5, and 3/5 of the replay time). This procedure is repeated twice per second for 100 seconds. Each trial yields slightly different results due to the membrane noise and the general nature of recurrent neural networks. Representative examples of this distracted replay can be seen in the Results section. Only one spatio-temporal distractor location is tested per simulation run. This is the experimental subphase. Following this, for an additional hundred seconds, presentation of the recall cue continues identically, but the distractor signal is deactivated. This is the control subphase. An example of this recall can be seen in Fig 4.
Fig 4
Control activity.
Spike raster plots of a section of control activity, demonstrating rapid replay. Time given relative to cue onset. Presented as both unsorted and sorted to demonstrate subtlety of structre. In the sorted case, different population groups, or clusters, are highlighted by different colors.
Control activity.
Spike raster plots of a section of control activity, demonstrating rapid replay. Time given relative to cue onset. Presented as both unsorted and sorted to demonstrate subtlety of structre. In the sorted case, different population groups, or clusters, are highlighted by different colors.This full process is repeated a total of 80 times, consisting 5 trials for each of the 16 distractor conditions (4 possible locations and 4 possible times).
Evaluation measures
In order to analyze the effects of the distractors, we first need to understand our control conditions and develop measures of effect. Firstly, we approach the control condition. We concern ourselves primarily with the temporal accuracy and binary presence of the sequence elements during replay. In order to do this, we first, for each trial, pool the spike traces for the neurons in each of the 5 trained neuron groups. We then convolve this with a 2 ms Gaussian kernel, producing a population rate trace for each group. For each cue, we take a -10 to +25 ms window around the cue, threshold above 10 Hz, and run a peak detector (using the PeakUtils Python module) to obtain the position of the maximum rate for each neuron group. We then histogram the times across all cues and trials in which all elements had a peak above the threshold (96% passing rate for the control condition), producing a mean time (relative to the cue) and timing variance for each sequence element in the undistracted recall condition.Similar pooling (and thresholding) of peak timing into histograms is done for each of the 16 trial conditions, with all conditions having readout success rates of 0.95 or higher. We can qualitatively compare the peak time histograms from this with the one for the control condition (which will be discussed in the results section). We wish as well to develop quantitative distraction measures. For this, we propose two novel measures: the “disruption index” and the “deviance index.”The deviance index is based on the simple difference between mean peak times for the control condition, and peak times for the experimental condition. As such, it is designed to emphasize the bulk deviation from the control condition, and is defined as follows for a single trial i:Here, Nelements = 5 is the number of sequence elements, is the peak time of element n for trial i, is the mean peak time of element n for the control condition, and is the variance of the distribution of peak times for element n in the control condition.The disruption index is based on the differences in times between subsequent sequence elements, comparing this set of differences for each experimental trial i and for the control condition. As such, is it designed to emphasize relative rather than absolute disruption of sequence recall. The variable definitions are the same as for , and it is defined as follows:Using these two measures, as well as the previously mentioned qualitative analysis, we evaluate the effects of the distractors on rapid sequence replay.
Results
We ask three primary questions here regarding distracted recall. Firstly, we ask what are the effects of the distractor, i.e. what happens to the recall process in the presence of such a signal. The second question we ask is, can we categorize or classify these effects in a meaningful way? The third question is, assuming we have successfully categorized distractor effects, can we predict which effect category a particular distractor will tend to fall into, and if so, how?We find that we can broadly classify distractors into two different classes, which we refer to “irrelevant” and “relevant” distractors. What we mean by irrelevant is, distractors that tend to have the effect of adding noise to the replay timing or amplitude but do not tend to disrupt the ordering of the replay. Relevant distractors, on the other hand, actively disrupt the replay, disordering it or leading to a high number of failed recalls. We find that the relevant distractors tend to be trained stimuli that are presented prior to their “natural” recall time in the replay process, and irrelevant distractors may be untrained stimuli, or trained stimuli presented at or after their “natural” recall time in the replay process.
Connections and representations
Before examining distractor behavior, we wish to examine how the sequences are stored in the network after the learning process (Fig 5). In order to do this, we look closely at the recurrent excitatory connection matrix. We histogram all the connection weights, the trained recurrent weights, and six categories of connection (all thresholded for above-zero values) (Figs 6, 7 and 8) from five pooled control-only runs. The categories are “one forward” (i.e. the next trained sequence element), “n forward” (i.e. to any trained sequence element subsequent to the next one), “to external” (i.e. outgoing from a trained sequence element to the untrained portion of the network), “one backward” (i.e. to the last trained sequence element), “n backward” (i.e. to any trained sequence element previous to the last one), and “from external” (i.e. incoming from the untrained portion of the network to a trained sequence element).
Fig 5
Sorted weight matrix throughout training.
Demonstrative examples of the sorted weight matrix throughout sequence training. Arbitrary color scale.
Fig 6
Weight categories.
Demonstrative llustration of different weight categories for a simple 3-element sequence. “Recurrent” refers to the recurrent weights within each input cluster. “One forward (backward)” refers to feedforward (feedbackward) weights from one input cluster to the next (to the previous). “N forward (backward)” refers to feedforward (feedbackward) weights between not-immediately-sequential input clusters. “From external” refers to weights going from the background portion of the network to the input clusters, and “To external” refers to weights going from the input clusters to the background portion of the network.
Fig 7
Connection strength histograms.
Strength histograms of all connections (i.e. for the whole excitatory matrix) (A), and trained recurrent connections (within clusters) (B), along with means and standard deviations for each distribution.
Fig 8
Connection strength histograms for important categories of connections.
Strength histograms of the important categories of connections, along with means and standard deviations for each distribution. (A) “one forward,” or connections from one input cluster to the subsequent input cluster. (B) “n forward,” or connections from an input cluster to other input clusters occuring more than one step after them in the trained sequence. (C) “to external,” or connections going from input clusters to untrained (or at least indirectly trained) portions of the netork. (D) “one backward,” or connections from one input cluster to the immediately previous one in the trained sequence (the low count is due to rounding, thresholding, and sparse matrix storage artifacts). (E) “n-backward,” or connections from an input cluster to earlier (but not immediately previous) input clusters in the trained sequence. (F) “from external,” or connections from the untrained (or indirectly trained) portion of the network to the input clusters.
Sorted weight matrix throughout training.
Demonstrative examples of the sorted weight matrix throughout sequence training. Arbitrary color scale.
Weight categories.
Demonstrative llustration of different weight categories for a simple 3-element sequence. “Recurrent” refers to the recurrent weights within each input cluster. “One forward (backward)” refers to feedforward (feedbackward) weights from one input cluster to the next (to the previous). “N forward (backward)” refers to feedforward (feedbackward) weights between not-immediately-sequential input clusters. “From external” refers to weights going from the background portion of the network to the input clusters, and “To external” refers to weights going from the input clusters to the background portion of the network.
Connection strength histograms.
Strength histograms of all connections (i.e. for the whole excitatory matrix) (A), and trained recurrent connections (within clusters) (B), along with means and standard deviations for each distribution.
Connection strength histograms for important categories of connections.
Strength histograms of the important categories of connections, along with means and standard deviations for each distribution. (A) “one forward,” or connections from one input cluster to the subsequent input cluster. (B) “n forward,” or connections from an input cluster to other input clusters occuring more than one step after them in the trained sequence. (C) “to external,” or connections going from input clusters to untrained (or at least indirectly trained) portions of the netork. (D) “one backward,” or connections from one input cluster to the immediately previous one in the trained sequence (the low count is due to rounding, thresholding, and sparse matrix storage artifacts). (E) “n-backward,” or connections from an input cluster to earlier (but not immediately previous) input clusters in the trained sequence. (F) “from external,” or connections from the untrained (or indirectly trained) portion of the network to the input clusters.Several major features stand out. First of all, in Fig 7, we note that the overall distribution of weights in the matrix appears to be unimodal and heavy-tailed, as has been observed in biology [19]. We note as well the properties of the “trained recurrent” weights (i.e. in-group or in-cluster, or connections between all the neurons in a single input group). They appear to take on a bimodal distribution, with the medians of each peak being just above zero, and around 1.75 nS. This is not very surprising, as classical STDP weakens connections in one direction while strengthening them in the other (when connections in both directions exist) under most spiking conditions [20], and the entire cluster being stimulated at once will clearly lead to a high level of internal STDP-induced synaptic changes. Strong recurrency also suggests a high capacity for pattern completion [21] or, under the proper inhibitory and refractory conditions (which do not exist in this model), sustained self-activation of an individual sequence element.Looking at the overall weight distribution in Fig 7, we note that it appears unimodal and heavy-tailed, as has been noted in numerous experimental and theoretical studies examining the distribution of synaptic weights in cortical slices, suggesting the capacity for storage of complex representations [12, 19].The “one-forward” category, which appears to be unimodal and to possess a median similar to that of the higher component of the “trained recurrent” category, is by far the strongest connection category (meaning it has an exceptionally high peak value, like the “trained recurrent” category, and minimal excusrion into lower values, unlike the “trained recurrent” category). This is to be expected, as the population spiking order between one cluster and the next is always the same during training, leading to extremely strong potentiation in the sequential direction. Similarly, the “one-backward” category is exceedingly weak (with many connection matrix entries even being driven to zero, hence the low count in this category—an artifact of sparse matrix storage), as it is subject to the same effect in the opposite direction. This indicates that the network rather strongly stores the direction of the sequence, an indication vindicated by the observed strong rapid replay. It also predicts that the strongest effect of any trained distractor will be on itself and the subsequenct sequence element, as these two categories are the strongest. This begins to suggest that the sequence is stored in the form of a cell assembly for each element, which is in turn connected in a vaguely synfire chain-like fashion to the subsequent element. We can call this structure an assembly sequence.The “n-forward,” “n-backward,” and “from external” categories show relatively average connection strengths, indicating little in the way of second-order links and external influence on the sequence, allowing us to predict that the external distractor class will not have a strongly disruptive effect. The “to external” connection category of weights, on the other hand, is strongly bimodal, with medians slightly above zero and around 2 nS, which indicates that while each trained cluster does not project outward (aside from to the next cluster) very often, when it does, it does so rather strongly, suggesting the possibility to form directional associations with untrained elements in a patchy fashion. While this possible expansion of the representation into the background portion of the network is a potentially an intriguing future line of inquiry, it falls outside the scope of this paper.
The control condition
Recall that following initialization, training (after which synaptic plasticity is frozen), and relaxation, we present the network with a series of recall cues both with and without distractor signals, in order to obtain both experimental and control trials. In order to qualitatively understand distraction effects, we must first understand the undistracted (i.e. control) condition. The distribution of peak times is shown below (Fig 9), and the fraction of trials passing the threshold veto is 96% (as described in the “Evaluation measures” subsection). We can see immediately that replay is fairly reliable, with timing variance increasing slightly with each subsequent sequence element. The slight offset, which is present in all similar figures, is the result of neuronal conductance integration times.
Fig 9
Control peak distributions.
Distribution of peak recall times for each element in the control condition (i.e. undistracted recall). Plot (A) shows this as a violin plot (densities estimated with a 50 point Gaussian kernel) and plot (B) as overlapping histograms. Times are given relative to cue onset. The reliable ordering indicates a clear recall of the trained sequence.
Control peak distributions.
Distribution of peak recall times for each element in the control condition (i.e. undistracted recall). Plot (A) shows this as a violin plot (densities estimated with a 50 point Gaussian kernel) and plot (B) as overlapping histograms. Times are given relative to cue onset. The reliable ordering indicates a clear recall of the trained sequence.
External distractors
We define external distractors as those that are not spatially correspondent with any trained stimuli. We investigate how recall changes in the presence of these external distractors. We present the same sort of peak time distributions for the external experimental conditions (1 position and 4 delays), including the percentages of each condition passing the threshold veto (Figs 10 and 11). Also included is an example raster of distracted replay for each of the 4 timing conditions (which we have selected to examine how otherwise identical distractors at different temporal points in the recall process affect it).
Fig 10
Experimental peak distributions for early external distractors.
Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the first two temporal variants of external distraction, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for external distractor with 0 ms delay. (B) Example replay for external distractor with 0 ms delay. (C) Violin plot for external distractor with 1 ms delay. (D) Example replay for external distractor with 1 ms delay.
Fig 11
Experimental peak distributions for late external distractors.
Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the second two temporal variants of external distraction, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for external distractor with 2 ms delay. (B) Example replay for external distractor with 2 ms delay. (C) Violin plot for external distractor with 3 ms delay. (D) Example replay for external distractor with 3 ms delay.
Experimental peak distributions for early external distractors.
Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the first two temporal variants of external distraction, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for external distractor with 0 ms delay. (B) Example replay for external distractor with 0 ms delay. (C) Violin plot for external distractor with 1 ms delay. (D) Example replay for external distractor with 1 ms delay.
Experimental peak distributions for late external distractors.
Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the second two temporal variants of external distraction, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for external distractor with 2 ms delay. (B) Example replay for external distractor with 2 ms delay. (C) Violin plot for external distractor with 3 ms delay. (D) Example replay for external distractor with 3 ms delay.The external distractor does not have a highly disruptive effect on the replay timing, as predicted. Examining the density violins, it adds some additional variance to the timing of the recall process, but does not fundamentally alter it. This suggests that input outside of or not directly linked to the trained sequence elements have only a mild noising effect, and that the internal representation of the trained sequence is kept, for the most part, local to the input cluster—i.e. it only weakly or rather sparsely extends to untrained neurons (at least in the “input” direction), as expected from the representation examination. We classify this type of distraction as “irrelevant” distraction.
Internal distractors
Similarly, an assortment of equivalent violin plots for each internal, or trained, experimental condition (3 positions and 4 delays), including the percentages of each condition passing the threshold veto (Figs 12, 13, 14, 15, 16 and 17), is presented. Also included is an example raster of distracted replay for each condition. For clarity of viewing, these are separated into temporally early and temporally late distractors. Temporally early is used to mean at the same time as or immediately following the recall cue, and temporally late means later in the recall process.
Fig 12
Experimental peak densities for early distractors at position A. Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the first two temporal variants of distraction at position A, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for distractor A with 0 ms delay. (B) Example replay for distractor A with 0 ms delay. (C) Violin plot for distractor A with 1 ms delay. (D) Example replay for distractor A with 1 ms delay.
Fig 13
Experimental peak densities for late distractors at position A. Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the second two temporal variants of distraction at position A, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for distractor A with 2 ms delay. (B) Example replay for distractor A with 2 ms delay. (C) Violin plot for distractor A with 3 ms delay. (D) Example replay for distractor A with 3 ms delay.
Fig 14
Experimental peak densities for early distractors at position C. Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the first two temporal variants of distraction at position C, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for distractor C with 0 ms delay. (B) Example replay for distractor C with 0 ms delay. (C) Violin plot for distractor C with 1 ms delay. (D) Example replay for distractor C with 1 ms delay.
Fig 15
Experimental peak densities for late distractors at position C. Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the second two temporal variants of distraction at position C, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for distractor C with 2 ms delay. (B) Example replay for distractor C with 2 ms delay. (C) Violin plot for distractor C with 3 ms delay. (D) Example replay for distractor C with 3 ms delay.
Fig 16
Experimental peak densities for early distractors at position E. Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the first two temporal variants of distraction at position E, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for distractor E with 0 ms delay. (B) Example replay for distractor E with 0 ms delay. (C) Violin plot for distractor E with 1 ms delay. (D) Example replay for distractor E with 1 ms delay.
Fig 17
Experimental peak densities for late distractors at position E. Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the second two temporal variants of distraction at position E, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for distractor E with 2 ms delay. (B) Example replay for distractor E with 2 ms delay. (C) Violin plot for distractor E with 3 ms delay. (D) Example replay for distractor E with 3 ms delay.
Experimental peak densities for early distractors at position A. Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the first two temporal variants of distraction at position A, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for distractor A with 0 ms delay. (B) Example replay for distractor A with 0 ms delay. (C) Violin plot for distractor A with 1 ms delay. (D) Example replay for distractor A with 1 ms delay.Experimental peak densities for late distractors at position A. Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the second two temporal variants of distraction at position A, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for distractor A with 2 ms delay. (B) Example replay for distractor A with 2 ms delay. (C) Violin plot for distractor A with 3 ms delay. (D) Example replay for distractor A with 3 ms delay.Experimental peak densities for early distractors at position C. Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the first two temporal variants of distraction at position C, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for distractor C with 0 ms delay. (B) Example replay for distractor C with 0 ms delay. (C) Violin plot for distractor C with 1 ms delay. (D) Example replay for distractor C with 1 ms delay.Experimental peak densities for late distractors at position C. Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the second two temporal variants of distraction at position C, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for distractor C with 2 ms delay. (B) Example replay for distractor C with 2 ms delay. (C) Violin plot for distractor C with 3 ms delay. (D) Example replay for distractor C with 3 ms delay.Experimental peak densities for early distractors at position E. Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the first two temporal variants of distraction at position E, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for distractor E with 0 ms delay. (B) Example replay for distractor E with 0 ms delay. (C) Violin plot for distractor E with 1 ms delay. (D) Example replay for distractor E with 1 ms delay.Experimental peak densities for late distractors at position E. Violin plots (densities estimated with a 50 point Gaussian kernel) of the peak times for the second two temporal variants of distraction at position E, with listed percentages of trials passing veto. A raster plot of an example distracted recall is also included. The spiral indicates the location of the distractor. (A) Violin plot for distractor E with 2 ms delay. (B) Example replay for distractor E with 2 ms delay. (C) Violin plot for distractor E with 3 ms delay. (D) Example replay for distractor E with 3 ms delay.Several important things in these plots can be quickly noted. First of all, it must be stated that a “good” replay is considered to consist of local maxima in the subpopulation spike rate for the stimulus groups occurring in their trained order. Such good (i.e. complete and correctly ordered) replays only occur in cases where the spatio-temporal position of the distractor is at or subsequent to the spatio-temporal position of the corresponding trained stimulus during undistracted replay. This suggests, at least in the rapid replay regime, that once a stimulus has been activated in replay, distractions at its spatial location act like external distractors.Furthermore, as predicted in the representation analysis, the strongest effects appear to apply to the group at the position of the distractor, and the subsequent one (in the form of activation and subsequent refractory / inhibitory shutdown, and premature replay, respectively). We refer to these strong, disruptive effects as arising from “relevant” distractors.Additionally, as predicted, there is no significant “backward” effect from any internal distractor. This lines up with the same sort of “irrelevant” distraction seen with external distractors. We can summarize the properties of “relevant” and “irrelevant” distractors as follows. Trained distractors presented early tend to disrupt replay (i.e. “relevant”), and untrained distractors and trained distractors presented late tend to only add noise to it (i.e. “irrelevant”). We will continue to discuss in detail the reasons we believe this occurs (beyond the representation / storage analysis presented earlier) in the discussion section, but, though it was not subsequently given much thought or studied systematically, certain elements of the suspected causes were observed as early as 1993 [10].
Evaluation measure results
The quantitative measures we had previously developed are now deployed. We hope to determine if either of these measures, or some combination of the two, can be used in our distractor classification scheme. Arrays of disruption and deviance index values for the various experimental conditions are displayed in Fig 18. Effectively, negative values indicate events occuring “early,” while positive values indicate events occuring “late” (in the case of the disruption index, events are ordered pairs of stimuli, while in the case of the deviance index, events are individual stimuli).
Fig 18
Disruption and deviance index arrays.
An array of disruption and deviance index values for each experimental condition.
Disruption and deviance index arrays.
An array of disruption and deviance index values for each experimental condition.Examining this leads to the observation that the external distractor leads to a minimal disruption and deviance compared to the other distractors. We note as well that any distractor (with the exception of the distractor at position A, i.e. the starting distractor) presented near the end of the replay also has a relatively low disruption and deviance. Conversely, we see that the strongest effects come from the distractors at the middle (C) or end (E) positions, and that said effects are strongest when the distractors are presented at the same time as the recall cue. This further confirms our initial evaluations and subsequent predictions regarding what kind of distractors are “relevant” and “irrelevant.”It is difficult to arrive at a direct determination of relevancy or irrelevancy using these quantitative measures, though, certainly, there are strong tendencies. For our value sets, thresholding the disruption index at -0.05 results in all lesser measures being relevant and all greater measures being irrelevant, with the exception of distractor E with a 3 ms delay, which is, under this scheme, classified as irrelevant despite being mildly relevant. Similar mostly-reliable rules can probably be made for other conditions and parameter sets, but are not expected to be universal.An important property to note here is that the system (i.e. the recall of the learned representation) is, for a large class of distractors, extremely stable and robust against distractions, even in the absence of any top-down attention or feedback mechanism. This stability is intrinsic to the learned connectivity and the neuronal dynamics.
Discussion
While certain preliminary analyses in the context of synfire chains (the most extreme of the “rapid replay” spectrum) were made quite early on [10], on the whole, interference with the recall process in trained recurrent network has not been well studied. We attempted, with this paper, to provide a beginning basis and a set of simple measures and terms for studying these things, as they will become more and more important as self-training recurrent neural networks are deployed into various real-world and real-time applications, many of which will likely be in noisy environments where distracting input is a real possibility.Distractors have many potential sources, both internal and external. External distractors are primarily sensory stimuli—a heard noise or seen movement, for example. Distractors may be internal as well—a sudden, unexpected memory or thought popping into one’s head. They may even be hybrid in nature—a feeling of discomfort, pain, or unease, for example. Regardless of the source, the brain must deal with distractors on a regular basis, and it’s important to understand how neural circuits deal with them on a basic level.In examining this, we first will discuss the case of the so-called irrelevant distractor, as it is relatively straightforward to consider its phenomenology—simply put, by introducing additional activity into the network that is not correlated in a meaningful way with the recall process, the distractor and its reverberations add noise to the network activity, which can affect timing and readout precision, but do not significantly perturb the dynamics or trajectory of the system. This is nearly trivial to understand.What requires somewhat more thought and consideration is the case of the so-called relevant distractor—what network phenomena make it so disruptive to the replay process? When a group of similarly-tuned trained neurons is simultaneously activated, several things happen. First of all, a brief refractory period exists for the fired neurons, preventing immediate reactivation. On slightly longer timescales, several mechanisms conspire to continue to discourage rapid reactivation of the neuron group in question, including inhibitory activation and the raising of the neurons’ firing threshold via intrinsic plasticity. At the same time, after sufficient conduction and integration time, if the activation has been strong enough, the next group of neurons in the trained sequence activates as activity propagates through the network. When untrained neurons are activated as a distractor, this does not tend to activate (via secondary pathways) a trained group with sufficient coherency for these things to occur. Similarly, when a trained group of neurons is stimulated when it is already in the described post-activation state (i.e. late and out-of-sequence), its recurrent self-excitation and tendency to propagate activity is highly suppressed. On the other hand, if a group is stimulated early and out-of-sequence, it tends to induce all the described effects, disrupting any sequence replay that was already occurring or about to occur.Sequences are undoubtedly ubiquitous and important. That being said, most executed actions occur on a slower timescale than the replays produced by this model, a notable exception being saccades. Distraction studies have even been done on saccades, however, while comparing the results of such studies to the model in this paper is tempting, general belief in saccades being executed by the combined action of two separate circuits (specifically, a “when” network and a “where” network) make comparisons with our single network model difficult if not impossible [22-24].A partial in-silico replication of a biological experiment examining low-level sequence replay [3, 9] using a very closely related model to the one used in this paper examined the concept of distraction from a different perspective, i.e. by spatially moving the start cue within the sequence, rather than keeping the start cue constant and introducing an additional spurious signal as we have. Similar to our findings, a feedforward, rapid-replay dominant regime of pattern storage and recall is reported, with a recall speed independent of training speed. Instead of looking at qualitative evaluations of cue and distraction effects, these papers utilize a cumulative distribution of Spearman correlation coefficients (a rank-ordered derivative of Pearson correlation coefficients) between cluster rates. A rightward shift of this distribution indicates improved recall performance with reagrd to cluster ordering. We may then speculate what the result of such analyses on our current distractor paradigm would be. Overall, we expect that our irrelevant distractors would result in a similar rightward shift to the undistracted position, while relevant distractors, by strongly disrupting the ordering between cluster peaks, would result in an insignificant shift or possibly even a leftward shift. We would expect the results regarding presentation of the start cue at alternative locations to produce the same results as the previous in-silico experiment. However, we have not adopted this analysis method in this paper, as it provides a less specific exploration of the phenomenology of what the distractor actually does to the replay dynamics.We believe our observations and predictions hold in general for the rapid replay regime of trained sequences, though it remains to be seen if it holds as well for a slower replay regime involving extended switching between self-sustained activity in cell assemblies or moving attractors. Though it falls outside the scope of this study, we can still speculate and provide some background. The rapid replay (or synfire-dominated) regime, as we call it, is the most natural and obvious way to recall sequences of neural activations. However, as stated, the recall speed is invariant to the training speed, depending only on the size of the sequence and neuronal conduction and integration delays—e.g. a three element sequence trained over a three second activation would replay faster than an five element sequence trained over a half-second activation. The fact that not only is timescale not captured during training, but is also rapid and invariant during recall means that this regime cannot directly drive, for example, motor activation, but exists most likely as a one in a hierarchy of components of memory and planning. Various mechanisms of slow replay have been proposed (e.g. [8, 25–27]), but they rely on modulatory signals or differing neural architecture or dynamics. It is as such a much more difficult question to ask how distraction affects slow replay, as there are numerous ways, both hierarchical and non-hierarchical, in which it may be implemented, and not an exceedingly obvious way as in the case of rapid replay. We can, however, recognize certain strong effects that are a part of distracted rapid replay that would, in all likelihood, not be a factor in slow distracted replay. First of all, the refractory effect mentioned would not be a concern. The inhibitory aftereffect mentioned might also be lessened, depending on the timescales and architectures involved. This suggests that the primary point of distraction would be the premature activation of a sequence element linked to the distractor. Overall, in the slower regime, a more robust and less disruptive response would be expected in most cases. Day-to-day observation of human and animal behavior confirms this suspicion; if recall across all timescales were as sensitive to distraction as it is in the rapid replay regime, living in a busy, noisy world would be much more difficult indeed.
Conclusion
We have studied the effects of distraction on cued recall in trained recurrent neural networks. Specifically, we have first presented a recurrent neural network that learns simple sequences in an unsupervised fashion, and, once trained, is capable of rapidly recalling them upon the receipt of a recall cue. We have then examined this rapid recall process under both distracted and undistracted conditions. We note two general families of effects from distractors—either the addition of noise to the recall process, or the disruption or disordering of the recall process. We refer to these two categories as irrelevant and relevant distractors, respectively. By examining the conditions in which each type of distraction occurs, we arrive at the conclusion that in the rapid recall regime, relevant, or highly disruptive distractors, tend to be those which correspond to trained stimuli, and are presented early relative to the “natural” replay time in the recall process of the stimulus to which they correspond. Similarly, irrelevant (or minimally disruptive) distractors, tend to be those which correspond to either untrained stimuli, or trained stimuli presented later than their “natural” replay time. Due to the dynamics of recurrent neural networks in the cortex and hippocampus, we believe the principles of interference or distraction outline here should apply, in general, to any neural phenomena which exist primarily in a non-hierarchical rapid replay regime.27 Nov 2019PONE-D-19-26936Hey, look over there: Distraction effects on rapid sequence recallPLOS ONEDear Dr. Miner,Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.Why did the author choose SORN model? What is its advantage compared with the Hodgkin-Huxley model?During the training session, why did the authors choose 100 s? If the training duration is much shorter or longer than 100 s, whether the results will be changed?Why did the authors disable synaptic plasticity after training?Please, compare your model results with experimental data. A figure showing experimental data and comparison of model dynamics would therefore strengthen this paper.Please, discuss where relevant and irrelevant distractors come from. For example, both external environmental influences as well as previously learned internal dynamics from within the brain could create these two scenarios. A new simulation to test how quick/easy it is to train a network on a new related and unrelated sequence would be an interesting addition to the paper.There are probably too many figures in the paper. Some could be combined into multi-panel figures, for example figures 16 and 17 could be in the same figure as separate panels. Similarly for other figure pairs.Please, make all source code available including scripts for specific experiments run so that others can replicate the results.The English needs to be improved. The authors should double check typo, grammar, etc.In the abstract, the authors should check the usage of “arrive at”.Line 52, “top-town”=>”top-down”; Line 56, “rchitecture” => “architecture”In the Figure 1 legend, “The excitatory portion of the of the network” => “The excitatory portion of the network”Line 145, “an precisely” =>”a precisely”Line 218, “we histogram”, histogram is a noun, not a verb.Line 220: "contro-only" -> "control-only"Line 331: "Weope" ?? Please check spelling and grammar.Line 260, “connections strengths”=> “connection strengths”Line 291, “desnity” =>“density”Line 331, what is “Weope”?Line 352, it has two “for”, delete one of the “for”;Line 355, “lrge” =>”large”Line 415, “holds”=>”hold”jLine 450, what is “said process”?It’s better to add left and right In Fig 4 legend, such as “unsorted (left) and sorted (right)”.In Figure 9, the label of panel (A) and (B) are missing. The label of x axis in the panel B is missing.The label of y axis in the panel B of Fig10-17.Fig 18, the label of x axis of panel A is missing. In the panel B, based on the value of 0.0, 0.001, 0.002, and 0.003, the unit should be s, not ms.We would appreciate receiving your revised manuscript by Jan 11 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter.To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocolsPlease include the following items when submitting your revised manuscript:A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'.A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'.An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'.Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out.We look forward to receiving your revised manuscript.Kind regards,Gennady Cymbalyuk, Ph.D.Academic EditorPLOS ONEJournal Requirements:1. When submitting your revision, we need you to address these additional requirements.Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found athttp://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf2. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide.[Note: HTML markup is below. Please do not edit.]Reviewers' comments:Reviewer's Responses to QuestionsComments to the Author1. Is the manuscript technically sound, and do the data support the conclusions?The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.Reviewer #1: YesReviewer #2: Yes**********2. Has the statistical analysis been performed appropriately and rigorously?Reviewer #1: YesReviewer #2: N/A**********3. Have the authors made all data underlying the findings in their manuscript fully available?The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.Reviewer #1: NoReviewer #2: No**********4. Is the manuscript presented in an intelligible fashion and written in standard English?PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.Reviewer #1: YesReviewer #2: Yes**********5. Review Comments to the AuthorPlease use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)Reviewer #1: Very clearly written paper on using a trained recurrent neural network paradigmto teach the network to produce sequential activation of neural ensembles. The authorsthen perform tests of distracting the network using random noise (irrelevantdistractor) and by specifically manipulating the temporal response of the network.The work is compelling and interesting and should shed light on neural mechanismsof working memory and how it is disturbed from environmental or internal states.Although I have no methodological concerns, the authors should make all source codeavailable including scripts for specific experiments run so that others can replicatethe results. These should be posted to github.com in a public repository or on ModelDB.The authors could do a better job of comparing their model results with experimentaldata. This would enhance the likelihood that biologists will take their work seriously.A figure showing experimental data and comparison of model dynamics would thereforestrengthen this paper.In addition, the authors could discuss where relevant and irrelevant distractors comefrom. For example, both external environmental influences as well as previously learnedinternal dynamics from within the brain could create these two scenarios. A new simulationto test how quick/easy it is to train a network on a new related and unrelated sequencewould be an interesting addition to the paper.There are probably too many figures in the paper. Some could be combined intomulti-panel figures, for example figures 16 and 17 could be in the same figure asseparate panels. Similarly for other figure pairs.line 220: "contro-only" -> "control-only"Line 331: "Weope" ?? Please check spelling and grammar.Reviewer #2: In this paper, the authors studied the influence of two different types of distractors on the network dynamics during the recall of the encoded sequence by using a recurrent network model. The network learns simple sequences with the synaptic weights modulated by two plasticity mechanisms: 1) spike timing-dependent plasticity (STDP), and 2) homeostatic synaptic plasticity. They examined the recall with different types of distractors, and concluded that they can broadly classify them as relevant and irrelevant distractors.Overall, the article is well organized and presented. However, the following revisions should be considered:Major issues:1. Why did the author choose SORN model? What is its advantage compared with the Hodgkin-Huxley model?2. During the training session, why did the authors choose 100 s? If the training duration is much shorter or longer than 100 s, whether the results will be changed?3. Why did the authors disable synaptic plasticity after training?Minor issues:1. The English needs to be improved. The authors should double check typo, grammar, etc.1) In the abstract, the authors should check the usage of “arrive at”.2) Line 52, “top-town”=>”top-down”; Line 56, “rchitecture” => “architecture”3) In the Figure 1 legend, “The excitatory portion of the of the network” => “The excitatory portion of the network”4) Line 145, “an precisely” =>”a precisely”5) Line 218, “we histogram”, histogram is a noun, not a verb.6) Line 260, “connections strengths”=> “connection strengths”7) Line 291, “desnity” =>“density”8) Line 331, what is “Weope”?9) Line 352, it has two “for”, delete one of the “for”;10) Line 355, “lrge” =>”large”11) Line 415, “holds”=>”hold”j2. Line 450, what is “said process”?3. It’s better to add left and right In Fig 4 legend, such as “unsorted (left) and sorted (right)”.4. In Figure 9, the label of panel (A) and (B) are missing. The label of x axis in the panel B is missing.5. The label of y axis in the panel B of Fig10-17.6. Fig 18, the label of x axis of panel A is missing. In the panel B, based on the value of 0.0, 0.001, 0.002, and 0.003, the unit should be s, not ms.**********6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.If you choose “no”, your identity will remain anonymous but your review may still be made public.Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.Reviewer #1: NoReviewer #2: No[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.]While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.20 Feb 2020Dear Editor and Reviewers,We thank you for your work in reviewing this manuscript, and apologize for the delay in revisions. We hope that we have sufficiently revised the paper in order to address the reviewers’ concerns. We reproduce below the reviewer comments (in italics), with responses (in normal text) where appropriate.Reviewer #1:Comment #1: Very clearly written paper on using a trained recurrent neural network paradigm to teach the network to produce sequential activation of neural ensembles. The authors then perform tests of distracting the network using random noise (irrelevant distractor) and by specifically manipulating the temporal response of the network. The work is compelling and interesting and should shed light on neural mechanisms of working memory and how it is disturbed from environmental or internal states.Although I have no methodological concerns, the authors should make all source code available including scripts for specific experiments run so that others can replicate the results. These should be posted to github.com in a public repository or on ModelDB.Response: We have made the code available on Github ( https://github.com/DCMiner/HeyLookOverThere )Comment #2: The authors could do a better job of comparing their model results with experimental data. This would enhance the likelihood that biologists will take their work seriously.A figure showing experimental data and comparison of model dynamics would therefore strengthen this paper.Response: We agree with the reviewer, but unfortunately we were unable to find appropriate experimental data. All experiments we could find featuring a distracted recall paradigm were of too long a timescale and at too high an abstraction level to be realistically comparable. The one hopeful exception, we investigated in more detail, were microsaccade studies. However, the neural architecture behind saccade planning and execution is believed to be rather different from our model’s architecture, and thus is realistically incomparable as well.Comment #3: In addition, the authors could discuss where relevant and irrelevant distractors come from. For example, both external environmental influences as well as previously learned internal dynamics from within the brain could create these two scenarios. A new simulation to test how quick/easy it is to train a network on a new related and unrelated sequence would be an interesting addition to the paper.Response: Thank you for the comment. We have added this point to the discussion section. However, while we understand the interest, we think that the training of new, additional sequences falls outside of the intended scope of this manuscript.Comment #4: There are probably too many figures in the paper. Some could be combined into multi-panel figures, for example figures 16 and 17 could be in the same figure as separate panels. Similarly for other figure pairs.Response: While we do agree that the manuscript has a large number of figures – in particular, figures 10 – 17, as you noted – we found that reducing their size to combine them led to a lack of figure readability in the plots, particularly with regard to the shape of the violin plot distributions and the exact spatio-temporal distractor position. We believe that this readability is important to an understanding of the results and have therefore somewhat reluctantly decided to leave these figures as they are.Comment #5: line 220: "contro-only" -> "control-only"Response: Corrected.Comment #6: Line 331: "Weope" ?? Please check spelling and grammar.Response: Corrected.Reviewer #2:Comment #1: In this paper, the authors studied the influence of two different types of distractors on the network dynamics during the recall of the encoded sequence by using a recurrent network model. The network learns simple sequences with the synaptic weights modulated by two plasticity mechanisms: 1) spike timing-dependent plasticity (STDP), and 2) homeostatic synaptic plasticity. They examined the recall with different types of distractors, and concluded that they can broadly classify them as relevant and irrelevant distractors.Overall, the article is well organized and presented. However, the following revisions should be considered:Comment #1: Why did the author choose SORN model? What is its advantage compared with the Hodgkin-Huxley model?Response: SORN is a model family of network architectures and combinations of diverse plasticity mechanisms considering inter alias the integrate-and-fire model to describe the neuronal dynamics. The Hodgkin-Huxley model is a detailed single-neuron model encompassing the dynamics of ion channels on a neuron’s membrane. Although more detailed, considering the Hodgkin-Huxley model in large, adaptive neuronal networks would require too much computational resources. Furthermore, the phenomena examined in this study are on the network level and, thus, we expect that the detailed dynamics of individual ion channels do not have a significant influence on these.Comment #2: During the training session, why did the authors choose 100 s? If the training duration is much shorter or longer than 100 s, whether the results will be changed?Response: The network requires several tens of seconds to learn, after which it reaches a saturated learning state indicated by the fact that the synaptic weights do not change significantly any more. 100 s was found to be sufficient to ensure that this saturation was stable without requiring an excessive amount of simulation time. Due to the saturation, a longer training duration would not significantly change the synaptic weights and, thus, we would obtain the same results as shown in the manuscript. Significantly shorter training (10 s or 20 s training time, for example) would result in reduced learning, which would influence already the unperturbed recall. However, the systematic analysis of reduced training time is beyond the scope of the manuscript.Comment #3: Why did the authors disable synaptic plasticity after training?Response: We perform here a separation of time scales. If synaptic plasticity were left on after training, the network would consider the cue signal as learning signal and overlearn it, as it is repeated alone and strongly many times during the testing phase. This is not intended and, thus, we consider that recall happens on a much faster time scale than learning.Minor issues:1. The English needs to be improved. The authors should double check typo, grammar, etc.#1) In the abstract, the authors should check the usage of “arrive at”.-Usage is correct.#2) Line 52, “top-town”=>”top-down”; Line 56, “rchitecture” => “architecture”-Fixed#3) In the Figure 1 legend, “The excitatory portion of the of the network” => “The excitatory portion of the network”-Fixed#4) Line 145, “an precisely” =>”a precisely”-Fixed#5) Line 218, “we histogram”, histogram is a noun, not a verb.-“Histogram” can be used as a verb, as is evidenced by the common usage of “histogramming,” for example. Wiktionary supports this assertion.#6) Line 260, “connections strengths”=> “connection strengths”-Fixed#7) Line 291, “desnity” =>“density”-Fixed#8) Line 331, what is “Weope”?-Fixed (typo of “we hope”)#9) Line 352, it has two “for”, delete one of the “for”;-Fixed#10) Line 355, “lrge” =>”large”-Fixed#11) Line 415, “holds”=>”hold”j-Fixed2. Line 450, what is “said process”?-This usage of said means recently referred to – in this case, the recall process. We have edited the text for clarity.3. It’s better to add left and right In Fig 4 legend, such as “unsorted (left) and sorted (right)”.-Fixed4. In Figure 9, the label of panel (A) and (B) are missing. The label of x axis in the panel B is missing.-This is intentional and intended to reduce clutter, as the labels are given in the caption. This and similar choices were discussed with multiple colleagues and agreed upon for many of the “busier” figures, as many found the initial figures with all duplicate labels present to be overly cluttered and difficult to read.5. The label of y axis in the panel B of Fig10-17.-see Point 4 above6. Fig 18, the label of x axis of panel A is missing. In the panel B, based on the value of 0.0, 0.001, 0.002, and 0.003, the unit should be s, not ms.-see Point 4 abovePlease find attached the relevant revision files.Sincerely,Daniel Miner and Christian TetzlaffSubmitted filename: Response Letter_corr.pdfClick here for additional data file.12 Mar 2020Hey, look over there: Distraction effects on rapid sequence recallPONE-D-19-26936R1Dear Dr. Miner,We are pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it complies with all outstanding technical requirements.The figure resolution is pretty low in the paper. Please make sure it has high resolution when it is published.Within one week, you will receive an e-mail containing information on the amendments required prior to publication. When all required modifications have been addressed, you will receive a formal acceptance letter and your manuscript will proceed to our production department and be scheduled for publication.Shortly after the formal acceptance letter is sent, an invoice for payment will follow. To ensure an efficient production and billing process, please log into Editorial Manager at https://www.editorialmanager.com/pone/, click the "Update My Information" link at the top of the page, and update your user information. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.If your institution or institutions have a press office, please notify them about your upcoming paper to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, you must inform our press team as soon as possible and no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.With kind regards,Gennady Cymbalyuk, Ph.D.Academic EditorPLOS ONEAdditional Editor Comments (optional):Reviewers' comments:Reviewer's Responses to QuestionsComments to the Author1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.Reviewer #1: All comments have been addressedReviewer #2: All comments have been addressed**********2. Is the manuscript technically sound, and do the data support the conclusions?The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.Reviewer #1: YesReviewer #2: Yes**********3. Has the statistical analysis been performed appropriately and rigorously?Reviewer #1: YesReviewer #2: Yes**********4. Have the authors made all data underlying the findings in their manuscript fully available?The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.Reviewer #1: YesReviewer #2: Yes**********5. Is the manuscript presented in an intelligible fashion and written in standard English?PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.Reviewer #1: YesReviewer #2: Yes**********6. Review Comments to the AuthorPlease use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)Reviewer #1: (No Response)Reviewer #2: The authors significantly improved the manuscript and addressed all my original questions. I have one minor comment: the figure resolution is pretty low in the paper. Please make sure it has high resolution when it is published.**********7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.If you choose “no”, your identity will remain anonymous but your review may still be made public.Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.Reviewer #1: NoReviewer #2: No24 Mar 2020PONE-D-19-26936R1Hey, look over there: Distraction effects on rapid sequence recallDear Dr. Miner:I am pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.For any other questions or concerns, please email plosone@plos.org.Thank you for submitting your work to PLOS ONE.With kind regards,PLOS ONE Editorial Office Staffon behalf ofDr. Gennady CymbalyukAcademic EditorPLOS ONE