Abstract
The retrieval properties of attractor neural networks are studied, subject to a modification of the representation of the stimulus input. In the present study the stimulus imposes an initial state on the network and then remains as a persistent potential input (local field) of weakened amplitude, but containing the same number of errors as the initial, strong input. The dynamics of this network is analysed assuming three different types of error distributions in the persistent stimulus, i.e. Gaussian, discrete and hidden units. It is found that, below saturation of the free network, the persistent stimulus does not damage the retrieval properties of the free network up to rather high values of the field amplitude. Even above saturation, for all loading levels, the network corrects errors if the field amplitude is not too low. The system is studied by solving mean-field equations for its attractors. The solutions are then compared with simulations.
The results lead to the conclusion that the memory loading level at which the network stops acting as an effective associative memory is determined, biologically, by the size of a plausible stimulus input and by the level of errors that can be biologically tolerated.
In order to extend the study to a potential sequence of error-correcting oversaturated networks we introduce random synaptic dilution. It is intended to prevent much correlation between the error bits in the dynamics of the two coupled networks. This compound system is studied by simulations. Finally, we discuss a potential theoretical project in which the dynamics of the coupled networks can be studied analytically.