Do you want to publish a course? Click here

Self-control dynamics for sparsely coded networks with synaptic noise

66   0   0.0 ( 0 )
 Added by Desire Bolle
 Publication date 2004
  fields Physics
and research's language is English




Ask ChatGPT about the research

For the retrieval dynamics of sparsely coded attractor associative memory models with synaptic noise the inclusion of a macroscopic time-dependent threshold is studied. It is shown that if the threshold is chosen appropriately as a function of the cross-talk noise and of the activity of the memorized patterns, adapting itself automatically in the course of the time evolution, an autonomous functioning of the model is guaranteed. This self-control mechanism considerably improves the quality of the fixed-point retrieval dynamics, in particular the storage capacity, the basins of attraction and the mutual information content.

rate research

Read More

97 - D. Bolle , R. Heylen 2007
The inclusion of a macroscopic adaptive threshold is studied for the retrieval dynamics of both layered feedforward and fully connected neural network models with synaptic noise. These two types of architectures require a different method to be solved numerically. In both cases it is shown that, if the threshold is chosen appropriately as a function of the cross-talk noise and of the activity of the stored patterns, adapting itself automatically in the course of the recall process, an autonomous functioning of the network is guaranteed. This self-control mechanism considerably improves the quality of retrieval, in particular the storage capacity, the basins of attraction and the mutual information content.
63 - D. Bolle , R. Heylen 2006
The inclusion of a macroscopic adaptive threshold is studied for the retrieval dynamics of layered feedforward neural network models with synaptic noise. It is shown that if the threshold is chosen appropriately as a function of the cross-talk noise and of the activity of the stored patterns, adapting itself automatically in the course of the recall process, an autonomous functioning of the network is guaranteed.This self-control mechanism considerably improves the quality of retrieval, in particular the storage capacity, the basins of attraction and the mutual information content.
The dynamics of neural networks is often characterized by collective behavior and quasi-synchronous events, where a large fraction of neurons fire in short time intervals, separated by uncorrelated firing activity. These global temporal signals are crucial for brain functioning. They strongly depend on the topology of the network and on the fluctuations of the connectivity. We propose a heterogeneous mean--field approach to neural dynamics on random networks, that explicitly preserves the disorder in the topology at growing network sizes, and leads to a set of self-consistent equations. Within this approach, we provide an effective description of microscopic and large scale temporal signals in a leaky integrate-and-fire model with short term plasticity, where quasi-synchronous events arise. Our equations provide a clear analytical picture of the dynamics, evidencing the contributions of both periodic (locked) and aperiodic (unlocked) neurons to the measurable average signal. In particular, we formulate and solve a global inverse problem of reconstructing the in-degree distribution from the knowledge of the average activity field. Our method is very general and applies to a large class of dynamical models on dense random networks.
To identify emerging microscopic structures in low temperature spin glasses, we study self-sustained clusters (SSC) in spin models defined on sparse random graphs. A message-passing algorithm is developed to determine the probability of individual spins to belong to SSC. Results for specific instances, which compare the predicted SSC associations with the dynamical properties of spins obtained from numerical simulations, show that SSC association identifies individual slow-evolving spins. This insight gives rise to a powerful approach for predicting individual spin dynamics from a single snapshot of an equilibrium spin configuration, namely from limited static information, which can be used to devise generic prediction tools applicable to a wide range of areas.
In this chapter we discuss how the results developed within the theory of fractals and Self-Organized Criticality (SOC) can be fruitfully exploited as ingredients of adaptive network models. In order to maintain the presentation self-contained, we first review the basic ideas behind fractal theory and SOC. We then briefly review some results in the field of complex networks, and some of the models that have been proposed. Finally, we present a self-organized model recently proposed by Garlaschelli et al. [Nat. Phys. 3, 813 (2007)] that couples the fitness network model defined by Caldarelli et al. [Phys. Rev. Lett. 89, 258702 (2002)] with the evolution model proposed by Bak and Sneppen [Phys. Rev. Lett. 71, 4083 (1993)] as a prototype of SOC. Remarkably, we show that the results obtained for the two models separately change dramatically when they are coupled together. This indicates that self-organized networks may represent an entirely novel class of complex systems, whose properties cannot be straightforwardly understood in terms of what we have learnt so far.
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا