Do you want to publish a course? Click here

Effects of degree distribution in mutual synchronization of neural networks

45   0   0.0 ( 0 )
 Added by Sheng-Jun Wang
 Publication date 2007
  fields Physics
and research's language is English




Ask ChatGPT about the research

We study the effects of the degree distribution in mutual synchronization of two-layer neural networks. We carry out three coupling strategies: large-large coupling, random coupling, and small-small coupling. By computer simulations and analytical methods, we find that couplings between nodes with large degree play an important role in the synchronization. For large-large coupling, less couplings are needed for inducing synchronization for both random and scale-free networks. For random coupling, cutting couplings between nodes with large degree is very efficient for preventing neural systems from synchronization, especially when subnetworks are scale-free.



rate research

Read More

Characterizing the in uence of network properties on the global emerging behavior of interacting elements constitutes a central question in many areas, from physical to social sciences. In this article we study a primary model of disordered neuronal networks with excitatory-inhibitory structure and balance constraints. We show how the interplay between structure and disorder in the connectivity leads to a universal transition from trivial to synchronized stationary or periodic states. This transition cannot be explained only through the analysis of the spectral density of the connectivity matrix. We provide a low dimensional approximation that shows the role of both the structure and disorder in the dynamics.
112 - Alex Arenas 2006
We investigate the connection between the dynamics of synchronization and the modularity on complex networks. Simulating the Kuramotos model in complex networks we determine patterns of meta-stability and calculate the modularity of the partition these patterns provide. The results indicate that the more stable the patterns are, the larger tends to be the modularity of the partition defined by them. This correlation works pretty well in homogeneous networks (all nodes have similar connectivity) but fails when networks contain hubs, mainly because the modularity is never improved where isolated nodes appear, whereas in the synchronization process the characteristic of hubs is to have a large stability when forming its own community.
We investigate the dynamical role of inhibitory and highly connected nodes (hub) in synchronization and input processing of leaky-integrate-and-fire neural networks with short term synaptic plasticity. We take advantage of a heterogeneous mean-field approximation to encode the role of network structure and we tune the fraction of inhibitory neurons $f_I$ and their connectivity level to investigate the cooperation between hub features and inhibition. We show that, depending on $f_I$, highly connected inhibitory nodes strongly drive the synchronization properties of the overall network through dynamical transitions from synchronous to asynchronous regimes. Furthermore, a metastable regime with long memory of external inputs emerges for a specific fraction of hub inhibitory neurons, underlining the role of inhibition and connectivity also for input processing in neural networks.
The response of degree-correlated scale-free attractor networks to stimuli is studied. We show that degree-correlated scale-free networks are robust to random stimuli as well as the uncorrelated scale-free networks, while assortative (disassortative) scale-free networks are more (less) sensitive to directed stimuli than uncorrelated networks. We find that the degree-correlation of scale-free networks makes the dynamics of attractor systems different from uncorrelated ones. The dynamics of correlated scale-free attractor networks result in the effects of degree correlation on the response to stimuli.
In this work, we address the question whether a sufficiently deep quantum neural network can approximate a target function as accurate as possible. We start with simple but typical physical situations that the target functions are physical observables, and then we extend our discussion to situations that the learning targets are not directly physical observables, but can be expressed as physical observables in an enlarged Hilbert space with multiple replicas, such as the Loshimidt echo and the Renyi entropy. The main finding is that an accurate approximation is possible only when the input wave functions in the dataset do not exhaust the entire Hilbert space that the quantum circuit acts on, and more precisely, the Hilbert space dimension of the former has to be less than half of the Hilbert space dimension of the latter. In some cases, this requirement can be satisfied automatically because of the intrinsic properties of the dataset, for instance, when the input wave function has to be symmetric between different replicas. And if this requirement cannot be satisfied by the dataset, we show that the expressivity capabilities can be restored by adding one ancillary qubit where the wave function is always fixed at input. Our studies point toward establishing a quantum neural network analogy of the universal approximation theorem that lays the foundation for expressivity of classical neural networks.
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا