Structured information in sparse-code metric neural networks
Abstract:
Sparse-code networks have retrieval abilities which are strongly dependent on the firing threshold for the neurons. If the connections are spatially uniform, the macroscopic properties of the network can be measured by the overlap between neurons and learned patterns, and by the global activity. However, for nonuniform networks, for instance small-world networks, the neurons can retrieve fragments of patterns without performing global retrieval. Local overlaps are needed to describe the network. We characterize the structure type of the neural states using a parameter that is related to fluctuations of the local overlaps, with distinction between bump and block phases. Simulation of neural dynamics shows a competition between localized (bump), structured (block) and global retrieval. When the network topology randomness increases, the phase-diagram shows a transition from local to global retrieval. Furthermore, the local phase splits into a bump phase for low activity and a block phase for high activity. A theoretical approach solves the asymptotic limit of the model, and confirms the simulation results which pbkp_redicts the change of stability from bumps to blocks when the storage ratio increases. © 2011 Elsevier B.V. All rights reserved.
Año de publicación:
2012
Keywords:
- network topology
- Associative memory
- Threshold dynamics
- Structured information
- Small world
Fuente:
Tipo de documento:
Article
Estado:
Acceso restringido
Áreas de conocimiento:
- Aprendizaje automático
- Ciencias de la computación
- Ciencias de la computación
Áreas temáticas:
- Programación informática, programas, datos, seguridad
- Métodos informáticos especiales
- Funcionamiento de bibliotecas y archivos