Abstract
In analyzing a complex network in the real world, it is ideally of great help to recognize its universality class. While biological networks, in particular, grow under various ‘learning rules’, their impacts on scaling have not yet been characterized enough. Here we applied the Hodge-Kodaira decomposition, a topological method to count global loops, to neural networks with different learning rules and edge densities. Interestingly, the networks which evolved under different learning rules showed different scalings with edge densities. The causal learning rule scaled similarly to its underlying graph (i.e. Erdös-Renyi random graph, in this study), on which a network can grow, while the Hebbian-like rule did not.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.