Abstract

We discuss the concept of probabilistic neural networks with a fixed internal representation being models for machine understanding. Here, ‘understanding’ is interpretted as the ability to map data to an already existing representation which encodes an a priori organisation of the feature space. We derive the internal representation by requiring that it satisfies the principles of maximal relevance and of maximal ignorance about how different features are combined. We show that, when hidden units are binary variables, these two principles identify a unique model—the hierarchical feature model—which is fully solvable and provides a natural interpretation in terms of features. We argue that learning machines with this architecture possess a number of interesting properties, such as the continuity of the representation with respect to changes in parameters and data, the possibility of controlling the level of compression and the ability to support functions that go beyond generalisation. We explore the behaviour of the model with extensive numerical experiments and argue that models in which the internal representation is fixed reproduce a learning modality which is qualitatively different from that of traditional models, such as restricted Boltzmann machines.

Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.