A collaboration between SISSA and the University of Pennsylvania indicates the existence of an efficient process of sensory coding in rats, suggesting a general principle for optimal use of computational resources. The study, published in eLife, paves the way for an understanding of the underlying neuronal mechanisms behind this efficiency and for the development of artificial intelligence systems based on similar principles.
The brains of human beings at rest are supposed to use a fifth of the energy produced by their bodies, i.e., about 20 W of power. In fact, the cost would be much more if our brains were not equipped with an efficient coding mechanism that allows us to represent only the information which is really useful in the vast, ongoing stream of sensory stimuli. A new study conducted by SISSA and the University of Pennsylvania, published in eLife, shows the existence of similar efficient coding processes for visual stimuli in rodents. The results add support to the important efficient coding theory of sensory perception, and pave the way towards new experimental approaches for understanding underlying neuronal mechanisms and the development of new training protocols for artificial vision systems.
In the early ’60s, British scientist Horace Barlow proposed the efficient coding hypothesis. “According to this theory our brain can efficiently represent sensory information using a neural code that minimizes the number of impulses, and thus the energy, necessary to encode and transmit the information,” explains Davide Zoccolan, director of the Visual Neuroscience Laboratory at the Scuola Internazionale Superiore di Studi Avanzati—SISSA. “This happens in particular in the visual system, because of the decreasing number of neurons in the deeper areas of the cortex that reduces the representational capacity.”
According to information theory, which underpins the efficient coding hypothesis, an efficient sensory system should preferentially allocate computational resources to represent those statistical features of the environment that are more informative about its state. In the case of the visual system, this means encoding the most informative features of the natural images that surround us.
Vijay Balasubramanian, a computational neuroscientist at the University of Pennsylvania, has been working on this topic over the past decade: “We analyzed thousands of images of natural landscapes by transforming them into binary images, made up of black and white pixels, and decomposing them into different textures defined by specific statistics,” says the researcher. “We noticed that different kinds of textures have different variability in nature and human subjects are better at recognizing those which vary the most. It is as if our brains assign resources where they are most necessary.”
Until now there was no proof that similarly efficient perceptions of visual textures occur in other species. In a new study, published in eLife, the teams of Zoccolan and Balasubramanian have established that the effect occurs in rodents.
Riccardo Caramellino, first author of the study, together with Andrea Buccellato and Anna Carboncino, trained the animals to discriminate binary images made of random black and white pixels from textures created according to specific probabilistic criteria, as was previously done with human subjects. They then analyzed the results using a mathematical model of an ‘ideal observer’ developed by Eugenio Piasini, co-first author of the paper. The scientists observed that rodents, like humans, are most sensitive to the textures that vary the most in nature.
Source: Read Full Article