Category-specific perceptual learning of robust object recognition modelled using deep neural networks
Hojin Jang and
Frank Tong
PLOS Computational Biology, 2025, vol. 21, issue 9, 1-19
Abstract:
Object recognition in real-world environments requires dealing with considerable ambiguity, yet the human visual system is highly robust to noisy viewing conditions. Here, we investigated the role of perceptual learning in the acquisition of robustness in both humans and deep neural networks (DNNs). Specifically, we sought to determine whether perceptual training with object images in Gaussian noise, drawn from certain animate or inanimate categories, would lead to category-specific or category-general improvements in human robustness. Moreover, might DNNs provide viable models of human perceptual learning? Both before and after training, we evaluated the noise threshold required for accurate recognition using novel object images. Human observers were quite robust to noise before training, but showed additional category-specific improvement after training with only a few hundred noisy object examples. In comparison, standard DNNs initially lacked robustness, then showed both category-general and category-specific learning after training with the same noisy examples. We further evaluated DNN models that were pre-trained with moderately noisy images to match human pre-training accuracy. Notably, these models only showed category-specific improvement, matching the overall pattern of learning exhibited by human observers. A layer-wise analysis of DNN responses revealed that category-general learning effects emerged in the lower layers, whereas category-specific improvements emerged in the higher layers. Our findings provide support for the notion that robustness to noisy visual conditions arises through learning, humans likely acquire robustness from everyday encounters with real-world noise, and additional category-specific improvements exhibited by humans and DNNs involve learning at higher levels of visual representation.Author summary: We explored how humans and artificial neural networks learn to recognize objects under noisy and ambiguous conditions, which is crucial for making sense of complex, real-world environments. Humans are naturally adept at identifying objects even when visibility is poor, like on a rainy or snowy day, or when objects are partially hidden. We wanted to ask if humans or neural networks receive training with very noisy images of objects, do they get better at the task? Also, if they are trained specifically with animate or inanimate object images, would recognition improve in general or only for the trained category? We found that humans became better at recognizing new object images in noisy conditions, but only for the categories they were trained on. Artificial networks initially struggled with noisy images but showed some general improvement from training, plus further benefits for the trained category. Interestingly, networks that were pre-trained to mimic the initial robustness of human observers only showed category-specific benefits of training, mirroring the effects of training in humans. Our findings highlight how humans adapt to challenging visual conditions, suggesting that learning plays an important role in understanding and navigating noisy, real-world settings.
Date: 2025
References: Add references at CitEc
Citations:
Downloads: (external link)
https://journals.plos.org/ploscompbiol/article?id=10.1371/journal.pcbi.1013529 (text/html)
https://journals.plos.org/ploscompbiol/article/fil ... 13529&type=printable (application/pdf)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:plo:pcbi00:1013529
DOI: 10.1371/journal.pcbi.1013529
Access Statistics for this article
More articles in PLOS Computational Biology from Public Library of Science
Bibliographic data for series maintained by ploscompbiol ().