• Login
    View Item 
    •   SMARTech Home
    • Georgia Tech Neuro
    • GT Neuro Seminar Series
    • View Item
    •   SMARTech Home
    • Georgia Tech Neuro
    • GT Neuro Seminar Series
    • View Item
    JavaScript is disabled for your browser. Some features of this site may not work without it.

    Investigating the basis of noise-robust object recognition in humans and convolutional neural networks

    Thumbnail
    View/Open
    ftong.mp4 (245.3Mb)
    ftong_videostream.html (1.096Kb)
    transcript.txt (55.61Kb)
    thumbnail.jpg (46.33Kb)
    Date
    2021-11-29
    Author
    Tong, Frank
    Metadata
    Show full item record
    Abstract
    It has been claimed that convolutional neural networks (CNNs) have now achieved human-level performance at object recognition tasks. However, modest changes to the object stimuli or to the viewing conditions can sometimes cause state-of-the-art CNNs to fail, raising questions as to whether they truly process visual information in a manner that mimics the human visual system. Here, I will present behavioral and neuroimaging data demonstrating the robustness of human vision when tasked with recognizing objects in severe levels of visual noise. Our functional MRI studies demonstrate the powerful role of top-down attentional feedback in dampening neural responses to visual noise, clutter, and competing overlapping objects. In experiments that directly pit human observers and CNNs, we find that humans outperform CNNs by a large margin and that they are affected by white noise and spatially correlated (‘pink’) noise in qualitatively different ways. We developed a noise-training procedure, generating noisy images of objects with low signal-to-noise ratio, to investigate whether CNNs can acquire robustness that better matches human vision. After noise training, CNNs could outperform human observers while exhibiting more similar qualitative patterns of performance. Moreover, noise-trained CNNs provided a better model for predicting human recognition thresholds on an image-by-image basis. Layer-specific analyses revealed that the contaminating effects of noise were dampened, rather than amplified, across successive stages of the noise-trained network. Our findings suggest that CNNs can learn noise-robust representations that better approximate human visual processing, though it remains an open question as to how the incorporation of top-down attention mechanisms might further improve the correspondence between artificial and biological visual systems.
    URI
    http://hdl.handle.net/1853/65552
    Collections
    • GT Neuro Seminar Series [109]

    Browse

    All of SMARTechCommunities & CollectionsDatesAuthorsTitlesSubjectsTypesThis CollectionDatesAuthorsTitlesSubjectsTypes

    My SMARTech

    Login

    Statistics

    View Usage StatisticsView Google Analytics Statistics
    facebook instagram twitter youtube
    • My Account
    • Contact us
    • Directory
    • Campus Map
    • Support/Give
    • Library Accessibility
      • About SMARTech
      • SMARTech Terms of Use
    Georgia Tech Library266 4th Street NW, Atlanta, GA 30332
    404.894.4500
    • Emergency Information
    • Legal and Privacy Information
    • Human Trafficking Notice
    • Accessibility
    • Accountability
    • Accreditation
    • Employment
    © 2020 Georgia Institute of Technology