Repository logo

Infoscience

  • English
  • French
Log In
Logo EPFL, École polytechnique fédérale de Lausanne

Infoscience

  • English
  • French
Log In
  1. Home
  2. Academic and Research Output
  3. Conferences, Workshops, Symposiums, and Seminars
  4. Capsule networks, but not convolutional networks explain global configurational visual effects
 
conference paper

Capsule networks, but not convolutional networks explain global configurational visual effects

Doerig, Adrien  
•
Bornet, Alban
•
Herzog, Michael H.
2019
Perception
41st European Conference on Visual Perception (ECVP)

In human vision, perception of local features depends on all elements in the visual field and their exact configuration. For example, observers performed a vernier discrimination task. When a surrounding square was added to the vernier, the task became much more difficult: a classic crowding effect. Crucially, adding more flanking squares improved performance (uncrowding). In addition, in displays of squares and stars, small changes in the configuration changed performance strongly. Here, we show that convolutional neural networks fail to address the global aspects of configuration because, first, the target and the flankers’ representations at a given layer are pooled within the receptive fields of the subsequent layer, leading to poor performance. Second, far away elements cannot interact with the vernier to produce uncrowding. We show that capsule networks, a new kind of neural network that explicitly takes configuration into account, can capture the experimental results well.

  • Details
  • Metrics
Logo EPFL, École polytechnique fédérale de Lausanne
  • Contact
  • infoscience@epfl.ch

  • Follow us on Facebook
  • Follow us on Instagram
  • Follow us on LinkedIn
  • Follow us on X
  • Follow us on Youtube
AccessibilityLegal noticePrivacy policyCookie settingsEnd User AgreementGet helpFeedback

Infoscience is a service managed and provided by the Library and IT Services of EPFL. © EPFL, tous droits réservés