Skip to main content
eScholarship
Open Access Publications from the University of California

Bridging DFT and DNNs: A neural dynamic process model of scene representation, guided visual search and scene grammar in natural scenes

Creative Commons 'BY' version 4.0 license
Abstract

We extend our previous neural dynamic models of visual search and scene memory (Grieben et al. (2020); Grieben and Schöner (2021)) to move beyond classical “laboratory” stimuli. The new model can autonomously explore a natural scene and build a scene memory of recognized objects and their locations. It is also capable of guided visual search for object categories in the scene. This is achieved by learning object templates for object recognition, and feature guidance templates for visual search and associating them to categorical concepts. We address how preattentive shape can be extracted from the visual input and how scene guidance, specifically, scene gramar (Võ, 2021), emerge. For the first time, we embed feature extraction by a headless deep convolutional neural network (CNN) in a neural dynamic (DFT) architecture by learning a mapping from the distributed feature representation of the CNN to the localist representation of a dynamic neural field.

Main Content
For improved accessibility of PDF content, download the file to your device.
Current View