
Project Introduction
We inspire our architecture on the magnocellular and parvocellular streams in the visual cortex in order to make it more robust to blurs, occlusions, and other out-of-distribution test images. The magnocellular stream is color-blind and carries coarse-grained information about slight brightness changes, while the parvocellular stream carries more fine-grained details that are informative about local color differences (Bear et al. 2007). These parallel processes interact through complex feed-forward, lateral, and top-down connectivity in the brain (Medathati et al. 2016), but we will simplify our approach and consider the feed-forward aspects of this neural circuit. We will simulate the magnocellular and parvocellular streams by using two parallel CNNs, the outputs of which are merged through an averaging layer and passed on to a fully-connected network that produces a classification of images. Architectures similar to our implementation have been used to identify gestures by Khurana et al. [2019]. We will assess the performance of people on a subset of the testing set that our algorithm is tested on as a benchmark for optimal results. We hypothesize that our Dual-stream CNN will be more robust to occlusions and blurriness that are partly out-of-distribution compared to traditional CNNs. We also expect our algorithm to achieve closer to human performance
See Magno/Parvo CNN Research Paper Here and GitHub Repo Here