Yamins, D. L. & DiCarlo, J. J. Using goal-driven deep learning models to understand sensory cortex. Nat. Neurosci. 19, 356–365 (2016).
Kriegeskorte, N. Deep neural networks: a new framework for modeling biological vision and brain information processing. Annu. Rev. Vis. Sci. 1, 417–446 (2015).
Heeger, D. J. Half-squaring in responses of cat striate cells. Vis. Neurosci. 9, 427–443 (1992).
Rust, N. C., Schwartz, O., Movshon, J. A. & Simoncelli, E. P. Spatiotemporal elements of macaque V1 receptive fields. Neuron 46, 945–956 (2005).
Pillow, J. W. et al. Spatio-temporal correlations and visual signalling in a complete neuronal population. Nature 454, 995–999 (2008).
Vintch, B., Movshon, J. A. & Simoncelli, E. P. A convolutional subunit model for neuronal responses in macaque V1. J. Neurosci. 35, 14829–14841 (2015).
Ustyuzhaninov, I. et al. Digital twin reveals combinatorial code of non-linear computations in the mouse primary visual cortex. Preprint at bioRxiv https://doi.org/10.1101/2022.02.10.479884 (2022).
Maheswaranathan, N. et al. Interpreting the retinal neural code for natural scenes: from computations to neurons. Neuron 111, 2742–2755 (2023).
David, S. V., Hayden, B. Y. & Gallant, J. L. Spectral receptive field properties explain shape selectivity in area V4. J. Neurophysiol. 96, 3492–3505 (2006).
Yamins, D. L. et al. Performance-optimized hierarchical models predict neural responses in higher visual cortex. Proc. Natl Acad. Sci. USA 111, 8619–8624 (2014).
Cadieu, C. et al. A model of V4 shape selectivity and invariance. J. Neurophysiol. 98, 1733–1750 (2007).
Bahri, Y., Dyer, E., Kaplan, J., Lee, J. & Sharma, U. Explaining neural scaling laws. Proc. Natl Acad. Sci. USA 121, e2311878121 (2024).
Liu, Z. et al. A ConvNet for the 2020s. In Proc. IEEE/CVF Conference on Computer Vision and Pattern Recognition 11966–11976 (IEEE, 2022).
Kaplan, J. et al. Scaling laws for neural language models. Preprint at arXiv https://doi.org/10.48550/arXiv.2001.08361 (2020).
Schrimpf, M. et al. Integrative benchmarking to advance neurally mechanistic models of human intelligence. Neuron 108, 413–423 (2020).
Hinton, G., Vinyals, O. & Dean, J. Distilling the knowledge in a neural network. Preprint at arXiv https://doi.org/10.48550/arXiv.1503.02531 (2015).
Han, S., Mao, H. & Dally, W. J. Deep compression: compressing deep neural networks with pruning, trained quantization and huffman coding. In Proc. 4th International Conference on Learning Representations (ICLR, 2016).
Butts, D. A. Data-driven approaches to understanding visual neuron activity. Annu. Rev. Vis. Sci. 5, 451–477 (2019).
Pierzchlewicz, P. et al. Energy guided diffusion for generating neurally exciting images. In Proc. Advances in Neural Information Processing Systems 36 (eds Oh, A. et al.) (NeurIPS, 2023).
Bashivan, P., Kar, K. & DiCarlo, J. J. Neural population control via deep image synthesis. Science 364, eaav9436 (2019).
DiMattina, C. & Zhang, K. Adaptive stimulus optimization for sensory systems neuroscience. Front. Neural Circuits 7, 101 (2013).
Pillow, J. W. & Park, M. in Closed Loop Neuroscience (ed. El Hady, A.) Ch. 1 (Elsevier, 2016).
Seung, H. S., Opper, M. & Sompolinsky, H. Query by committee. In Proc. Fifth Annual Workshop on Computational Learning Theory 287–294 (Association for Computing Machinery, 1992).
Cowley, B. & Pillow, J. W. High-contrast “gaudy” images improve the training of deep neural network models of visual cortex. In. Proc. Advances in Neural Information Processing Systems 33 (eds Larochelle, H. et al.) (NeurIPS, 2020).
Frankle, J. & Carbin, M. The lottery ticket hypothesis: finding sparse, trainable neural networks. In Proc. 7th International Conference on Learning Representations 8954–8995 (ICLR, 2019).
He, Y., Zhang, X. & Sun, J. Channel pruning for accelerating very deep neural networks. In Proc. IEEE International Conference on Computer Vision 1398–1406 (IEEE, 2017).
Wang, T. et al. Large-scale calcium imaging reveals a systematic V4 map for encoding natural scenes. Nat. Commun. 15, 6401 (2024).
Du, F., Núñez-Ochoa, M. A., Pachitariu, M. & Stringer, C. Towards a simplified model of primary visual cortex. Preprint at bioRxiv https://doi.org/10.1101/2024.06.30.601394 (2024).
Kamali, F., Suratgar, A. A., Menhaj, M. & Abbasi-Asl, R. Compression-enabled interpretability of voxelwise encoding models. PLoS Comput. Biol. 21, e1012822 (2025).
Cowley, B. R., Williamson, R. C., Acar, K., Smith, M. A. & Yu, B. M. Adaptive stimulus selection for optimizing neural population responses. In Proc. Advances in Neural Information. Processing Systems 30 (eds Guyon, I. et al.) (NeurIPS, 2017).
Willeke, K. F. et al. Deep learning-driven characterization of single cell tuning in primate visual area V4 unveils topological organization. Preprint at bioRxiv https://doi.org/10.1101/2023.05.12.540591 (2023).
Szegedy, C. et al. Intriguing properties of neural networks. In Proc. 2nd International Conference on Learning Representations (ICLR, 2014).
Guo, C. et al. Adversarially trained neural representations may already be as robust as corresponding biological neural representations. In Proc. 39th International Conference on Machine Learning (eds Chaudhuri, K. et al.) 8072–8081 (PMLR, 2022).
Berardino, A., Ballé, J., Laparra, V. & Simoncelli, E. Eigen-distortions of hierarchical representations. In Proc. Advances in neural information processing systems 30 (eds Guyon, I. et al.) (NeurIPS, 2017).
Krizhevsky, A., Sutskever, I. & Hinton, G. E. ImageNet classification with deep convolutional neural networks. Commun. ACM https://doi.org/10.1145/3065386 (2017).
DiCarlo, J. J., Zoccolan, D. & Rust, N. C. How does the brain solve visual object recognition? Neuron 73, 415–434 (2012).
Dapello, J. et al. Simulating a primary visual cortex at the front of cnns improves robustness to image perturbations. In Proc. Advances in Neural Information Processing Systems 33 (eds Larochelle, H. et al.) (NeurIPS, 2020).
Federer, C., Xu, H., Fyshe, A. & Zylberberg, J. Improved object recognition using neural networks trained to mimic the brain’s statistical properties. Neural Netw. 131, 103–114 (2020).
Kornblith, S., Norouzi, M., Lee, H. & Hinton, G. Similarity of neural network representations revisited. In Proc. 36th International Conference on Machine Learning (eds Chaudhuri, K. & Salakhutdinov, R.) 3519–3529 (PMLR, 2019).
Majaj, N. J., Hong, H., Solomon, E. A. & DiCarlo, J. J. Simple learned weighted sums of inferior temporal neuronal firing rates accurately predict human core object recognition performance. J. Neurosci. 35, 13402–13418 (2015).
Cadena, S. A. et al. Deep convolutional models improve predictions of macaque V1 responses to natural images. PLoS Comput. Biol. 15, e1006897 (2019).
Conway, B. R. The organization and operation of inferior temporal cortex. Annu. Rev. Vis. Sci. 4, 381–402 (2018).
Pasupathy, A., Popovkina, D. V. & Kim, T. Visual functions of primate area V4. Annu. Rev. Vis. Sci. 6, 363–385 (2020).
Heeger, D. J. Normalization of cell responses in cat striate cortex. Vis. Neurosci. 9, 181–197 (1992).
Carandini, M. & Heeger, D. J. Normalization as a canonical neural computation. Nat. Rev. Neurosci. 13, 51–62 (2012).
Coen-Cagli, R., Kohn, A. & Schwartz, O. Flexible gating of contextual influences in natural vision. Nat. Neurosci. 18, 1648 (2015).
Burg, M. F. et al. Learning divisive normalization in primary visual cortex. PLoS Comput. Biol. 17, e1009028 (2021).
Ruff, D. A. & Cohen, M. R. A normalization model suggests that attention changes the weighting of inputs between visual areas. Proc. Natl Acad. Sci. USA 114, E4085–E4094 (2017).
Verhoef, B.-E. & Maunsell, J. H. Attention-related changes in correlated neuronal activity arise from normalization mechanisms. Nat. Neurosci. 20, 969–977 (2017).
Ungerleider, L. G., Galkin, T. W., Desimone, R. & Gattass, R. Cortical connections of area V4 in the macaque. Cereb. Cortex 18, 477–499 (2008).
Semedo, J. D. et al. Feedforward and feedback interactions between visual cortical areas use different population activity patterns. Nat. Commun. 13, 1099 (2022).
Jun, N. Y. et al. Coordinated multiplexing of information about separate objects in visual cortex. eLife 11, e76452 (2022).
Day-Cooney, J., Cone, J. J. & Maunsell, J. H. Perceptual weighting of V1 spikes revealed by optogenetic white noise stimulation. J. Neurosci. 42, 3122–3132 (2022).
Shahbazi, E., Ma, T., Pernuš, M., Scheirer, W. & Afraz, A. Perceptography unveils the causal contribution of inferior temporal cortex to visual perception. Nat. Commun. 15, 3347 (2024).
Smith Breault, M. SciDraw: monkey brain. Zenodo https://doi.org/10.5281/zenodo.3926117 (2025).
Stan, P. L. & Smith, M. A. Recent visual experience reshapes V4 neuronal activity and improves perceptual performance. J. Neurosci. 44, e1764232024 (2024).
Issar, D., Williamson, R. C., Khanna, S. B. & Smith, M. A. A neural network for online spike classification that improves decoding accuracy. J. Neurophysiol. 123, 1472–1485 (2020).
Pospisil, D. A. & Bair, W. The unbiased estimation of the fraction of variance explained by a model. PLoS Comput. Biol. 17, e1009212 (2021).
Cadena, S. A. et al. Diverse task-driven modeling of macaque v4 reveals functional specialization towards semantic tasks. PLoS Comput. Biol. 20, e1012056 (2024).
Sponheim, C. et al. Longevity and reliability of chronic unit recordings using the Utah, intracortical multi-electrode arrays. J. Neural Eng. 18, 066044 (2021).
Degenhart, A. D. et al. Stabilization of a brain–computer interface via the alignment of low-dimensional spaces of neural activity. Nat. Biomed. Eng. 4, 672–685 (2020).
Kleiner, M., Brainard, D. & Pelli, D. What’s new in psychtoolbox-3? Perception 36, 1–16 (2007).
Cohen, M. R. & Kohn, A. Measuring and interpreting neuronal correlations. Nat. Neurosci. 14, 811–819 (2011).
Walker, E. Y. et al. Inception loops discover what excites neurons most using deep predictive models. Nat. Neurosci. 22, 2060–2065 (2019).
Thomee, B. et al. YFCC100M: the new data in multimedia research. Commun. ACM 59, 64–73 (2016).
Deng, J. et al. ImageNet: a large-scale hierarchical image database. In Proc. 2009 IEEE Conference on Computer Vision and Pattern Recognition 248–255 (IEEE, 2009).
Lewi, J., Butera, R. & Paninski, L. Sequential optimal design of neurophysiology experiments. Neural Comput. 21, 619–687 (2009).
Benda, J., Gollisch, T., Machens, C. K. & Herz, A. V. From response to stimulus: adaptive sampling in sensory physiology. Curr. Opin. Neurobiol. 17, 430–436 (2007).
DiMattina, C. & Zhang, K. Active data collection for efficient estimation and comparison of nonlinear neural models. Neural Comput. 23, 2242–2288 (2011).
Klindt, D., Ecker, A. S., Euler, T. & Bethge, M. Neural system identification for large populations separating “what” and “where”. In Proc. Advances in Neural Information Processing Systems 30 (eds Guyon, I. et al.) (NeurIPS, 2017).
Richards, B. A. et al. A deep learning framework for neuroscience. Nat. Neurosci. 22, 1761–1770 (2019).
Doerig, A. et al. The neuroconnectionist research programme. Nat. Rev. Neurosci. 24, 431–450 (2023).
Chollet, F. et al. Keras. Keras https://keras.io (2015).
Paszke, A. et al. PyTorch: an imperative style, high-performance deep learning library. In Proc. Advances in Neural Information Processing Systems 32 (eds Wallach, H. et al.) (NeurIPS, 2019).
He, K., Zhang, X., Ren, S. & Sun, J. Deep residual learning for image recognition. In Proc. 2016 IEEE Conference on Computer Vision and Pattern Recognition 770–778 (IEEE, 2016).
Howard, A. G. et al. Mobilenets: efficient convolutional neural networks for mobile vision applications. Preprint at arXiv https://doi.org/10.48550/arXiv.1704.04861 (2017).
Huang, G., Liu, Z., Van Der Maaten, L. & Weinberger, K. Q. Densely connected convolutional networks. In Proc. 2017 IEEE Conference on Computer Vision and Pattern Recognition 2261–2269 (IEEE, 2017).
Chollet, F. Xception: deep learning with depthwise separable convolutions. In Proc. 2017 IEEE Conference on Computer Vision and Pattern Recognition 1800–1817 (IEEE, 2017).
Szegedy, C., Ioffe, S., Vanhoucke, V. & Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. In Proc. Thirty-First AAAI Conference on Artificial Intelligence 4278–4284 (AAAI, 2017).
Simonyan, K. & Zisserman, A. Very deep convolutional networks for large-scale image recognition. Preprint at arXiv https://doi.org/10.48550/arXiv.1409.1556 (2014).
Zoph, B., Vasudevan, V., Shlens, J. & Le, Q. V. Learning transferable architectures for scalable image recognition. In Proc. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 8697–8710 (IEEE, 2018).
Kubilius, J. et al. CORnet: modeling the neural mechanisms of core object recognition. Preprint at bioRxiv https://doi.org/10.1101/408385 (2018).
Salman, H., Ilyas, A., Engstrom, L., Kapoor, A. & Madry, A. Do adversarially robust imagenet models transfer better? In Proc. Advances in Neural Information Processing Systems 33 (eds Larochelle, H. et al.) (NeurIPS, 2020).
Zhuang, C. et al. Unsupervised neural network models of the ventral visual stream. Proc. Natl Acad. Sci. USA 118, e2014196118 (2021).
Ioffe, S. & Szegedy, C. Batch normalization: accelerating deep network training by reducing internal covariate shift. In Proc. 32nd International Conference on Machine Learning (eds Bach, F. & Blei, D.) 448–456 (PMLR, 2015).
Lurz, K.-K. et al. Generalization in data-driven models of primary visual cortex. In Proc. 9th International Conference on Learning Representations (ICLR, 2021).
Turishcheva, P., Burg, M. F., Sinz, F. H. & Ecker, A. S. Reproducibility of predictive networks for mouse visual cortex. In Proc. Advances in Neural Information Processing Systems 37 (eds Globerson, A. et al.) (NeurIPS, 2024).
Kingma, D. P. & Ba, J. Adam: a method for stochastic optimization. Preprint at arXiv https://doi.org/10.48550/arXiv.1412.6980 (2014).
Han, S., Pool, J., Tran, J., & Dally, W. Learning both weights and connections for efficient neural network. In Proc. Advances in Neural Information Processing Systems 28 (eds Cortes, C. et al.) (NeurIPS, 2015).
Luo, J.-H., Wu, J. & Lin, W. ThiNet: a filter level pruning method for deep neural network compression. In Proc. IEEE International Conference on Computer Vision 5068–5076 (IEEE, 2017).
Li, H., Kadav, A., Durdanovic, I., Samet, H. & Graf, H. P. Pruning filters for efficient convnets. In Proc. 5th International Conference on Learning Representations 1683–1695 (ICLR, 2017).
Goodfellow, I. J., Shlens, J. & Szegedy, C. Explaining and harnessing adversarial examples. Preprint at arXiv https://doi.org/10.48550/arXiv.1412.6572 (2015).
Veerabadran, V. et al. Subtle adversarial image manipulations influence both human and machine perception. Nat. Commun. 14, 4933 (2023).
Morcos, A., Raghu, M. & Bengio, S. Insights on representational similarity in neural networks with canonical correlation. In Proc. Advances in Neural Information Processing Systems 31 (eds Bengio, S. et al.) (NeurIPS, 2018).
Williams, A. H., Kunz, E., Kornblith, S. & Linderman, S. Generalized shape metrics on neural representations. In Proc. Advances in Neural Information Processing Systems 34 (eds Ranzato, M. et al.) (NeurIPS, 2021).
Pattisapu, S. SciDraw: monkey brain. Zenodo https://doi.org/10.5281/zenodo.17553661 (2025).
Cowley, B. R., Stan, P. L., Pillow, J. W. & Smith, M. A. Data from “Compact deep neural network models of visual cortex”. Carnegie Mellon University https://doi.org/10.1184/R1/30500090 (2025).
Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J. & Wojna, Z. Rethinking the inception architecture for computer vision. In Proc. 2016 IEEE Conference on Computer Vision and Pattern Recognition 2818–2826 (2016).
Pospisil, D. A., Pasupathy, A. & Bair, W. ‘Artiphysiology’ reveals V4-like shape tuning in a deep network trained for image classification. eLife 7, e38242 (2018).
Olah, C. et al. Zoom in: an introduction to circuits. Distill 5, e00024-001 (2020).
Goh, G. et al. Multimodal neurons in artificial neural networks. Distill 6, e30 (2021).
Ponce, C. R. et al. Evolving images for visual neurons using a deep generative network reveals coding principles and neuronal preferences. Cell 177, 999–1009 (2019).
Gallant, J. L., Connor, C. E., Rakshit, S., Lewis, J. W. & Van Essen, D. C. Neural responses to polar, hyperbolic, and cartesian gratings in area V4 of the macaque monkey. J. Neurophysiol. 76, 2718–2739 (1996).
Pasupathy, A. & Connor, C. E. Responses to contour features in macaque area V4. J. Neurophysiol. 82, 2490–2502 (1999).
Tanigawa, H., Lu, H. D. & Roe, A. W. Functional organization for color and orientation in macaque V4. Nat. Neurosci. 13, 1542–1548 (2010).
Nandy, A. S., Sharpee, T. O., Reynolds, J. H. & Mitchell, J. F. The fine structure of shape tuning in area V4. Neuron 78, 1102–1115 (2013).
Li, M., Liu, F., Juusola, M. & Tang, S. Perceptual color map in macaque visual area V4. J. Neurosci. 34, 202–217 (2014).
Okazawa, G., Tajima, S. & Komatsu, H. Image statistics underlying natural texture selectivity of neurons in macaque V4. Proc. Natl Acad. Sci. USA 112, E351–E360 (2015).
Lieber, J. D., Oleskiw, T. D., Simoncelli, E. P. & Movshon, J. A. Responses of neurons in macaque V4 to object and texture images. Preprint at bioRxiv https://doi.org/10.1101/2024.02.20.581273 (2025).
Carlson, E. T., Rasquinha, R. J., Zhang, K. & Connor, C. E. A sparse object coding scheme in area V4. Curr. Biol. 21, 288–293 (2011).
Abbasi-Asl, R. et al. The DeepTune framework for modeling and characterizing neurons in visual cortex area V4. Preprint at bioRxiv https://doi.org/10.1101/465534 (2018).
Hofer, H., Carroll, J., Neitz, J., Neitz, M. & Williams, D. R. Organization of the human trichromatic cone mosaic. J. Neurosci. 25, 9669–9679 (2005).
Brainard, D. H. Color and the cone mosaic. Annu. Rev. Vis. Sci. 1, 519–546 (2015).
Simonyan, K., Vedaldi, A. & Zisserman, A. Deep inside convolutional networks: visualising image classification models and saliency maps. In Proc. Workshop at 2nd International Conference on Learning Representations (eds. Bengio, Y. & LeCun, Y.) (ICLR, 2014).

