Research
What makes a system intelligent? The biological brain is our only example of general intelligence, yet modern AI systems are beginning to demonstrate similarly broad capabilities. Our lab approaches intelligence as a scientific problem, seeking the core design principles that enable adaptable, goal‑directed behavior in both brains and machines and studying how these principles scale to impact society.
Because the brain emerged through a complex evolutionary process, we reverse‑engineer neural circuits by simulating evolution in silico, toggling components such as architecture class, objective function, data stream and learning rule (depicted below). At the same time, we investigate how AI systems interact with human values and economic structures, working toward safety and alignment frameworks that support equitable outcomes.
By evaluating the models generated from these components against both neural recordings, behavioral data and societal outcomes, our long‑term goal is not only to build normative accounts of how intelligent behavior arises but also to guide the design of AI systems that are capable, aligned with human values and beneficial to society.
Below are some representative papers that are relevant to the above questions. When possible, I try to link to the freely accessible preprint; however, this may differ from the final published version. For a full publication list, see my CV. See here for presentations on some of this work, as well as here and here for a short or long form (includes past work) video overview, respectively.
AI Safety & Society
Our lab is also engaged in research on AI alignment and the societal impact of AI. Representative papers include:
-
A. Nayebi. An AI capability threshold for rent‑funded universal basic income in an AI‑automated economy. In this economic analysis we derive conditions under which AI‑generated profits could sustainably finance a universal basic income. We show that AI systems must achieve only ~5–6× existing automation productivity to fund an 11%‑of‑GDP UBI, and that raising the public revenue share to about 33% lowers this threshold to ~3x. 2025. [code][summary]
-
A. Nayebi. Intrinsic barriers and practical pathways for human‑AI alignment: an agreement‑based complexity analysis. This paper formalizes AI alignment as a multi‑objective optimization problem and identifies information‑theoretic lower bounds showing that once either the number of objectives or the number of agents is large enough, no interaction or rationality can avoid intrinsic alignment overheads. These results highlight fundamental complexity‑theoretic constraints and provide guidelines for safer, scalable human–AI collaboration. 2025. [summary][talk recording]
NeuroAI (Selected Papers)
(*: joint first author; †: joint senior author)
-
R. D. Keller, A. Tornell, F. C. Pei, X. Pitkow, L. Kozachkov†, A. Nayebi†.
Autonomous behavior and whole‑brain dynamics emerge in embodied zebrafish agents with model‑based intrinsic motivation. 2025. [summary] -
T. Chung, Y. Shen, Nathan C. L. Kong, A. Nayebi.
Task‑optimized convolutional recurrent networks align with tactile processing in the rodent brain. 2025. [code][summary] -
J. Feather*, M. Khosla*, N. A. Ratan Murthy*, A. Nayebi*. Brain-model evaluations need the NeuroAI Turing Test. 2025. [code][summary][Brain Inspired podcast]
-
A. Nayebi, R. Rajalingham, M. Jazayeri, G.R. Yang. Neural foundations of mental simulation: future prediction of latent representations on dynamic scenes. Advances in Neural Information Processing Systems (NeurIPS), Volume 36 (2023): 70548-70561. (Selected for spotlight presentation) [code][summary][NeurIPS 5 min talk recording][MIT CBMM talk recording][MIT News article][MIT CBMM short video]
-
A. Nayebi*, N.C.L. Kong*, C. Zhuang, J.L. Gardner, A.M. Norcia, D.L.K. Yamins. Mouse visual cortex as a limited resource system that self-learns an ecologically-general representation. PLOS Computational Biology, Volume 19 (2023): 1-36. [code][summary][talk recording]
-
A. Nayebi, J. Sagastuy-Brena, D.M. Bear, K. Kar, J. Kubilius, S. Ganguli, D. Sussillo, J.J. DiCarlo, D.L.K. Yamins. Recurrent connections in the primate ventral visual stream mediate a tradeoff between task performance and network size during core object recognition. Neural Computation, Volume 34 (2022): 1652-1675. [code][summary]
-
A. Nayebi, A. Attinger, M.G. Campbell, K. Hardcastle, I.I.C. Low, C.S. Mallory, G.C. Mel, B. Sorscher, A.H. Williams, S. Ganguli, L.M. Giocomo, D.L.K. Yamins. Explaining heterogeneity in medial entorhinal cortex with task-driven neural networks. Advances in Neural Information Processing Systems (NeurIPS), Volume 34 (2021). (Selected for spotlight presentation) [code][summary][talk recording]
-
C. Zhuang, S. Yan, A. Nayebi, M. Schrimpf, M.C. Frank, J.J. DiCarlo, D.L.K. Yamins. Unsupervised neural network models of the ventral visual stream. Proceedings of the National Academy of Sciences of the United States of America (PNAS), Volume 118 (2021). [code][talk recording]
-
A. Nayebi*, S. Srivastava*, S. Ganguli, D.L.K. Yamins. Identifying learning rules from neural network observables. Advances in Neural Information Processing Systems (NeurIPS), Volume 33 (2020). (Selected for spotlight presentation) [code][summary][talk recording][blogpost][The Economist news article]
- D. Kunin*, A. Nayebi*, J. Sagastuy-Brena*, S. Ganguli, J. Bloom, D.L.K. Yamins. Two routes to scalable credit assignment without weight symmetry. Proceedings of the 37th International Conference on Machine Learning (ICML), PMLR 119 (2020):5511-5521. [code][summary][talk recording]
-
A. Nayebi*, D.M. Bear*, J. Kubilius*, K. Kar, S. Ganguli, D. Sussillo, J.J. DiCarlo, D.L.K. Yamins. Task-driven convolutional recurrent models of the visual system. Advances in Neural Information Processing Systems (NeurIPS), Volume 31 (2018): 5290-5301. [code][summary][blogpost]
-
L.T. McIntosh*, N. Maheswaranathan*, A. Nayebi, S. Ganguli, S.A. Baccus. Deep learning models of the retinal response to natural scenes. Advances in Neural Information Processing Systems (NIPS), Volume 29 (2016): 1369-1377. [code]