Skip to main content

Hello,I am SANJEEV NAMJOSHI I am a machine learning engineer at VERSES.AI. My theoretical research interests lie at the intersection of neuroscience and machine learning. Specifically, I am interested in the computational modeling and simulation of living systems.

Sanjeev Namjoshi, Ph.D

I am a machine learning engineer at VERSES.AI. My theoretical research interests lie at the intersection of neuroscience, machine learning, and statistical physics. Specifically, I am interested in the computational modeling and simulation of living systems. These research areas heavily overlap with a diverse number of fields including embodied cognition, control theory, robotics, complex systems analysis, neuroeconomics, and theory of mind.

I am also interested in bringing wider recognition and visibility to Active Inference research. My first textbook on Active Inference will be available from MIT Press in March 2026. I am currently writing a second textbook on the related Bayesian Mechanics field which also be published by MIT Press.

Active Inference, and the more general Free Energy Principle, provide a mathematical description of the brain and living systems that may be key to unlocking further progress toward artificial general intelligence as well as numerous industrial applications in areas such as robotics, logistics, and autonomous vehicles. Since the beginning of 2023 I have led engagements with the Active Inference Institute who have kindly hosted my chapter presentations, code review, and feedback sessions for the work-in-progress textbook.

Textbooks on Active Inference and Bayesian Mechanics

My textbook, Fundamentals of Active Inference, will be available in March 2026, published by The MIT Press. The supplemental material associated with this textbook, including Jupyter notebooks and proofs for all equations in the book, is currently in preparation and will be released after the book’s publication. I am currently working on a second textbook, Fundamentals of Bayesian Mechanics. The textbooks are focused for an engineering audience – machine learning engineers and applied engineers – who are interested in applying active inference in the real world.

Collectively, these books aim to provide a technical introduction to the research areas of Active Inference, the Free Energy Principle, and Bayesian Mechanics. The mathematical ideas within these fields have the potential to revolutionize many industries and greatly advance theoretical AI research beyond what has already been achieved through deep learning and large language models in the last decade. My intention for this work is to condense over 20 years of progress and literature in the field to bring a wider degree of recognition and interest from machine learning researchers and engineers in academia and industry.

The textbooks are aimed at an upper undergraduate/early graduate level and aims to be fairly self-contained, requiring little knowledge of neuroscience and related literature but focusing heavily on building intuition, understanding the mathematics, and implementation in Python code. The prerequisites for the book are basic Python programming knowledge, probability theory, multivariate calculus, basic linear algebra, and high-school physics, topics usually covered in undergraduate natural science and engineering degrees.

Background

Active Inference is a fast-growing field within computational neuroscience that integrates ideas from machine learning, Bayesian statistics, physics, and other fields to mathematically describe the brain and human/animal behavior. The original ideas were conceived by the neuroscientist Dr. Karl Friston who has worked with a number of collaborators in the last two decades to expand the scope and applicability of the theory to the brain and living systems more generally. Active Inference has been successfully applied to many different areas of neuroscience and represents the culmination of decades of prior research. It provides an interdisciplinary way to look at a broad range of fields ranging from robotics, economics, cybernetics/control theory, machine learning and more.

The Free Energy Principle is a broader and more theoretical physics-based theory also created by Friston. The Free Energy Principle provides the theoretical and philosophical scaffold behind Active Inference and a mathematical modeling framework known as Bayesian Mechanics to model living systems that undergo “self-organization”. That is, such systems adapt to changes to their surroundings to maintain their structure over time. Under the Free Energy Principle, it is suggested that biological or artificial organisms follows the imperative of minimizing a statistical quantity known as variational free energy (negative evidence lower bound) which serves as a universal objective function in approximate Bayesian inference. Organisms that successfully minimize this quantity are the ones that are able to (actively) infer the current (and future) states of their environments and thus restrict themselves to preferable states conducive to their survival.

Supplemental Materials

Although the supplemental materials are not yet available you can click on the links below to learn about the content I have planned. Supplemental materials will be added after the book’s publication.

Code notebooksCompendium of proofsHistory of active inference and the free energy principle

Research And Talks

  • The Hidden Math Beyond All Living Systems (Machine Learning Street Talk, October 2024) [Link]
  • Education, Expectation-Maximisation, Evolution (Active Inference Insights Episode 18, March 2024) [Link]
  • How Active Inference is Bridging Neuroscience and AI (Hidden Layers Episode 10, February 2024) [Link]
  • Machine Learning Street Talk Podcast (Patreon Access Only, January 2024) [Link]
  • Engineering Explained: Bayesian mechanics (September 2023) [Link]
  • Developing Next-Gen Active Inference Tools: Broadening Accessibility, Educational Resources, and the Software Ecosystem (3rd Annual Active Inference Symposium, September 2023) [Link]

In January and February 2020 I gave two talks at the Deep Learning Meetup in Austin, TX on the history of AI. The aim of these talks was to weave together a unique historical perspective on AI that integrates developments in philosophy, mathematics, and neuroscience and the various paths toward AI as we think of it today. The recordings of these talks can be found below:

  • The History of AI Part I: Before the AI Winter (1822-1980) [Link]
  • The History of AI Part II: Modern Developments in AI and Neuroscience (1980-Present) [Link]

Active inference and Bayesian mechanics textbooks: My primary research focus is the development of a textbook and other educational resources around the fields of Active Inference and Bayesian Mechanics. The book will be published by The MIT Press. I have primarily written this book outside of work but from January-July 2023 I was on sabbatical to focus on it exclusively. More detailed information is given on the textbook page.

Animal learning-based AI: In collaboration with Roy Clymer we have written a paper showcasing his work on an AI model inspired by animal learning. The model is based on a model of the synapse but is entirely different in construction from feed-forward neural networks and does not learn through backpropogation; superficially, the model has some similarities with spiking neural networks. It includes an agent-environment interaction loop and is capable of fairly complex behavioral chaining and learning in continuous time without any explicit calculations of gradients. More information and proof of concept results (using a Roomba) can be found on Roy’s website linked above and in this YouTube video.

  • Clymer RE, Namjoshi SV. A computational model of behavioral adaptation to solve the credit assignment problem. 2023. arXiv:2311.18134v1 [q-bio.NC]. [arXiv]

Asynchronous federated learning: In collaboration with my former co-workers at KUNGFU.AI. As part of my work at KUNGFU.AI we had to utilize federated learning in order to prevent the mixture of private patient data stored in different AWS cloud accounts. Federated learning as implemented by tools such as Flwr require synchronous learning where the slowest dataset to train becomes the bottleneck for the other datasets. With our asynchronous federated learning setup it is possible to eliminate issues caused by bottlenecking datasets. This work is currently implemented in Python [Github].

  • Namjoshi SV, Green R, Sharma K, Si Z. Serverless Federated Learning with flwr-serverless. 2023. arXiv:2310.15329v1 [cs.LG]. [arXiv]

In December 2023, I joined VERSES.AI as a machine learning engineer.

At KUNGFU.AI I served as lead on a project to develop breast-cancer risk prediction software. The project involved developing a deep learning model using over 30 terrabytes of mammography data and temporal patient outcome data from hospitals around the world. The goal of the model was to provide calibrated risk scores with a 5-year predictive risk window that radiologists could use to provide more informed recommendations for their patients. This project required the application of federated learning to datasets stored on different cloud servers that could interact through VPC peering.

At Hypergiant I served as lead on a project to develop an agent-based logistics simulator. The simulation was designed such that the user could change the initial conditions of the simulator to produce different outcome data. This data could then be used for forecasting to see how various business metrics would be expected to change as a result of changing the parameters of the simulation. This project served as a test-bed for asking “what-if” questions useful in business analytics.

As part of my transition toward industry I worked in Dr. Wei Zhang’s lab on a computer vision project to detect brain cancer in 3D MRI images. I also contributed some bioinformatics analysis on a paper investigating the oncogene FGFR3-TACC3.

  • Lyu Q, Namjoshi SV, McTyre E, Topaloglu U, Barcus R, Chan MD, Cramer CK, Debinski W, Gurcan MN, Lesser GJ, Lin HK, Munden RF, Pasche BC, Sai KKS, Strowd RE, Tatter SB, Watabe K, Zhang W, Wang G, Whitlow CT. A transformer-based deep-learning approach for classifying brain metastases into primary organ sites using clinical whole-brain MRI images. Patterns (N Y). 2022. 3(11):100613. doi: 10.1016/j.patter.2022.100613. [PubMed]
  • Li T, Mehraein-Ghomi F, Forbes ME, Namjoshi SV, Ballard EA, Song Q, Chou PC, Wang X, Parker Kerrigan BC, Lang FF, Lesser G, Debinski W, Yang X, Zhang W. HSP90-CDC37 functions as a chaperone for the oncogenic FGFR3-TACC3 fusion. Mol Ther. 2022. 30(4):1610-1627. doi: 10.1016/j.ymthe.2022.02.009. [PubMed]

After my graduate work I wished to refine my knowledge in computational neuroscience. I worked with Dr. Ken Kishida’s lab on a number of different projects applying reinforcement learning and Bayesian statistics to analyze human subjects interacting with game theory-inspired tasks that investigated social norms, decision-making, and subjective consciousness. Among my major contributions was optimizing a model that could decompose a fast-scan cyclic voltammetry signal recording from patients undergoing deep-brain stimulation to distinguish different neurotransmitter measurements.

My graduate research was performed in Kimberly Raab-Graham’s lab where I worked on a mix of molecular biology and bioinformatics applied to proteomics and transcriptomics. My early work focused on functional and network analysis of proteins affected by excitation or inhibition of the protein complex mTORC1 expressed in rat or mouse hippocampal neurons (via mass spectrometry). Many of these proteins controlled by mTORC1 expression are involved in memory processes and dysregulated in neurodegenerative diseases. Main paper related to this project:

  • Niere F, Namjoshi S, Song E, Dilly GA, Schoenhard G, Zemelman BV, Mechref Y, Raab-Graham KF. Analysis of Proteins That Rapidly Change Upon Mechanistic/Mammalian Target of Rapamycin Complex 1 (mTORC1) Repression Identifies Parkinson Protein 7 (PARK7) as a Novel Protein Aberrantly Expressed in Tuberous Sclerosis Complex (TSC). Mol Cell Proteomics. 2016. 15(2):426-44. doi: 10.1074/mcp.M115.055079. [PubMed]

I have also worked with the RNA-binding protein fragile-X mental retardation protein (FMRP) whose dysregulation is implicated in neuronal dysfunction (autism spectrum disorders, depression). The researched involved utilizing RNA immunopreciptation followed by Illumina sequencing to identify the populations of RNA sequestered or released in RNA granules controlled by FMRP. Many of these RNA molecules could potentially be used in therapeutic treatments, particularly for the development of better rapidly-acting antidepressants. Main paper related to this project:

  • Heaney CF, Namjoshi SV, Uneri A, Bach EC, Weiner JL, Raab-Graham KF. Role of FMRP in rapid antidepressant effects and synapse regulation. Mol Psychiatry. 2021. 26(6):2350-2362. doi: 10.1038/s41380-020-00977-z. [PubMed]

Additionally, I wrote a review which covers the general process of devising experiments for the purpose of extracting and analyzing RNA populations to study local mRNA translation under different experimental conditions.

  • Namjoshi SV and Raab-Graham KF. Screening the Molecular Framework Underlying Local Dendritic mRNA Translation. Front Mol Neurosci. 2017. doi: 10.3389/fnmol.2017.00045. eCollection 2017. [PubMed]

Other papers from my time in the Raab-Graham Lab:

  • Niere F, Uneri A, McArdle CJ, Deng Z, Egido-Betancourt HX, Cacheaux LP, Namjoshi SV, Taylor WC, Wang X, Barth SH, Reynoldson C, Penaranda J, Stierer MP, Heaney CF, Craft S, Keene CD, Ma T, Raab-Graham KF. Aberrant DJ-1 expression underlies L-type calcium channel hypoactivity in dendrites in tuberous sclerosis complex and Alzheimer’s disease. Proc Natl Acad Sci USA. 2023. 120(45). doi: 10.1073/pnas.2301534120. [PubMed]
  • Raab-Graham KF, Workman ER, Namjoshi S, Niere F. Pushing the threshold: How NMDAR antagonists induce homeostasis through protein synthesis to remedy depression. Brain Res. 2016. 1647:94-104. doi: 10.1016/j.brainres.2016.04.020. [PubMed]
  • Wolfe SA, Workman ER, Heaney CF, Niere F, Namjoshi S, Cacheaux LP, Farris SP, Drew MR, Zemelman BV, Harris RA, Raab-Graham KF. FMRP regulates an ethanol-dependent shift in GABABR function and expression with rapid antidepressant properties. Nat Commun. 2016. 7:12867. doi: 10.1038/ncomms12867. [PubMed]

I also worked in Dr. Scott Stevens’ Lab focusing on yeast genetics. This lab provided my background in experimental molecular biology. I performed functional analysis of proteins in the spliceosome, the primary protein complex in the cell which carries out RNA processing.

Experience/Education

  • Jan 2024 – present (Full-time): Machine learning engineer at VERSES.AI
  • Jan 2023 – Jul 2023: Sabbatical to write Active Inference Textbook
  • Jun 2021 – Dec 2023 (Full-time): Senior Machine Learning Engineer at KUNGFU.AI
  • Jun 2020 – May 2021 (Full-time): Data Scientist at Hypergiant
  • Apr 2020 – Jun 2020 (Contract): Data Scientist at HookBang
  • Sep 2019 – Mar 2020 (Full-time): Data Scientist at Yonder (formally known as New Knowledge)
  • Sep 2018 – Sep 2019 (Full-time): Postdoctoral Researcher in Dr. Wei Zhang’s lab
  • Dec 2013 – Apr 2017 (Full-time): Graduate Student Research Assistant in Dr. Kimberly Raab-Graham’s lab
  • Jun 2021 – Dec 2023 (Full-time): Senior Machine Learning Engineer at KUNGFU.AI
  • Jun 2011 – Aug 2013 (Full-time): Graduate Student Research Assistant in Dr. Scott Steven’s lab
  • Sep 2010 – May 2017: Ph.D. in Neuroscience from the University of Texas at Austin
  • Aug 2006 – May 2010: B.S. in Neurobiology from the University of Texas at Austin

Article Title

June 5, 2025 in Article

Section 01

Coming soon!
Read More
June 7, 2025 in Article

Blog article #1

Coming soon!
Read More