Cross-Species Transfer Learning for Electrophysiology-to-Transcriptomics Mapping in Cortical GABAergic Interneurons
2026-03-11 • Machine Learning
Machine Learning
AI summaryⓘ
The authors used data from mouse and human brains to study special brain cells called GABAergic interneurons, which help control brain activity. They checked that previous methods for linking how these cells work electrically and their genetic types still work with new data. They tested different ways to predict cell types, including a new model that looks closely at specific features and can learn from mouse data to improve predictions in human data. Their work shows that learning from mouse brains can help understand human brain cells better. Overall, their approach confirms past results and offers new tools for studying brain cell diversity.
single-cell electrophysiologytranscriptomicsGABAergic interneuronsPatch-seqsparse PCArandom forestBiLSTMattention mechanismcross-species transfer learningmouse visual cortex
Authors
Theo Schwider, Ramin Ramezani
Abstract
Single-cell electrophysiological recordings provide a powerful window into neuronal functional diversity and offer an interpretable route for linking intrinsic physiology to transcriptomic identity. Here, we replicate and extend the electrophysiology-to-transcriptomics framework introduced by Gouwens et al. (2020) using publicly available Allen Institute Patch-seq datasets from both mouse and human cortex. We focus on GABAergic inhibitory interneurons to target a subclass structure (Lamp5, Pvalb, Sst, Vip) that is comparable and conserved across species. After quality control, we analyzed 3,699 mouse visual cortex neurons and 506 human neocortical neurons from neurosurgical resections. Using standardized electrophysiological features and sparse PCA, we reproduced the major class-level separations reported in the original mouse study. For supervised prediction, a class-balanced random forest provided a strong feature-engineered baseline in mouse data and a reduced but still informative baseline in human data. We then developed an attention-based BiLSTM that operates directly on the structured IPFX feature-family representation, avoiding sPCA and providing feature-family-level interpretability via learned attention weights. Finally, we evaluated a cross-species transfer setting in which the sequence model is pretrained on mouse data and fine-tuned on human data for an aligned 4-class task, improving human macro-F1 relative to a human-only training baseline. Together, these results confirm reproducibility of the Gouwens pipeline in mouse data, demonstrate that sequence models can match feature-engineered baselines, and show that mouse-to-human transfer learning can provide measurable gains for human subclass prediction.