Skip to main content

Spring Deadline: Sunday, March 1 @ 11:59pm PT. Click here to apply.

Universal Neurons in GPT-2: Emergence, Persistence, and Functional Impact

Universal Neurons in GPT-2: Emergence, Persistence, and Functional Impact

December 1, 2025

We investigate the phenomenon of neuron universality in independently trained GPT-2 Small models, examining how these universal neurons—neurons with consistently correlated activations across models—e...

Accepted to Interplay @ COLM 2025

Authors: Advey Nandan, Tim Chou, Amrit Lalith

We investigate the phenomenon of neuron universality in independently trained GPT-2 Small models, examining how these universal neurons—neurons with consistently correlated activations across models—emerge and evolve throughout training. By analyzing five GPT-2 models at three checkpoints (100k, 200k, 300k steps), we identify universal neurons through pairwise correlation analysis of activations over a dataset of 5 million tokens. Universal neurons emerge early, increasing consistently through training, notably in deeper layers. Universal neurons are highly stable over time, especially in later layers. Ablating universal neurons significantly increases loss and KL divergence, confirming their causal importance to model predictions. Layer-wise ablation reveals that ablating universal neurons in the first layer causes a disproportionately large increase in both KL divergence and loss.

Begin Your Journey

The application takes 10 minutes and is reviewed on a rolling basis. We look for strong technical signal—projects, coursework, or competition results—and a genuine curiosity to do real research.

If admitted, you will join a structured pipeline with direct mentorship to take your work from ideation to top conference submission at venues like NeurIPS, ACL, and EMNLP.