Massive activations are identified as unexpectedly large activation values within Large Language Models (LLMs), which often stand orders of magnitude above others (for instance, 100,000x larger). Interestingly, these massive activations are relatively input-invariant, yet play a pivotal role as crucial bias terms that are integral to the functionality of LLMs. The study delves deeper to understand the impact of these activations on attention mechanisms, thereby influencing attention probabilities and introducing implicit bias terms within the self-attention output.
The presence of massive activations across various models suggests a fundamental aspect of deep neural networks yet to be completely understood. Going beyond mere characterization, these findings urge the scientific community to investigate the origins and implications of such disparities in activation values. Such understanding could lead to more robust and balanced network designs, with potential applications in interpretability studies and anomaly detection. Learn more…