Emergence of symbolic abstraction heads for in-context learning in large language models

Ali Al-Saeedi, Aki Harma

Research output: Chapter in Book/Report/Conference proceedingConference article in proceedingAcademicpeer-review

Abstract

Large Language Models (LLMs) based on self-attention circuits are able to perform, at inference time, novel reasoning tasks, but the mechanisms inside the models are currently not fully understood. We assume that LLMs are able to generalize abstract patterns from the input and form an internal symbolic internal representation of the content. In this paper, we study this by analyzing the performance of small LLM models trained with sequences of instantiations of abstract sequential symbolic patterns or templates. It is shown that even a model with two layers is able to learn an abstract template and use it to generate correct output representing the pattern. This can be seen as a form of symbolic inference taking place inside the network. In this paper, we call the emergent mechanism abstraction head. Identifying mechanisms of symbolic reasoning in a neural network can help to find new ways to merge symbolic and neural processing.
Original languageUndefined/Unknown
Title of host publicationProceedings of Bridging Neurons and Symbols for Natural Language Processing and Knowledge Graphs Reasoning @ COLING 2025
EditorsKang Liu, Yangqiu Song, Zhen Han, Rafet Sifa, Shizhu He, Yunfei Long
Place of PublicationAbu Dhabi, UAE
PublisherELRA and ICCL
Pages86-96
Number of pages11
Publication statusPublished - 2025

Cite this