What is an "Abstract Reasoner"? Revisiting Experiments and Arguments about Large Language Models

Tian Yun
Brown University
Chen Sun
Brown University
Ellie Pavlick
Brown University


Abstract

Recent work has argued that large language models (LLMs) are not "abstract reasoners", citing their poor zero-shot performance on a variety of challenging tasks as evidence. We revisit these experiments in order to add nuance to the claim. First, we show that while LLMs indeed perform poorly in a zero-shot setting, even tuning a small subset of parameters for input encoding can enable near-perfect performance. However, we also show that this finetuning does not necessarily transfer across datasets. We take this collection of empirical results as an invitation to (re-)open the discussion of what it means to be an "abstract reasoner", and why it matters whether LLMs fit the bill.


Language Models for Abstract Textual/Visual Reasoning Task (ACRE)

Illustration of the use of language models for text-based and image-based versions of ACRE.

Illustration of the use of language models for text-based and image-based versions of ACRE. Each data example will be formulated into a prompt for an LLM to make a prediction for the query. In textual reasoning task, each context frame is represented by a frame caption. In visual reasoning tasks, each context frame is represented by an encoded frame representation.


Overview of Experimental Settings

Illustration of our experimental settings.

In Setting (a), we freeze the whole LLM and run evaluations. This is treated as language baseline when image captions are inputs on abstract visual reasoning tasks. In Settings (b) and (c), we freeze the pretrained transformer blocks and finetune only the input layers (i.e., token embedding layer and visual encoder). In Setting (c), we freeze the token embedding layer to study the impact of tuning the visual encoder in a controlled setting.

Note that the inputs are pure language in Settings (a) and (b), while the inputs are language prompts with image representations in Setting (c).


Results

1. Frozen Pretrained LLMs Achieve Low Performance Across A Large Suite of Abstract Reasoning Tasks.


We first seek to replicate Gendron et al's finding that frozen pretrained LLMs acheive low performance across a large suite of reasoning tasks. We observe that even though there are small gaps between the original results and the reproduced results, the performance of the pretrained LLMs are still low, indicating that their poor transfer ability to abstract reasoning tasks. What requires additional investigation, however, is whether this poor transfer is interpretable as a lack of abstract reasoning ability.

2. On Abstract Reasoning Tasks, LLaMA2 With A Finetuned Token Embedding Layer Performs Significantly Better Than Its Pretrained Counterpart, AND Can Perform On Par With the LoRA-finetuned LLaMA2.


Given that pretrained LLMs perform poorly off-the-shelf, it is natural to ask whether they can be adapted to these task, and if so, just how much adaptation is necessary. We explore two ways to finetune the LLMs: (1) finetuning all layers with low-rank adaptation (LoRA); (2) finetunning only the embedding layer of the LLMs. LoRA finetuning has become a standard way of adapting a model to a task and represents an upper bound on how well the model could be made to perform the task under the most permissive conditions. In contrast, finetuning just the embedding layer represents a conceptually different type of transfer with respect to the question of this paper. Namely, finetuning just the embeddings is analogous to changing just the input to the system--e.g., ensuring the input is in the format the system expects--but leaving the system itself unchanged.

We observe that LoRA-finetuned models perform significantly better than their pretrained counterparts, and can even solve ACRET and RAVENT perfectly. Moreover, LLaMA2 with a finetuned embedding layer can perform on par with the LoRA-finetuned LLaMA2.

3. The Pattern Holds in Abstract Visual Reasoning Task: LLaMA2 with Train-from-scratch Visual Encoder Can Perform Significantly Better Than Their Language-only Counterpart.


Given the surprising effectiveness of finetuning just the embedding layer of LLaMA2 on text-only abstract reasoning tasks, we hypothesize that the frozen transformer blocks of a pretrained LLM will perform well on abstract visual reasoning tasks if the visual encoder is tuned for the task. That is, we follow the multimodal LLM framework (MLLM) which consists of a visual backbone, a language backbone, and a linear projection layer which maps visual representations to language latent space. We keep the transformer blocks and the token embedding layer of language backbone frozen, and only train the visual encoder and the projection layer. If this MLLM with a trained visual encoder can perform better than its language backbone with oracle visual perception, then it provides further evidence for the above interpretation of the frozen LLM as a highly transferable system.

On ACRE, we observe that LLaMA2 with train-from-scratch visual encoders can perform significantly better than their language-only counterparts, and can even outperform majority of the multimodal state-of-the-art, including IV-CL and LRR, which are pretrained with video data. On MEWL, we observe the same pattern that LLaMA2 with learned visual encoders can outperform prior state-of-the-art and also the language baselines which assume perfect visual perception.


Discussion: What is an Abstract Reasoner?

The question of whether LLMs are abstract reasoners has consequences for how we understand and thus how we develop increasingly advanced artificial intelligence. The challenge is that there is no consensus for what it means to be an abstract reasoner. In their recent work, Gendron et al operationalize abstract reasoning as the ability to transfer zero-shot to a range of complex reasoning tasks. They find that LLMs perform poorly on this evaluation, and thus conclude that they are not abstract reasoners.

In this work, we reproduce Gendron et al's findings, but push back against their interpretation. In particular, we provide new experiments which show that tuning just the embedding layer is remarkably effective. Indeed, across a variety of textual and multimodal tasks, frozen pretrained LLMs can achieve high levels of performance as long as the input representations are adapted sufficiently for each task.

It seems too stringent a criteria to require that that abstract reasoners perform arbitrary tasks on arbitrary inputs without adaptation. By way of counterargument, consider the good old fashioned AI (GOFAI) systems of the 1990s, which typically included symbolic systems internally, e.g., databases implemented in SQL or rules for logical inference implemented in PROLOG. By most intuitive definitions, these databases and rules would be considered "abstract" and the tasks the systems performed over them would be "reasoning". But we would not expect these systems to operate well over a database implemented in MongoDB, or to apply rules defined by Python. Rather, the need to operate on representations of a particular format is a consequence of, not an exception to, the system's abstraction.

Of course, we don't claim that the internal processing of an LLM is exactly analogous to that of a GOFAI system. Of course, in an LLM, tuning the input embedding layer might do more than simply "rerepresent", but rather might encode some task-specific processing as well. But interpreted loosely, the analogy is useful for highlighting how the question of adaptability and transferability relates to the question of abstraction and reasoning.

Indeed, this relationship has been considered in depth by philosophers of AI, long before LLMs. For example, Dennett appeals to transferability in his attempt to describe the difference between human cognition and simpler computational systems:
Consider the lowly thermostat...we might agree to grant it the capacity for about half a dozen different beliefs...it can believe the room is too cold or too hot, that the boiler is on or off...and so forth...suppose we de-interpret its beliefs and desires, it can believe the A is too F or G...and so forth....by attaching the thermostatic control mechanism to different input and output devices, it could be made to regulate the amount of water in a tank, or the speed of a train for instance...But as systems become perceptually richer and behaviorally more versatile, it becomes harder and harder to make substitutions in the actual links of the system to the world without changing the organization of the system itself. ...There comes to be a two-way constraint of growing specificity between the device and the environment. Fix the device in any one state and it demands a very specific environment in which to operate properly (you can no longer switch it easily from regulating temperature to regulating speed or anything else); but at the same time, if you do not fix the state it is in, but just plunk it down in a changed environment, its sensory attachments will be sensitive and discriminative enough to respond appropriately to the change...
Although Dennett is not discussing the notion of "abstract reasoners" per se, he observes that intelligent systems do not transfer well unless they are allowed to adapt. Indeed, Dennett argues that this is a defining property, one that differentiates human-like intelligence from simpler (albeit perhaps more abstract) systems such as thermostats.

Dennett's argument is relevant here not because LLMs are human-like or even human-level in their reasoning abilities (they are far from it!). Rather, Dennett articulates a position that is implicit in contemporary discussions about LLMs and "abstract reasoning". That is, that we care about how well a system adapts to new environments because adapting well to new environments is a hallmark of intelligence. Indeed, this is often cited explicitly as the motivation for studies of this nature (e.g., "the question of whether or not LLMs can perform human-like reasoning remains open..." (Gendron et al. 2024)). But if evaluating human-likeness or human-levelness is the motivation for studying abstract reasoning, then arguments such as Dennett's provide a compelling case against using zero-shot transfer ability as a relevant metric.

Of course, there is another, more practical, argument for why we might care about whether LLMs are abstract reasoners, which is simply that we want LLMs to transfer well zero-shot to many tasks in order to facilitate easier, cheaper, and more efficient development of systems. Indeed, the thermostat's highly abstract design is a feature, not a bug. This type of hardware abstraction is what allows similar components and control mechanisms to be readily repurposed to support many types of use cases. A "human like" thermostat might be very undesirable.

Thus, before seeking to answer the question of whether LLMs are "abstract reasoners", we must first determine, as a community, why we care. Do we care because we want to understand how human-like they are, or do we care because we want to facilitate more efficient technological progress? Almost certainly, we care about both, but we should not expect the same experiments to bear on both lines of inquiry. Finding clarity around these questions - what is an abstract reasoner and why do we care about building one? - is the essential next step if we are to make progress toward either, or both, goals.


Paper and BibTex

Tian Yun, Chen Sun, Ellie Pavlick.
What is an “Abstract Reasoner”? Revisiting Experiments and Arguments about Large Language Models
CONLL 2025.




  @inproceedings{yun2025what,
	title={What is an "Abstract Reasoner"? Revisiting Experiments and Arguments about Large Language Models
	},
	author={Tian Yun and Chen Sun and Ellie Pavlick},
	booktitle={The 29th Conference on Computational Natural Language Learning},
	year={2025}
  }
					


Acknowledgements

We would like to thank all reviewers and the area chair for their valuable feedback. We would like to thank Apoorv Khandelwal for the help to present this work at CONLL and the feedback on the presentation. We would also like to thank Samuel Musker, Calvin Luo, and other members of the SuperLab at Brown University for their discussions and insights. The project depicted is sponsored in part by a Young Faculty Award from the Defense Advanced Research Projects Agency, Grant #D24AP00261. The content of the information does not necessarily reflect the position, or the policy of the government and no official endorsement of this work should be inferred.

This template was originally made by Phillip Isola and Richard Zhang for a colorful ECCV project; the code can be found here.