CMU Research: AI Intelligence May Emerge From Data Compression, Not Just Big Datasets

CMU Research Challenges AI Pre-training Paradigm with Compression-Based Intelligence
New research from Carnegie Mellon University (CMU) suggests that artificial intelligence might achieve problem-solving capabilities through information compression alone, potentially bypassing the need for extensive pre-training on massive datasets. This groundbreaking work challenges the conventional wisdom in AI development, which heavily relies on large datasets and computationally intensive models.
The Core Idea: Compression as Intelligence
Researchers Isaac Liao and Professor Albert Gu propose that lossless information compression can be a driving force behind intelligent behavior. Their system, CompressARC, tackles abstract reasoning tasks by finding the most efficient way to represent information, effectively learning the underlying patterns and rules of a given problem without prior training.
Testing on ARC-AGI
To validate their hypothesis, Liao and Gu tested CompressARC on the Abstraction and Reasoning Corpus (ARC-AGI), a benchmark designed to assess AI systems' abstract reasoning skills. ARC-AGI presents visual puzzles where AI must infer rules from examples to solve new instances.
- The Challenge: ARC-AGI puzzles involve grid-based images requiring the AI to understand concepts like object persistence, goal-directed behavior, counting, and basic geometry.
- Human Performance: The average human solves 76.2 percent of ARC-AGI puzzles, with experts reaching 98.5 percent.
- OpenAI's o3: OpenAI's o3 model achieved 75.7 percent in computational limits and 87.5 percent with unlimited time on ARC-AGI.
- CompressARC's Results: CompressARC achieved 34.75 percent accuracy on the training set and 20 percent on the evaluation set. Notably, each puzzle takes approximately 20 minutes on a consumer GPU, a stark contrast to the massive computational resources used by other leading methods.
A Novel Approach to AI
CompressARC deviates significantly from typical AI methodologies:
- No Pre-training: The system is randomly initialized and trains in real-time using only the specific puzzle it needs to solve. It requires no external training data.
- No Search: Unlike systems that explore numerous potential solutions, CompressARC relies solely on gradient descent to incrementally adjust parameters and minimize errors.
- Compression as Inference: The core principle is using compression to find the shortest description of a puzzle that accurately reproduces the examples and the solution when unpacked.
- Custom Architecture: While borrowing structural elements from transformers, CompressARC is a custom neural network designed for compression, not an LLM or standard transformer.
- Decoder-Focused: The neural network acts primarily as a decoder. During encoding, the system fine-tunes network settings and data to create the most compressed representation.
Theoretical Foundations
The connection between compression and intelligence is rooted in computer science concepts:
- Kolmogorov Complexity: The shortest program that produces a given output.
- Solomonoff Induction: A theoretical model for prediction equivalent to optimal compression.
These concepts suggest that efficient compression requires pattern recognition, regularity identification, and understanding underlying data structures—hallmarks of intelligence.
Previous Research and Implications
This work builds on previous findings, such as a 2023 DeepMind paper showing that large language models could outperform specialized compression algorithms. While DeepMind demonstrated compression in trained models, Liao and Gu's research suggests that the compression process itself can generate intelligence from scratch.
This research challenges the AI industry's trend towards larger models and more extensive datasets. CompressARC offers a potential alternative path, demonstrating that intelligence might emerge from efficient information representation rather than sheer scale.
Limitations and Future Directions
CompressARC's current limitations include struggles with tasks requiring counting, long-range pattern recognition, rotations, reflections, or simulating agent behavior. The research has not yet undergone peer review, and its 20 percent accuracy on unseen puzzles, while promising for a pre-training-free approach, is lower than human and top AI system performance.
Critics may argue that CompressARC's success might be tied to specific puzzle structures and may not generalize broadly. However, if validated, this research could offer a more resource-efficient path to AI development and unlock a crucial component of general intelligence.
Image Credits:
- Illustration: Vibrant Colored Futuristic Cubes (Eugene Mymrin via Getty Images)
- Three example ARC-AGI benchmarking puzzles (Isaac Liao / Albert Gu)
- An animated GIF showing the multi-step process of CompressARC solving an ARC-AGI puzzle (Isaac Liao)
- Photo of a C-clamp compressing books (Getty Images)
Original article available at: https://arstechnica.com/ai/2025/03/compression-conjures-apparent-intelligence-in-new-puzzle-solving-ai-approach/