1. Introduction: Why This Matters

In mainstream media and industry, large language models (LLMs) are routinely described as "intelligent" or even showing "sparks of AGI." But current LLMs are self-supervised sequence models: extremely capable, but optimised to predict the next token in text. They are not grounded in the physical world and do not possess the embodied, causal, socially situated learning that characterises human intelligence.

This article argues:

  1. Calling LLMs "intelligent" in the human sense is misleading.
  2. A credible path toward AGI requires embodied, multimodal, socially grounded, continual learning.

2. What LLMs Actually Do: Stochastic Parrots, Not Reasoners

Bender et al. (2021) describe LLMs as "stochastic parrots": systems that can generate fluent text via statistical correlations but without grounding, understanding, or agency.1 They warn that LLMs:

Multiple studies show that LLMs struggle with tasks requiring genuine reasoning or causal understanding, especially when problems deviate from training distributions.5,6


3. Scaling Is Hitting Technical and Economic Limits

Between 2020–2024, bigger models + more compute = better performance. Today, major constraints are clear:

  1. Flattening scaling laws — marginal gains require exponentially more compute.8
  2. Data exhaustion — high-quality text is nearly "mined out".
  3. Macroeconomic limits — The 2025 Storm report argues current AI valuations rest on unrealistic productivity assumptions.8

If scaling alone could yield AGI, we would not need elaborate scaffolds like chain-of-thought prompting, retrieval, tool use, memory systems, or Monte Carlo search. These exist precisely because text prediction alone cannot produce causal world understanding.


4. The Brain Is a Prediction Machine — But Predicting the World, Not Tokens

Predictive-processing theories (Friston 2010; Clark 2013) conceptualise the brain as a hierarchical prediction engine minimising error signals across sensory channels.2,3 But crucially, the brain predicts:

A 2024 meta-analysis by Costa et al. identifies a domain-general "Dynamic Prediction Network" across cognition.4 Humans are indeed prediction machines — just not token-prediction machines.


5. How LLM Concept Representations Diverge from Human Concepts

Xu et al. (2025) compared human conceptual representations with LLM embeddings for 4,442 concepts.9 They found:

This aligns with the symbol grounding problem: LLMs define symbols only via other symbols, not via embodied experience.10


6. Why Better Data Alone Won't Fix It

Two structural barriers remain:

  1. Embodiment: Human cognition arises from closed-loop interaction with the world. LLMs lack sensorimotor grounding.10
  2. Continual learning: Humans learn over decades. LLMs forget catastrophically without specialised techniques.7,11

Multi-agent RL shows hints of emergent communication and culture,12,13 but these systems remain brittle and narrow.


7. A More Plausible Route Toward AGI

  1. Embodied agents acting in real or simulated environments.14
  2. Multimodal grounding linking concepts to perception and action.10
  3. Social learning including teaching, imitation, and cultural transmission.12,13
  4. Robust continual learning without catastrophic forgetting.7,11
  5. Causal and uncertainty-aware reasoning building explicit models of the world.15

LLMs may remain important modules — but not standalone artificial minds.


8. So, Are LLMs "Intelligent"?

If "intelligent" means surface competence, LLMs qualify. If it means grounded understanding, causality, agency, and lifelong learning — they do not.

Calling LLMs "intelligent" obscures real limitations and risks.


9. References

← Back to Blog