The classical objection reformulated and representation as a condition of all cognition
Part 1/3 of the article “Encoding, transduction, and world models”. Part 2/3 (chapters 3-5) is published. Part 3/3 (chapters 6-7) is in preparation. Original article in French.
The recurring critique that AI systems are “trapped in language” rests on a confusion between language and representation. Dreyfus (Heideggerian phenomenology, practical know-how vs symbolic manipulation) and Searle (Chinese Room, syntax vs semantics) identify a genuine tension: the distance between syntactic processing and lived semantics. The limitation of LLMs does not stem from confinement within language but from the joint absence of direct sensorimotor grounding, multimodal grounding founded on lived co-constitution, and a biographically organised episodic memory.
All intelligence — biological or artificial — implies a prior transformation of the world into manipulable internal states. The article distinguishes three forms of encoding: (1) the explicit encoder, a dedicated architectural module (Transformer, VAE, CLIP) that defines the informational geometry of the latent space; (2) implicit encoding, distributed across network layers (CNN, LSTM, self-attention), an emergent property of optimisation; (3) pipeline encoding, the set of decisions made by the data scientist upstream of the model — variable selection, transformation, incompleteness handling — which defines the model’s perception space. This third layer is epistemologically the most loaded: the model has no access to dimensions of the world that have not been encoded.