From Tokens to Thoughts: How LLMs and Humans Trade Compression for Meaning
Overview
Overall Novelty Assessment
The paper applies the Information Bottleneck principle to quantify compression-meaning trade-offs in LLM versus human conceptual representations, analyzing embeddings from over 40 models against human categorization benchmarks. It resides in the 'Information Bottleneck Principle in Semantic Categorization' leaf, which contains only three papers total. This is a relatively sparse research direction within the broader taxonomy of 21 papers across multiple branches, suggesting the work occupies a focused niche at the intersection of formal information theory and empirical LLM-human comparison.
The taxonomy reveals neighboring work in adjacent leaves: 'Entropy-Based Conceptual Importance Quantification' examines entropy measures in structured representations like AMR graphs, while 'Information-Theoretic Brain-LLM Alignment' uses compression metrics to align neural and model representations. The sibling papers in the same leaf—Evolution Compression Categorization and Color Categories Compression—apply information-theoretic frameworks to human cognitive evolution and perceptual domains respectively. The current work appears to bridge these by directly comparing LLM and human compression strategies across semantic categorization tasks, diverging from purely human-focused or purely neural alignment studies.
Among 28 candidates examined across three contributions, no clearly refuting prior work was identified. The information-theoretic framework contribution examined 10 candidates with zero refutations, the digitized benchmark contribution examined 8 with zero refutations, and the empirical findings on divergent optimization strategies examined 10 with zero refutations. This suggests that within the limited search scope, the specific combination of Information Bottleneck analysis applied to multi-model LLM-human comparison on classic categorization benchmarks appears relatively unexplored, though the analysis does not claim exhaustive coverage of the literature.
Based on the limited search of 28 semantically related candidates, the work appears to occupy a novel position combining formal information-theoretic analysis with large-scale empirical LLM-human comparison. The sparse population of its taxonomy leaf and absence of refuting candidates within the examined scope suggest distinctiveness, though the analysis acknowledges it cannot rule out relevant work outside the top-K semantic matches or citation network examined.
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors develop a unified framework based on Rate-Distortion Theory and Information Bottleneck principles to systematically measure and compare how LLMs and humans balance compression efficiency against semantic meaning preservation in conceptual organization. The framework introduces an L objective that combines information-theoretic complexity with geometric distortion to evaluate representational strategies.
The authors digitize and publicly release classic human categorization datasets from foundational cognitive science studies, comprising 1,049 items across 34 categories with membership and typicality ratings. These benchmarks provide high-quality empirical grounding for evaluating whether LLMs understand concepts as humans do.
Through analysis of 40+ LLMs against human benchmarks, the authors reveal that LLMs achieve superior information-theoretic efficiency but miss fine-grained semantic distinctions crucial for human understanding. Encoder models surprisingly outperform decoder models in human alignment, and training dynamics show conceptual structure develops through rapid formation followed by architectural reorganization.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
Contribution Analysis
Detailed comparisons for each claimed contribution
Information-theoretic framework for comparing LLM and human conceptual representations
The authors develop a unified framework based on Rate-Distortion Theory and Information Bottleneck principles to systematically measure and compare how LLMs and humans balance compression efficiency against semantic meaning preservation in conceptual organization. The framework introduces an L objective that combines information-theoretic complexity with geometric distortion to evaluate representational strategies.
[22] Concept Bottleneck Models PDF
[23] Concept bottleneck generative models PDF
[24] Agility to Handle Dynamics of Business Transformation PDF
[25] Towards human-agent communication via the information bottleneck principle PDF
[26] Efficient human-like semantic representations via the Information Bottleneck principle PDF
[27] Intervening in Black Box: Concept Bottleneck Model for Enhancing Human Neural Network Mutual Understanding PDF
[28] Toward human-like object naming in artificial neural systems PDF
[29] Concept-Based Explainable AI: Interpreting Deep Learning Models through Human-Readable Concepts in Financial Applications PDF
[30] Transferring Expert Cognitive Models to Social Robots via Agentic Concept Bottleneck Models PDF
[31] Conceptual Content in Deep PDF
Digitized cognitive psychology benchmarks for evaluating conceptual alignment
The authors digitize and publicly release classic human categorization datasets from foundational cognitive science studies, comprising 1,049 items across 34 categories with membership and typicality ratings. These benchmarks provide high-quality empirical grounding for evaluating whether LLMs understand concepts as humans do.
[32] Aligning machine and human visual representations across abstraction levels PDF
[33] Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM Diversity PDF
[34] Assessing AI-Generated Questions' Alignment with Cognitive Frameworks in Educational Assessment PDF
[35] Evaluating alignment between humans and neural network representations in image-based learning tasks PDF
[36] Evaluating (and improving) the correspondence between deep neural networks and human representations PDF
[37] From Pixels to Feelings: Aligning MLLMs with Human Cognitive Perception of Images PDF
[38] The Flexibility of Similarity Perception and Its Implications for Representationally Aligned Artificial Intelligence PDF
[39] Alignability-based free categorization. PDF
Empirical findings on divergent optimization strategies between LLMs and humans
Through analysis of 40+ LLMs against human benchmarks, the authors reveal that LLMs achieve superior information-theoretic efficiency but miss fine-grained semantic distinctions crucial for human understanding. Encoder models surprisingly outperform decoder models in human alignment, and training dynamics show conceptual structure develops through rapid formation followed by architectural reorganization.