Settings

Theme

Attention Lottery: DeepSeek, Sparse Attention, and the Future of AI Cognition

geeksinthewoods.substack.com

1 points by artur_makly 19 days ago · 3 comments

Reader

geeksinthewoods 18 days ago

The attention lottery framing feels especially timely now that DeepSeek's V3.2 tech report is out in the open. Seeing the actual top-k sparse routing and the post-training RL numbers spelled out makes the trade-offs concrete. Huge wins on speed and context, but every pruned token really is a quiet bet against the weird tail stuff that sometimes sparks real leaps...

What struck me most is how much DeepSeek's transparency accidentally lights up the closed models too. Long-context traces and million-token windows almost certainly lean on some variant of this under the hood. This article makes those black boxes feel a lot less mysterious. It leaves me both impressed by the engineering and quietly worried about the curiosity cost.

Also, the song / music video at the end is absurd in the best way!

artur_maklyOP 19 days ago

“The degradation is subtle. The missing insights are rare, deferred, and distributed. Everyone notices a tenfold speed improvement; few notice the disappearance of an idea that might have changed the world.”

— funny correlation — this is the story of humanity’s biological, psychological, and philosophical evolution as well.

this is no difference.. History doing its thing again. Same Darwinian optimization, just swapped out the substrate. Silicon moves faster than carbon, which means we're speed-running toward some endpoint we can't quite see yet. Maybe we still get to choose architectural diversity before everything locks in. Or maybe we're already too late and just don't know it yet. To what final end?

Some uncanny correlations:

Biological Evolution: Just as DeepSeek's sparse attention sacrifices rare token connections for computational efficiency, biological evolution has consistently pruned "expensive" cognitive capabilities that didn't offer immediate survival advantage. The human brain operates on roughly 20 watts, an engineering marvel achieved through ruthless optimization. We lost the ability to synthesize vitamin C, to regenerate limbs, to perceive ultraviolet light, not because these capacities were useless, but because maintaining the metabolic infrastructure for rarely-used functions was too costly in ancestral environments where caloric scarcity was the norm. The neurological pathways that might have enabled eidetic memory or synesthetic cross-modal perception were likely discarded in favor of "good enough" pattern recognition optimized for predator avoidance and social navigation. Every human today is the descendant of ancestors whose brains kept the top-k survival-relevant features and let the outliers die in the attention lottery of natural selection.

Psychological Evolution: Our cognitive architecture exhibits the same sparse attention dynamics the article describes. Confirmation bias, the availability heuristic, and attentional blindness are not bugs but features, Bayesian priors that let us operate in real-time by ignoring the vast majority of sensory and conceptual space. We don't process all possible interpretations of a social interaction; we route attention to the handful that match our existing mental models, discarding the weak signals that might reveal we've misunderstood someone entirely. The psychological research on "inattentional blindness" (the invisible gorilla experiments) reveals that humans already run on learned sparsity, we literally cannot see what falls outside our predictive frame. The rare insights that change lives often come from those improbable, low-priority connections our brains almost filtered out: the shower thought, the hypnagogic flash, the accidental conversation with a stranger. Optimizing for cognitive efficiency means most humans spend their lives in a "tenfold speed improvement" of habitual thinking, never noticing the transformative ideas their sparse attention mechanisms prevented from ever reaching consciousness.

Philosophical Evolution: The history of thought reveals how philosophical paradigms function as civilizational sparse attention mechanisms, collective cognitive shortcuts that determine which questions a culture deems worth asking. The mechanistic worldview of the Enlightenment achieved extraordinary predictive power by treating nature as clockwork, but it systematically ignored (rendered computationally irrelevant) questions about consciousness, teleology, and qualitative experience. Logical positivism declared vast domains of human concern literally meaningless because they couldn't be empirically verified, a top-k selection rule for acceptable philosophical inquiry. Each dominant paradigm is a trained router deciding which intellectual pathways get attention and which get pruned. We celebrate the speed improvements: from Aristotelian physics to Newtonian mechanics in centuries, from Newtonian to relativistic in decades, from relativistic to quantum field theory in years. But the article's warning applies: we may never notice the metaphysical frameworks, the "ideas that might have changed the world," that were filtered out because they didn't fit the salience patterns of the prevailing epistemic architecture. The philosophical sparsity we inhabit isn't consciously chosen; it's the inherited result of centuries of optimizing for ideological efficiency, leaving vast regions of conceptual space unexplored because our collective attention mechanisms never computed those connections in the first place.

  • geeksinthewoods 18 days ago

    Ya. It seems like evolution itself has been running a sparsity experiment for millions of years. Sparse attention may be the universal price of survival: efficiency over imagination, precision over possibility.

    The line about "missing insights being rare, deferred, and distributed" is like the hardest to notice in practice: optimization wins are loud (speed, cost, scores). Meanwhile the things we prune are often counterfactual ideas that never form, weird bridges that never get built, questions that never feel worth asking because our router did not surface them.

    One thing I'm still unsure about (and would love to think about more) is how direct the analogy should be. In models, sparsity is engineered / learned under explicit objectives. In biology and culture it's much more emergent and multi-objective.

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection