The overloading of common words is well underway: new language models have “thinking” modes, “reasoning” capabilities! What this means, in practice, is that they’ve learned to produce a special kind of text, the conversion of the linguistic if-then into a dynamo that spins and spins and, often, magically —
Here is one distinction among several: this process can only compound —
Human thinking often washes the dishes, then goes for a walk.
So, if you redefine “thinking” to mean “arriving at a solution through an iterative linguistic loop” … yes, that’s what these models do. That definition is IMHO pretty thin.
We talk about humans thinking harder, which is not the same as thinking longer. I think most people know from experience that thinking longer generally just makes you anxious. But that’s what the models do, and not only longer, but in parallel, all those step-by-step monologues spilling out simultaneously, somewhere in the dark of a data center. “Quantity has a quality all its own,” said Stalin, maybe …
Well, okay —
A couple of years ago, when I wondered if language models are in hell, I expressed some hope about the richness of multimodal training. So far, this hasn’t panned out. Rather than images anchoring text in a richer, more embodied realm, the marriage seems to have gone the opposite direction. The models chop images into sequences of tokens —
We are going to lose this word —
Other useful words, still with us, include: imagination, ingenuity, insight. Clarity, most of all. Clarity is what Einstein was seeking when he sat and thought hard about the relative motion of magnets and conductors. He wanted to push through language, beyond it, beyond even the formalism of physics —
I am still waiting for models that aspire to pack complex systems —
Think harder!
To the blog home page