Comment on LLMs’ “simulated reasoning” abilities are a “brittle mirage,” researchers find

<- View Parent
mindbleach@sh.itjust.works ⁨6⁩ ⁨hours⁩ ago

Please don’t mistake vindication for a lack of ambiguity. When this took off, we had no goddamn idea what the limit was. The fact it works half this well is still absurd.

Simple examples like addition were routinely wrong, but they were wrong in a way that indicated - the model might actually infer the rules of addition. That’s a compact way to predict a lot of arbitrary symbols. Seeing that abstraction emerge would be huge, even if it was limited to cases with a zillion examples. And it was basically impossible to reason about whether that was pessimistic or optimistic.

A consensus for “that doesn’t happen” required all of this scholarship. If we had not reached this point, the question would still be open. Remove all the hype from grifters insisting AGI is gonna happen now, oops I mean now, oops nnnow, and you’re still left with a series of advances previously thought impossible. Backpropagation doesn’t work… okay now it does. Training only plateaus… okay it gets better. Diffusion’s cute, avocado chairs and all, but… okay that’s photoreal video. It really took people asking weird questions on high-end models to distinguish actual reasoning capability from extremely similar sentence construction.

And if we’re there, can we please have models ask a question besides ‘what’s the next word?’

source
Sort:hotnewtop