Discussion about this post

User's avatar
Luis Herrero's avatar

I like this article. Your point is clearly exposed, and the illustration with the addition example makes it easy to keep in mind. Thanks.

Expand full comment
Ilia Kurgansky's avatar

The interpretability question is an interesting one, actually. We like our machines to be interpretable so that we can tweak and fix them, but we never pose this question about the humans in our teams. Or, rather, we assume that humans are self-aware enough to know their own reasoning.

Plenty of research to suggest that a lot of human reasoning is post hoc - first you just "know" what your favourite ice-cream is, and then you build scaffolding to justify why you think you said it was vanilla. Quite an uncomfortable reality. I can't recommend Blindsight enough, by Peter Watts.

There might come a time where we have to make do with uninterpretable systems - we already do in most ML applications, to be honest.

I still think LLMs are largely pointless though, but what is interesting is that the biggest fans of LLMs and magical thinking will ignore this research, just like most people do not internalise the research pointing at their own lack of self-consciousness and agency - an uncomfortable cognitive dissonance. "But it clearly knows how to do algebra correctly, though!"

Expand full comment
19 more comments...

No posts