Discussion about this post

User's avatar
Callum Hackett's avatar

Nice analysis. I completely agree with the principle that we should favour causal descriptions whenever possible. A big issue is that the possibility for this varies depending on who "we" are, and the biggest audience for LLMs is never going to have the technical nous to wield even basic causal descriptions (never mind degenerates like Hinton who should know better).

In some ways, we've already lived with these conditions for a long time. There are various computational artefacts more complex than thermostats where lay people rely on intentional descriptions - if a piece of deterministic software unexpectedly deletes a batch of emails for some opaque reason after a user interaction, a typical user will say that the system "misunderstood" what they were trying to do (implying false belief). That someone from IT could come along and understand the mechanism doesn't mean that the user should forego their own intentional description (how else would they communicate the issue with the IT guy? To some degree, we need stance translation).

The picture is more complicated with LLMs just because there's a decades-long background of anthropomorphising language in AI that has been designed to mislead. We never had to talk about these systems as 'learning', 'intelligent' or 'agentic' - we could have described them as data compressions, program libraries, cultural microcosms etc. - and perhaps there's still time to salvage an alternative vocabulary.

But while I think we could get comfortable with intentional description regardless, it's only going to be safe if we deconstruct widespread misunderstandings about intentional *autonomy*. When we say a thermostat wants a certain temperature, we don't confuse ourselves about the fact that the thermostat is parasitic on a setting encoded by a human. When we say that a piece of deterministic software "misunderstands" our intentions, we use a communicative metaphor because we realise that the issue lies in the *interaction* of the system and the user, not in the system itself.

LLMs have no autonomy and no appropriate intentional description separate from their prompting by humans, and we therefore ought to be able to think and talk about them in the same way as other complex systems. Our present difficulty is that there's enormous commercial pressure for people to suppress recognition of LLMs' parasitism on humans, so we're awash with lies about their potential independence from us.

There are many ways this might be addressed but I think a good start would be to get honest technical people thinking a bit more loosely about the boundaries of computational systems, as has been developed with the Systems Reply to the Chinese Room and the Extended Mind Thesis etc. If people intuitively grok that there isn't a clear boundary between them and the AI model, assumptions about autonomy lose a lot of their bite and intentional description becomes more favourable.

15 more comments...

No posts

Ready for more?