12 Comments
Aug 9Liked by Sean Trott

Very nice discussion. The LLMs as crowds metaphor ties to the joke:

An LLM walks into a bar. "What'll you have?" asks the bartender. The LLM looks around and replies "what is everyone else having?"

My current stance is the alien metaphor. Specifically I imagine an electrically complex gas cloud on Jupiter that after decades of listening to radio and TV begins generating and transmitting new episodes of "I Love Lucy", "All in the Family", etc. I accept LLM behavior as robust and flexible enough to count as intelligence, though a kind far different than our own.

Expand full comment
author

I hadn't heard that before, I like that joke!

And yes, I think that's a fair perspective. The challenge moving forward (in my view) will be testing the limits of that robustness and flexibility and coming up with a theory of LLM "cognition" that's satisfying and accurate.

Expand full comment

Absolutely agree.

Expand full comment
Aug 9Liked by Sean Trott

Maybe this falls somewhere between copies and crowds in your categorization, but Alison Gopnik's framing of LLMs as a "cultural technology", i.e. like the Internet, libraries, printed text that all serve to enhance cognitive capacity and aide knowledge transmission, is an interesting metaphor as well!

Her recent talk: https://www.youtube.com/watch?v=qoCl_OuyaDw

Expand full comment
author

This looks great, thanks! I didn't realize she'd given a talk on this topic—I'm a big fan of her work, so I'm looking forward to watching it.

I actually initially had another section I was going to include called "LLMs as tools" but I had a harder time finding good examples. I'm personally very partial to the idea that LLMs can be well-understood as a kind of cultural technology (much like language itself). Interesting that she mentions libraries specifically—I'm working on another post exploring a thought experiment in which LLMs are construed as big libraries.

Expand full comment

I can't remember if we've talked about "cognitive gadgets theory" as postulated by Celia Heyes? Tools of the mind!

https://www.educationnext.org/cognitive-gadgets-theory-might-change-your-mind-literally/

Expand full comment

I'm pretty partial to the "LLMs as tools" framing, but maybe I'm just biased because the only purpose I've personally found them useful for is for coding, like GitHub Copilot. So far it feels like the only substance underneath all the hype are really good (carefully-prompted) autocompletion tools.

Cool, looking forward to that post! :)

Expand full comment

Interestingly enough, I've talked to Murray Shanahan about writing something around what Richard Rorty would say about LLMs today. One idea in my mind: Rorty described metaphor as the key tool we use to shift away from a current vocabulary to a new one. Your essay here nicely explores what metaphors we use for our mental model of LLMs, but I wonder, will we ever see an LLM invoke new metaphors to shift its own vocabulary? (I have my doubts!)

Expand full comment
author

I'm also somewhat doubtful, though it'd be fascinating to see!

Expand full comment

Sigh...

I confess that I do not have exhaustive or even very extensive knowledge of current LLMs, but I am led to believe that all are linear mathematical interpolation and extrapolation systems. As always in mathematics one does well to study the properties of the general class of systems, after which it becomes very easy to fully understand they properties of restricted sub-classes, such as LLMs. Analogies can serve only to cloud or distort real understanding.

It is well known that linear mathematical interpolation and extrapolation processes are approximated only to quite a limited extent in the cognition of mammals, including humans. (Too little is known, as yet, of non-mammalian cognition to make confident generalizations.) Thus analogies between LLMs and human cognition are of very limited validity at best, and likely to significantly mislead.

Expand full comment
author

Yes, that's fair, and I think it'd be great if there were more theoretical/math work characterizing LLMs. I'm certainly not the first person to note this, but the ML field sometimes feels quite empirical and a little disconnected from the more theoretical perspective.

Expand full comment

It is somewhat odd. I've lost touch with most of the AIers I once knew, but I believe it is still true that many do have pretty good backgrounds in mathematics.

About 60 years ago, as I recall, I was eating with Richard Bellman and Robert Kalaba before Dick was to present a paper. We fell to talking about a mutual acquaintance who had earlier presented a paper featuring wildly exaggerated claims for AI. Dick, who had a famously sharp tongue, quipped that too many of the enthusiasts for artificial intelligence seemed to suffer a deficit of natural intelligence.

There was a little too much truth to it for comfort, but I think we would all have agreed that the real problem was (and remains) less one of deficient natural intelligence but of a failure to think like mathematicians regarding what is fundamentally a mathematic issue.

Expand full comment