Discussion about this post

User's avatar
Mike X Cohen, PhD's avatar

Nice write-up, Sean. Interesting to see how different people from different backgrounds learn about LLM mechanisms. If I may be so audacious as to humbly suggest my 90+ hour course on LLM architecture, training, and mechanistic interpretability, using ML methods to investigate internal activations during inference: https://github.com/mikexcohen/LLM_course

Expand full comment
Pete's avatar

I never thought about LLMs as an object of observation. My assumption before reading your text was that anything engineered can be considered deterministic - thanks for opening a new avenue of contemplation for me!

Expand full comment
6 more comments...

No posts