How might LLMs store facts | DL7

https://www.youtube.com/watch?v=9-Jl0dxWQs8 AI Alignment forum post from the Deepmind researchers referenced at the video’s start:https://www.alignmentforum.org/posts/… Anthropic posts about superposition referenced near the end:https://transformer-circuits.pub/2022…https://transformer-circuits.pub/2023… Some added resources for those interested in learning more about mechanistic interpretability, offered by Neel Nanda Mechanistic interpretability paper reading listhttps://www.alignmentforum.org/posts/… Getting started in mechanistic interpretabilityhttps://www.neelnanda.io/mechanistic-… An interactive demo of sparse autoencoders (made…

Transformers (how LLMs work) explained visually | DL5

If you’re interested in the herculean task of interpreting what these large networks might actually be doing, the Transformer Circuits posts by Anthropic are great. In particular, it was only after reading one of these that I started thinking of the combination of the value and output matrices as being a combined low-rank map from…

Backpropagation calculus | DL4

This one is a bit more symbol-heavy, and that’s actually the point. The goal here is to represent in somewhat more formal terms the intuition for how backpropagation works in part 3 of the series, hopefully providing some connection between that video and other texts/code that you come across later. For more on backpropagation:http://neuralnetworksanddeeplearning….https://github.com/mnielsen/neural-ne…http://colah.github.io/posts/2015-08-… https://colah.github.io/posts/2015-08-Backprop

Gradient descent, how neural networks learn | DL2

To learn more, I highly recommend the book by Michael Nielsenhttp://neuralnetworksanddeeplearning….The book walks through the code behind the example in these videos, which you can find here:https://github.com/mnielsen/neural-ne… MNIST database:http://yann.lecun.com/exdb/mnist/ Also check out Chris Olah’s blog:http://colah.github.io/His post on Neural networks and topology is particular beautiful, but honestly all of the stuff there is great. And if…