Entry tags:
Let's understand Large Language Models better
This is a good starting point:
"A Mathematical Framework for Transformer Circuits", Dec 2021
transformer-circuits.pub/2021/framework/index.html
"A Mathematical Framework for Transformer Circuits", Dec 2021
transformer-circuits.pub/2021/framework/index.html
no subject
generally, I would not assume their explanations are complete, even for these small models
it makes better sense to think about their approach as a viewpoint, and not as "The Truth"
(and especially listening to his caveats near 1:57:00)
no subject
And another frequent motif is that these things are good with fixing the weirdness of tokenizers
2:01:00 and for more complicated models, it is useful to think that attention heads are doing a lot of skip trigrams and doing other things on top of that
no subject