>What neuron(s) is responsible for capitalization in GPT?
It doesn't matter. Individual things like capitalization are vague and useless for interpretability. We know that incorrect capitalization will increase loss, so the model will need to figure how to do it correctly.
>Our understanding of what the neurons do is very limited.
The mathematical definition is right in the code. You can see the calculations they are doing.
>this bunch of weights are a redundancy because this set already achieves this function that this other set does and so can be pruned. Let's also tune this set so it never tries to call this other set while we're at it.
They are equivalent. If removing something does not increase loss then it was redundant behavior at least for the dataset that it is being tested against.
>It doesn't matter. Individual things like capitalization are vague and useless for interpretability. We know that incorrect capitalization will increase loss, so the model will need to figure how to do it correctly.
It matters for the point I was making. Capitalization is a simple example. There are far vague functions we'd certainly like the answers to.
>They are equivalent. If removing something does not increase loss then it was redundant behavior at least for the dataset that it is being tested against.
The level of understanding for both is not equivalent sorry.
At this point, you're just rambling on about something that has nothing to do with the point I was making. Good Day
It doesn't matter. Individual things like capitalization are vague and useless for interpretability. We know that incorrect capitalization will increase loss, so the model will need to figure how to do it correctly.
>Our understanding of what the neurons do is very limited.
The mathematical definition is right in the code. You can see the calculations they are doing.
>this bunch of weights are a redundancy because this set already achieves this function that this other set does and so can be pruned. Let's also tune this set so it never tries to call this other set while we're at it.
They are equivalent. If removing something does not increase loss then it was redundant behavior at least for the dataset that it is being tested against.