• XLE@piefed.social
    link
    fedilink
    English
    arrow-up
    6
    ·
    1 day ago

    There’s a bit more to it: Obviously, if a model gets more correct data pumped into it, it’s more likely to produce a correct output. But they found that at the core of every AI model they tested, when an incorrect output came along, certain nodes produced it. And they are some of the nodes at the earliest part of making the model - before data gets added.

    So with that in mind, the tl;dr is more like

    AI models have two goals: first be readable, then be correct. It appears the nodes causing incorrect outputs that are also intended to make the output readable.