• Kogasa
    link
    fedilink
    English
    arrow-up
    1
    ·
    il y a 18 heures

    The GPT architecture is well understood, the part that is hard to explain is the way information is encoded in the trained model’s parameters. It’s not magic, it’s just a highly opaque encoding.