• Kogasa
    link
    fedilink
    English
    arrow-up
    1
    ·
    22 hours ago

    The GPT architecture is well understood, the part that is hard to explain is the way information is encoded in the trained model’s parameters. It’s not magic, it’s just a highly opaque encoding.