• Kogasa
    link
    English
    -34 months ago

    It must have some internal models of some things, or else it wouldn’t be possible to consistently make coherent and mostly reasonable statements. But the fact that it has a reasonable model of things like grammar and conversation doesn’t imply that it has a good model of literally anything else, which is unlike a human for whom a basic set of cognitive skills is presumably transferable. Still, the success of LLMs in their actual language-modeling objective is a promising indication that it’s feasible for a ML model to learn complex abstractions.

    • @[email protected]
      link
      fedilink
      English
      264 months ago

      if I copy a coherent sentence into my clipboard, my clipboard becomes capable of consistently making coherent statements

      • Kogasa
        link
        English
        -4
        edit-2
        4 months ago

        Yes, but that’s not how LLMs work. My statement depends heavily on the fact that a LLM like GPT is coaxed into coherence by unsupervised or semi-supervised training. That the training process works is the evidence of an internal model (of language/related concepts), not just the fact that something outputs coherent statements.

        • @[email protected]
          link
          fedilink
          English
          14
          edit-2
          4 months ago

          if I have a bot pick a random book and copy the first sentence into my clipboard, my clipboard becomes capable of consistently making coherent statements. unsupervised training 👍

          • Kogasa
            link
            English
            24 months ago

            Did you forget to actually ban me? I dunno why you were going to, or why you think I don’t know how LLMs work, but that’s your business.

            • @Tja
              link
              English
              14 months ago

              I don’t think they know how lemmy works, let alone LLMs xD

        • adderaline
          link
          fedilink
          English
          114 months ago

          this isn’t necessarily true. patterns in data aren’t by nature proof of an underlying system of logic. if you run the line-fitting machine on any kind of data, its going to output a line. considering just how much data is encoded into these transformers, i don’t think we can conclusively say that it has a underlying conception of how language works, much less an understanding of the concepts that language represents. it could really just be using the vast quantities of data it has to output approximately correct statements. there’s absolutely structure there, but it doesn’t have to have the kind of structured understanding humans have about language to produce language, in the same way a less sophisticated machine learning model doesn’t have to know what kind of data its fitting a line to to make a line.

    • flere-imsaho
      link
      fedilink
      English
      164 months ago

      it doesn’t. that’s why we’re calling it “spicy autocompletion” .

      • Kogasa
        link
        English
        -74 months ago

        It does, which is why it’s autocompletion and not auto-gibberish.

    • @[email protected]
      link
      fedilink
      English
      154 months ago

      It must have some internal models of some things, or else it wouldn’t be possible to consistently make coherent and mostly reasonable statements.

      Talk about begging the question