• takeda@lemm.ee
    link
    fedilink
    English
    arrow-up
    138
    arrow-down
    5
    ·
    2 days ago

    trust but verify

    The thing is that LLM is a professional bullshitter. It is actually trained to produce text that can fool ordinary person into thinking that it was produced by a human. The facts come 2nd.

    • conditional_soup@lemm.ee
      link
      fedilink
      English
      arrow-up
      53
      arrow-down
      6
      ·
      2 days ago

      Yeah, I know. I use it for work in tech. If I encounter a novel (to me) problem and I don’t even know where to start with how to attack the problem, the LLM can sometimes save me hours of googling by just describing my problem to it in a chat format, describing what I want to do, and asking if there’s a commonly accepted approach or library for handling it. Sure, it sometimes hallucinate a library, but that’s why I go and verify and read the docs myself instead of just blindly copying and pasting.

      • lefaucet@slrpnk.net
        link
        fedilink
        English
        arrow-up
        33
        arrow-down
        1
        ·
        edit-2
        2 days ago

        That last step of verifying is often being skipped and is getting HARDER to do

        The hallucinations spread like wildfire on the internet. Doesn’t matter what’s true; just what gets clicks that encourages more apparent “citations”. Another even worse fertilizer of false citations is the desire to push false narratives by power-hungry bastards

        AI rabbit holes are getting too deep to verify. It really is important to keep digital hallucinations out of the academic loop, especially for things with life-and-death consequences like medical school

        • medgremlin@midwest.social
          link
          fedilink
          English
          arrow-up
          5
          ·
          2 days ago

          This is why I just use google to look for the NIH article I want, or I go straight to DynaMed or UpToDate. (The NIH does have a search function, but it’s terrible meaning it’s just easier to use google to find the link to the article I actually want.)

          • Detun3d@lemm.ee
            link
            fedilink
            English
            arrow-up
            2
            ·
            1 day ago

            I’ll just add that I’ve had absolutely no benefit, just time wasted, when using the most popular services such as ChatGPT, Gemini and Copilot. Yes, sometimes it gets a few things right, mostly things that are REALLY easy and quick to find even when using a more limited search engine such as Mojeek. Most of the time these services will either spit out blatant lies or outdated info. That’s one side of the issue with these services, and I won’t even get into misinformation injected by their companies. The other main issue I find for research is that you can’t get a broader, let alone precise picture about anything without searching for information yourself, filtering the sources yourself and learning and building better criteria yourself, through trial and error. Oftentimes it’s good info that you weren’t initially searching for what makes your time well spent and it’s always better to have 10 people contrast information they’ve gathered from websites and libraries based on their preferences and concerns than 10 people doing the same thing with information they were served by an AI with minimal input and even less oversight. Better to train a light LLM model (or setup any other kind of automation that performs even better) with custom parameters at your home or office to do very specific tasks that are truly useful, reliable and time saving than trusting and feeding sloppy machines from sloppy companies.

    • Zexks@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      23 hours ago

      So are people. Rule NUMBER 1 when the internet was first picking up is “Don’t believe everything you read on the internet”. it’s like all of you have forgotten. So many want to bitch so hard about Ai while completely ignoring the environment it was raised in and the PEOPLE who trained it. You know, all of us. This is a human issue not an AI issue.

    • Honytawk@feddit.nl
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      1
      ·
      24 hours ago

      So use things like perplexity.ai, which adds links to the web page where they got the information from right next to the information.

      So you can check yourself after an LLM made a bullshit summary.

      Trust but verify

    • Impleader@lemmy.world
      link
      fedilink
      English
      arrow-up
      25
      arrow-down
      2
      ·
      2 days ago

      I don’t trust LLMs for anything based on facts or complex reasoning. I’m a lawyer and any time I try asking an LLM a legal question, I get an answer ranging from “technically wrong/incomplete, but I can see how you got there” to “absolute fabrication.”

      I actually think the best current use for LLMs is for itinerary planning and organizing thoughts. They’re pretty good at creating coherent, logical schedules based on sets of simple criteria as well as making communications more succinct (although still not perfect).

      • takeda@lemm.ee
        link
        fedilink
        English
        arrow-up
        6
        ·
        1 day ago

        Sadly, the best use case for LLM is to pretend to be a human on social media and influence their opinion.

        Musk accidentally showed that’s what they are actually using AI for, by having Grok inject disinformation about South Africa.

      • Honytawk@feddit.nl
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        23 hours ago

        Can you try again using an LLM search engine like perplexity.ai?

        Then just click on the link next to the information so you can validate where they got that info from?

        LLMs aren’t to be trusted, but that was never the point of them.

      • sneekee_snek_17@lemmy.world
        link
        fedilink
        English
        arrow-up
        5
        arrow-down
        1
        ·
        2 days ago

        The only substantial uses i have for it are occasional blurbs of R code for charts, rewording a sentence, or finding a precise word when I can’t think of it

        • NielsBohron@lemmy.world
          link
          fedilink
          English
          arrow-up
          5
          arrow-down
          3
          ·
          edit-2
          2 days ago

          It’s decent at summarizing large blocks of text and pretty good for rewording things in a diplomatic/safe way. I used it the other day for work when I had to write a “staff appreciation” blurb and I couldn’t come up with a reasonable way to take my 4 sentences of aggressively pro-union rhetoric and turn it into one sentence that comes off pro-union but not anti-capitalist (edit: it still needed a editing pass-through to put it in my own voice and add some details, but it definitely got me close to what I needed)

          • sneekee_snek_17@lemmy.world
            link
            fedilink
            English
            arrow-up
            6
            arrow-down
            1
            ·
            2 days ago

            I’d say it’s good at things you don’t need to be good

            For assignments I’m consciously half-assing, or readings i don’t have the time to thoroughly examine, sure, it’s perfect

            • NielsBohron@lemmy.world
              link
              fedilink
              English
              arrow-up
              4
              ·
              2 days ago

              exactly. For writing emails that will likely never be read by anyone in more than a cursory scan, for example. When I’m composing text, I can’t turn off my fixation on finding the perfect wording, even when I know intellectually that “good enough is good enough.” And “it’s not great, but it gets the message across” is about the only strength of ChatGPT at this point.

    • Ketchup@reddthat.com
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 day ago

      I have two friends that work in tech, and I keep trying to tell them this. And they use it solely now: it’s both their google, and their research tool. I admit, at first I found it useful, until it kept being wrong. Either it doesn’t know the better/best way to do something that is common knowledge to a 15 year tech, while confidently presenting mediocre or incorrect steps. Or it makes up steps, menus, or dialog boxes that have never existed, or are from another system.

      I only trust it for writing pattern tasks: example, take this stream of conscious writing and structure it by X. But for information. Unless I’m manually feeding it attachments to find patterns in my good data— no way.

    • ByteJunk@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      arrow-down
      2
      ·
      2 days ago

      To be fair, facts come second to many humans as well, so I dont know if you have much of a point there…

    • Apepollo11@lemmy.world
      link
      fedilink
      English
      arrow-up
      5
      arrow-down
      11
      ·
      edit-2
      2 days ago

      That’s true, but they’re also pretty good at verifying stuff as an independent task too.

      You can give them a “fact” and say “is this true, misleading or false” and it’ll do a good job. ChatGPT 4.0 in particular is excellent at this.

      Basically whenever I use it to generate anything factual, I then put the output back into a separate chat instance and ask it to verify each sentence (I ask it to put <span> tags around each sentence so the misleading and false ones are coloured orange and red).

      It’s a two-pass solution, but it makes it a lot more reliable.

      • TheTechnician27@lemmy.world
        link
        fedilink
        English
        arrow-up
        20
        arrow-down
        3
        ·
        2 days ago

        It’s a two-pass solution, but it makes it a lot more reliable.

        So your technique to “make it a lot more reliable” is to ask an LLM a question, then run the LLM’s answer through an equally unreliable LLM to “verify” the answer?

        We’re so doomed.

        • Apepollo11@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          8
          ·
          edit-2
          2 days ago

          Give it a try.

          The key is in the different prompts. I don’t think I should really have to explain this, but different prompts produce different results.

          Ask it to create something, it creates something.

          Ask it to check something, it checks something.

          Is it flawless? No. But it’s pretty reliable.

          It’s literally free to try it now, using ChatGPT.

          • TheTechnician27@lemmy.world
            link
            fedilink
            English
            arrow-up
            11
            arrow-down
            1
            ·
            2 days ago

            I don’t think I should really have to explain this, but different prompts produce different results.

            Ron Swanson saying "I know more thab you" to a home improvement store employee

            • Apepollo11@lemmy.world
              link
              fedilink
              English
              arrow-up
              2
              ·
              2 days ago

              Hey, maybe you do.

              But I’m not arguing anything contentious here. Everything I’ve said is easily testable and verifiable.