Ok let’s give a little bit of context. I will turn 40 yo in a couple of months and I’m a c++ software developer for more than 18 years. I enjoy to code, I enjoy to write “good” code, readable and so.

However since a few months, I become really afraid of the future of the job I like with the progress of artificial intelligence. Very often I don’t sleep at night because of this.

I fear that my job, while not completely disappearing, become a very boring job consisting in debugging code generated automatically, or that the job disappear.

For now, I’m not using AI, I have a few colleagues that do it but I do not want to because one, it remove a part of the coding I like and two I have the feeling that using it is cutting the branch I’m sit on, if you see what I mean. I fear that in a near future, ppl not using it will be fired because seen by the management as less productive…

Am I the only one feeling this way? I have the feeling all tech people are enthusiastic about AI.

  • UmeU@lemmy.world
    link
    fedilink
    arrow-up
    38
    arrow-down
    6
    ·
    11 months ago

    AI is a really bad term for what we are all talking about. These sophisticated chatbots are just cool tools that make coding easier and faster, and for me, more enjoyable.

    What the calculator is to math, LLM’s are to coding, nothing more. Actual sci-fi style AI, like self aware code, would be scary if it was ever demonstrated to even be possible, which it has not.

    If you ever have a chance to use these programs to help you speed up writing code, you will see that they absolutely do not live up to the hype attributed to them. People shouting the end is nigh are seemingly exclusively people who don’t understand the technology.

    • anarchost@lemm.ee
      link
      fedilink
      arrow-up
      22
      ·
      11 months ago

      I’ve never had to double check the results of my calculator by redoing the problem manually, either.

    • Couldbealeotard@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      11 months ago

      Haven’t we started using AGI, or artificial general intelligence, as the term to describe the kind of AI you are referring to? That self aware intelligent software?

      Now AI just means reactive coding designed to mimic certain behaviours, or even self learning algorithms.

      • UmeU@lemmy.world
        link
        fedilink
        arrow-up
        1
        ·
        11 months ago

        That’s true, and language is constantly evolving for sure. I just feel like AI is a bit misleading because it’s such a loaded term.

        • Couldbealeotard@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          11 months ago

          I get what you mean, and I think a lot of laymen do have these unreasonable ideas about what LLMs are capable of, but as a counter point we have used the label AI to refer to very simple bits of code for decades eg video game characters.

    • Lmaydev
      link
      fedilink
      arrow-up
      6
      arrow-down
      4
      ·
      11 months ago

      AI is the correct term. It’s the name of the field of study and anything that mimics intelligence is an AI.

      Neural networks are a perfect example of an AI. What you actually code is very simple. A bunch of nodes that pass numbers forward through the system applying weights to the values. Their capabilities once trained far outstretch the simple code they run and seem intelligent.

      What you are referring to is general AI.

      • anarchost@lemm.ee
        link
        fedilink
        arrow-up
        6
        arrow-down
        2
        ·
        11 months ago

        It’s a misnomer, but if you want to pass off LLMs as “artificial intelligence” on technicality of definition, you’d also have to include

        advanced web search engines (e.g., Google Search), recommendation systems (used by YouTube, Amazon, and Netflix)

        etc.

        • Lmaydev
          link
          fedilink
          arrow-up
          3
          arrow-down
          1
          ·
          edit-2
          11 months ago

          Indeed you do.

          Neural networks are some of the original AIs.

        • BitSound@lemmy.world
          link
          fedilink
          arrow-up
          5
          arrow-down
          3
          ·
          11 months ago

          Yes those are also examples of AI, see relevant Wikipedia article:

          AI is whatever hasn’t been done yet

          We need better terms to specify exactly what we mean, e.g. a numeric scale of intelligence or maybe even something more complex like a radar chart.

    • evasive_chimpanzee@lemmy.world
      link
      fedilink
      arrow-up
      2
      ·
      11 months ago

      Yeah, this is the thing that always bothers me. Due to the very nature of them being large language models, they can generate convincing language. Also image “ai” can generate convincing images. Calling it AI is both a PR move for branding, and an attempt to conceal the fact that it’s all just regurgitating bits of stolen copywritten content.

      Everyone talks about AI “getting smarter”, but by the very nature of how these types of algorithms work, they can’t “get smarter”. Yes, you can make them work better, but they will still only be either interpolating or extrapolating from the training set.