• circuscritic
    link
    fedilink
    34
    edit-2
    1 month ago

    Awesome. Truly spectacular.

    Generative AI is so energy intensive ($$$), that Google is requiring users subscribe to Gemini.

    Google is entirely dependent on advertising sales. Ad revenue subsidizes literally everything else, from Android development to whichever 8-12 products and services they launch and subsequently cancel each year.

    Now, Google wants to remove web results and just use generative AI instead of search as it’s default user interface.

    So, like I said: Awesome.

    • @[email protected]
      link
      fedilink
      12
      edit-2
      1 month ago

      While I agree in principle, one thing I’d like to clarify is that TRAINING is super energy intensive, once the network is trained, it’s more or less static. Actually using the network isn’t dramatically more energy than any other indexed database lookup.

      • @[email protected]
        link
        fedilink
        221 month ago

        It’s static, yes, but the static price is orders of magnitude higher. It still involves loading the whole model into VRAM and performing matrix multiplication on trillions of numbers

        • @[email protected]
          link
          fedilink
          English
          51 month ago

          To be fair, I wouldn’t include “loading the whole model into VRAM” as part of the cost, given they can just keep it in there between different requests, and it might be down to hundreds of billions or dozens of billions instead of trillions… but even after all improvements it should still be orders of magnitude more expensive than normal search, which just makes their decision even crazier

        • @[email protected]
          link
          fedilink
          21 month ago

          Indexing and lookups on datasets as big as companies like Google and Amazon are running also take trillions of operations to complete, especially when you take into account the constant reindexing that needs to be done. In some cases, encoding data into a neural network is actually cheaper than storing the data itself. You can see this in practice with gaussian splatting point cloud capture, where they are training networks to guide points in the cloud at runtime, rather than storing the position of trillions of points over time.

      • @towerful
        link
        101 month ago

        Training will never stop, tho.
        New models will keep coming out, datasets and parameters are going to change.

        • @[email protected]
          link
          fedilink
          21 month ago

          I firmly believe it will slow down significantly. My prediction for the future is that there will be a much bigger focus on a few “base” models that will be tweaked slightly for different roles, rather than “from the ground up” retraining like we see now. The industry is already starting to move in that direction.