• Ace! _SL/S
    link
    fedilink
    English
    55
    edit-2
    9 months ago

    This comment sums it up pretty nicely:

    LOL innovative invention of swapping memory to storage…… maybe they can call it something cool like “cache”.

    Apple being “innovative” my ass, lmao

    • @[email protected]
      link
      fedilink
      English
      129 months ago

      Well, if that commenter had more than just a vague idea of caching and/or swapping, they would know that the right algorithm can make or break performance.

      That paper is not “we invented caching”, but “this is how we make some certain models work well despite constraints imposed by RAM and flash storage.”

      It’s a worthy job for an engineer or researcher. Not quite as innovative as the invention of the wheel, but still enough to write a paper on (and read it, if you can manage to understand it).

    • Hegar
      link
      fedilink
      109 months ago

      The easiest way to tell that something’s not really innovative is if the person describing it uses the word innovative.

    • @[email protected]
      link
      fedilink
      English
      -29 months ago

      35 upvotes in the technology community…man you guys really are just all knee-jerk reactionaries and it really knowledgeable tech at all. git gud

  • @[email protected]
    link
    fedilink
    English
    26
    edit-2
    9 months ago

    Everyone likes to trash machine learning because the power requirements are high, but what they don’t realize is that we’re in the very first days of this technology (well, first couple decades of the technology being around, first few years of it being advanced enough to have anything to show off). Every technology that got bundled together into your phone was equally as useless when it was first invented. Honestly, compared to the development of most other technologies I’ve looked at, the pace of development in AI has been shocking.

    Literally once a week, I see some news story about AI researchers delivering an order of magnitude speedup in some aspect of AI inference. The technique described here apparently allows for a 20x speedup on GPU’s.

    • @[email protected]
      link
      fedilink
      English
      109 months ago

      Whispercpp works off the ML cores on the m series chips. It’s faster than my 1080ti that I have in a server doing the same things–by orders of magnitude. And it sips power.

      Purpose built chips can be super powerful for their specific purposes.

  • @[email protected]
    link
    fedilink
    English
    39 months ago

    Tbh I’m more excited to see someone do use webnn, webgpu and petals together. Building smaller tighter models is good too.

  • @[email protected]
    link
    fedilink
    English
    19 months ago

    I don’t understand the innovation, I already run LLMs and stable diffusion on a laptop from 2011.

    I have no doubt it could be run on my Android phone.

  • @[email protected]
    link
    fedilink
    English
    -139 months ago

    Why the hell do we want to encourage people running MLMS on our phones?!! I don’t want to be part of some stupid pyramid scheme nonsense.