• MudMan@fedia.io
      link
      fedilink
      arrow-up
      5
      arrow-down
      1
      ·
      12 hours ago

      The idea is having tensor acceleration built into SoCs for portable devices so they can run models locally on laptops, tablets and phones.

      Because, you know, server-side ML model calculations are expensive, so offloading compute to the client makes them cheaper.

      But this gen can’t really run anything useful locally so far, as far as I can tell. Most of the demos during the ramp-up to these were thoroughly underwhelming and nowhere near what you get from server-side services.

      Of course they could have just called the “NPU” a new GPU feature and make it work closer to how this is run on dedicated GPUs, but I suppose somebody thought that branding this as a separate device was more marketable.

      • This is fine🔥🐶☕🔥@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        arrow-down
        1
        ·
        11 hours ago

        EU should introduce regulation that prohibits client-side AI/ML processing for applications that require internet access. Show the cost upfront. Let’s see how many people pay for that.

        • MudMan@fedia.io
          link
          fedilink
          arrow-up
          2
          arrow-down
          1
          ·
          9 hours ago

          That is a weird proposal.

          It’s definitely weird that everyone is panicking about data center processing costs but not about the exact same hardware powering high end gaming devices that have skyrocketed from 100W to 450W in a few years, but ultimately if you want to run a model locally you can run a model locally. I’m not sure how you’d regulate that, it’s just software.

          Hell, I don’t even think distributing the load is a terrible idea, it’s just that the models you can run locally in 40 TOPS kinda suck compared to the order of magnitude more processing you get on modern GPUs.