• Nomecks@lemmy.ca
    link
    fedilink
    arrow-up
    2
    ·
    edit-2
    2 months ago

    Intel sees the AI market as the way forward. NVIDIA’s AI business eclipses its graphics business by an order of magnitude now, and Intel wants in. They know that they rule the integrated graphics market, and can leverage that position to drive growth with things like edge processing for CoPilot.

    • brucethemoose@lemmy.world
      link
      fedilink
      arrow-up
      5
      arrow-down
      1
      ·
      edit-2
      2 months ago

      The localllama crowd is supremely unimpressed with Intel, not just because of software issues but because they just don’t have beefy enough designs, like Apple does, and AMD will soon enough. Even the latest chips are simply not fast enough for a “smart” model, and the A770 doesn’t have enough VRAM to be worth the trouble.

      They made some good contributions to runtimes, but seeing how they fired a bunch of engineers, I’m not sure that will continue.

      • Nomecks@lemmy.ca
        link
        fedilink
        arrow-up
        2
        arrow-down
        1
        ·
        edit-2
        2 months ago

        People running LLMs aren’t the target. People who use things like ChatGPT and CoPilot on low power PCs who may benefit from edge inference acceleration are. Every major LLM dreams of offloading compute on the end users. It saves them tons of money.

        • brucethemoose@lemmy.world
          link
          fedilink
          arrow-up
          3
          ·
          edit-2
          2 months ago

          One can’t offload “usable” LLMs without tons of memory bandwidth and plenty of RAM. It’s just not physically possible.

          You can run small models like Phi pretty quick, but I don’t think people will be satisfied with that for copilot, even as basic autocomplete.

          About 2x faster than Intel’s current IGPs is the threshold where the offloading can happen, IMO. And that’s exactly what AMD/Apple are producing.