Intel Arc Pro B70 Review

(pugetsystems.com)

54 points | by zdw 4 days ago

8 comments

  • speedgoose 53 minutes ago
    Time to first token is a very important performance metric, as I figured out using a Mac Studio M3 Ultra (that is quite slow on this aspect).

    But 32GB for a TDP of 230W is perhaps not super interesting. Especially because you probably want to have more than one card. It's a lot of heat. You could use the cards for heating up a building, but heatpumps exist.

    • bigyabai 39 minutes ago
      A lot of the TDP is reserved for running the shader units at full-power. My RTX 3070 Ti only pulls ~110w of it's 320w running CUDA inference on Gemma 26b and E4B.
      • Scaevolus 32 minutes ago
        It's not that it's reserving power, but rather that you hit some bottleneck on a 3070 Ti before running into thermal limits-- it's likely limited by either tensor core saturation or RAM throughput. Running the workload with Nvidia's profiling tools should make the bottleneck obvious.
        • lambda 5 minutes ago
          Generally the bottleneck is RAM throughput. Inference, in particular token generation, especially on a single user instance, is not all that computationally complex; you're doing some fairly simple calculations for each parameter, the time is dominated by just transferring each parameter from RAM to the cores. A 31B dense model like Gemma 4 has to transfer 31B parameters (at 16 bits per parameter for the full model, though on consumer hardware people generally run 4-8 bit quantizations) from RAM to the cores, that's a lot of memory transfer.

          Prompt processing or parallel token generation can do a bit more work per memory transfer, as you can use the same weights for a few different calculations in parallel. But even still, memory bandwidth is a huge factor.

  • cubefox 3 minutes ago
    Why are they still using their old Xe2/Battlemage architecture rather than their new Xe3/Celestial? They already used it in their Panther Lake chip set.
  • tempest_ 1 hour ago
    I would like one for the vram but I am sure they will be unobtainable after the initial stock sells out as I assume they were produced before the RAM prices went up.
  • driverdan 41 minutes ago
    From what I've read the Intel drivers are terrible and holding back using them for LLMs.
    • marshray 4 minutes ago
      I don't know about LLMs, but I tried an Intel card when Ubuntu Wayland couldn't initialize a 2 year old Nvidia. It just works.
    • martinald 36 minutes ago
      Don't think that's true. The drivers are bad (not sure terrible is fair, they have improved a lot) esp for older directx etc games. But Vulkan support is pretty good and that's all you need for LLMs really.
    • 999900000999 37 minutes ago
      Everyone has terrible drivers here aside from Nvidia.

      Intel looks like they'll leave the dedicated GPU space, so it's a bit doubtful if the drivers will ever catch up.

  • MostlyStable 55 minutes ago
    Is Intel still making GPUs? I have heard so many conflicting things about will they/won't they stay in the market.
    • numpad0 10 minutes ago
      Intel always had that habit of starting an internal conflict whenever whatever potential alternative revenue sources start to threaten their internal dependence on x86
    • 2OEH8eoCRo0 32 minutes ago
      I don't know what to believe when it comes to Intel news because they have so many haters.
    • dismalaf 39 minutes ago
      They'll always have iGPUs so whether or not they stay in the dGPU market depends mostly on whether or not people buy them. So they might not, whole market seems to be moving to SoCs/APUs/whatever you want to call them.
  • SparkyMcUnicorn 45 minutes ago
    Here are some llama.cpp benchmarks for it: https://www.phoronix.com/review/intel-arc-pro-b70-linux/3
  • XCSme 57 minutes ago
    Can you use those AI cards for gaming too?

    Or the makers intentionally nerf them, in order to better segment the markets/product lines?

    • ZiiS 45 minutes ago
      The drivers often need per game optimisations these will be missing but I doubt Intel would nerf them, just rely on you not paying a lot for RAM the game won't use.
      • XCSme 26 minutes ago
        I actually meant it in a different way. I would get it for local AI stuff, but being able to game on it would be a huge plus, otherwise I would need two different machines.
    • wmf 17 minutes ago
      They nerf gaming cards to make money on the pro cards. Since this is a pro card it's not nerfed.
  • 100ms 1 hour ago
    These seem amazing for hobbyist, but that TDP given the perf might be an issue deploying a lot of them
    • zrm 1 hour ago
      Its performance is pretty unbalanced. If you're using it for the couple of things that it's good at, the TDP is competitive.