xentr_theme_editor

  • Please do not post any links until you have 3 posts as they will automatically be rejected to prevent SPAM. Many words are also blocked due to being used in SPAM Messages. Thanks!

Intel B60 and B50 Arc Pro GPU

OK, I went and looked up DeepLink and it's apparently a dead feature. So how is Maxsun going to support the B60 Pro dual GPU then?

I guess Maxsun is betting on the adage a 'sucker is born every minute' ;)
 
Interesting to see Intel leaning hard into workstation AI with these—makes me wonder how well Battlematrix could scale with something like Stable Diffusion XL if software support matures. Anyone seen any real-world latency or inference benchmarks yet?
 
xentr_thread_starter
Interesting to see Intel leaning hard into workstation AI with these—makes me wonder how well Battlematrix could scale with something like Stable Diffusion XL if software support matures. Anyone seen any real-world latency or inference benchmarks yet?
No one has any real benchmark just like Strix Halo 128GB. As far as I know, none of the AI image generators support multiple GPU correctly. Other AI application like Ollama could use multiple GPU or at least spread the workload across the VRAM of multiple cards.
 
I kind a want like 3x ARC b50 pro cards and no where to be found yet. (2 for the office, 1 personal)

I was keeping an eye out as well, but haven't seen an option to buy them anywhere. I've seen reports that some regions overseas have seen minimal stock, but nothing that is available to consumers.
 
xentr_thread_starter
Someone asked Maxsun when they could order the B60 and they mentioned in a private WhatApps exchange that it will be ready for ordering on Aug 18th (the guy is in Singapore) for $1200 USD before tax and shipping.

 
xentr_thread_starter
More pricing news. GameMachines, a German retailer is listing the MaxSun model at 1500 Euro. Hydratechbuilds (US distro) is asking $3K USD. Single B60 is $1K.

 
More pricing news. GameMachines, a German retailer is listing the MaxSun model at 1500 Euro. Hydratechbuilds (US distro) is asking $3K USD. Single B60 is $1K.

with that much memory, these are targeted at AI workloads? how do the mainstream Arc GPU with less RAM do for AI workloads, in general?
 
xentr_thread_starter
Yep, AI inference workload but probably slower for AI training (haven't looked into it myself). According to eBay (all CAD):
  • A used RTX 8000 48GB (TU102 so 2nd gen card) is $3K CAD
  • modded RTX 4090D 48GB is $4.7K CAD (same kind of card in GN video)
  • Nvidia L20 48GB is $6.2K
  • Tesla A40 48GB is $10K
  • L40S 48GB is $12K
  • Tesla V100 32GB is $1.1K (Nvidia stop supporting it just recently, Volta is 1.5 gen)
  • Tesla M10 32GB is $200 (this is Maxwell so GTX 900 era)
That's why I was kicking myself for not seeing another user here selling their RTX 3090 for $800. The usual price in Vancouver is $900 to $1200. And you get 24GB, high bandwidth, and support SAGE Attention code to speed up Stable Diffusion and WAN video generation.

From my understanding, the memory is really there to load the bigger models. The memory bandwidth plays a role in the inference speed or tokens per second. So it could be slower than a RTX 3090 due to slower bandwidth but it won't be capped when you try to load a really big model. Once you run out of VRAM, it have to do memory swap between the GPU and system RAM which slows it down. Bigger model tends to give you better quality answers so having a card with 2x24GB helps but since data have to go thru the PCI-E bus (literally 2 GPU on 1 board), it may not be fast. At the end of the day, we need to see benchmarks. From my experience with Stable Diffusion, this card doesn't work because none of the apps support multiple GPU in a way that parallel the workload. You may be able to generate multiple images faster, but this card won't speed up generating a single image.

Not like a regular B770 won't work for local AI. You can still install Ollama or Stable Diffusion but you have to very careful in what kind of model you load. For Stable Diffusion, you end up picking a quantified version of the original model that fits in VRAM but gives you worst picture or close enough but slower. Full fat FLUX model and WAN video generation is like 20GB and that's not counting other stuff like text encoders that you need so it can translate prompts into image generation. Oh also, all the speed up tricks like SAGE Attention requires CUDA. Not sure if you can get away with ZLUDA (AMD implementation of CUDA) but Intel for sure doesn't have similar thing.

The 5070Ti Super looks very promising because it is rumored to have 24GB of memory and running the latest Nvidia tech. Right now 5070Ti is $1.2K CAD so if SUPER is only slightly more expensive, then the Intel B60 is moot.
 
xentr_thread_starter
Price for the B60 has been spotted. $600 USD from Central Computer but the listing is now pulled.

 
Back
Top