“While the B60 is designed for powerful ‘Project Battlematrix’ AI workstations sold as full systems ranging from $5,000 to $10,000, it will carry a roughly $500 per-unit price tag.”

  • brucethemoose@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    4 days ago

    The better choice depends on which software stack you intend to wrangle with, how long you intend to keep the cards, and your usage patterns, but the B580 is the sorta newer silicon.

    Exllamav3 is the shit these days (as you can fully offload 32Bs in 16GB with very little loss), and it’s theoretically getting AMD support way before Intel (unless Intel changes that).

    …Also, 2x 3060s may be a better option, depending on price.

    • hendrik@palaver.p3x.de
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      4 days ago

      It’s difficult to make that decision. I mainly tinker with generative AI. Try a chatbot or agent or try creative writing with it, mess with FramePack or LTX-video, or explore some text to speech or whatever I find interesting when I got some time to spare.
      Obviously I wouldn’t spend a lot of money just to mess around. So yeah. I currently just rent cloud GPUs by the hour and I’m fine with that. Once we get a very affordable and nice AI card with lots if fast VRAM, I think I’m going to buy one. But I’m really not sure if this one or any previous generation Radeon is what I’m looking for. Or me spending quite some time on ebay to find some old 30x0. And as a Linux user I’m not really fond of the Nvidia drivers, so that also doesn’t make it easier.

      • brucethemoose@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        ·
        edit-2
        4 days ago

        Oh yeah, you will run into a ton of pain sampling random projects on AMD/Intel. Most “experiments” only work out of the box on Nvidia. Some can be fixed, some can’t.

        A used 3090 is like gold if you can find one, yeah.

        And yes, I sympathize with Nvidia being a pain on linux… though it’s not so bad if you just output from your IGP or another card.

        And yes, stuff rented from vast.ai or whatever is cheap. So are APIs. TBH thats probably the way to go if budget is a big concern, and a 24GB B60 is out of the cards.