• HelloRoot@lemy.lol
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    5 days ago

    Seems pretty decent, but I wonder how it compares to an AI optimized desktop build with the same budget of 2000$.

    • mapumbaa@lemmy.zipOP
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      5 days ago

      It will probably kick the ass of that desktop. $2000 won’t get you far with a conventional build.

      • HelloRoot@lemy.lol
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        5 days ago

        Well, thats what I said “AI optimized”.

        Even my 5 year old 900$ rig can output like 4 tps.

        • mapumbaa@lemmy.zipOP
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          1
          ·
          4 days ago

          There is nothing “optimized” that will get you better inference performance of medium/large models at $2000.

          • HelloRoot@lemy.lol
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            4 days ago

            LLama 3 8B Instruct: 25tps

            DeepSeek R1 distill qwen 14b: 3.2tps

            To be fair: Motherboard, cpu and ram I bought 6 years ago with an nvidia 1660. Then I bought the Radeon RX 6600 XT on release in 2021, so 4 years ago. But it’s a generic gaming rig.

            I would be surprised if 2000$ worth of modern hardware, picked for this specific task would be worse than that mini PC.

            • mapumbaa@lemmy.zipOP
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              1
              ·
              edit-2
              4 days ago

              I promise. It’s not possible. But things change quickly of course.

              (Unless you’re lucky/pro and get your hands on some super cheap used high end hardware…)