• HelloRoot@lemy.lol
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    11 days ago

    Seems pretty decent, but I wonder how it compares to an AI optimized desktop build with the same budget of 2000$.

    • mapumbaa@lemmy.zipOP
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      11 days ago

      It will probably kick the ass of that desktop. $2000 won’t get you far with a conventional build.

      • humanspiral@lemmy.ca
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 days ago

        8700g with 256gb ram is possible on desktop. Half the APU performance, but less stupid bigger model > fast, for coding. No one seems to be using such a rig though.

      • HelloRoot@lemy.lol
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        11 days ago

        Well, thats what I said “AI optimized”.

        Even my 5 year old 900$ rig can output like 4 tps.

          • HelloRoot@lemy.lol
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            10 days ago

            LLama 3 8B Instruct: 25tps

            DeepSeek R1 distill qwen 14b: 3.2tps

            To be fair: Motherboard, cpu and ram I bought 6 years ago with an nvidia 1660. Then I bought the Radeon RX 6600 XT on release in 2021, so 4 years ago. But it’s a generic gaming rig.

            I would be surprised if 2000$ worth of modern hardware, picked for this specific task would be worse than that mini PC.

            • mapumbaa@lemmy.zipOP
              link
              fedilink
              English
              arrow-up
              2
              arrow-down
              1
              ·
              edit-2
              10 days ago

              I promise. It’s not possible. But things change quickly of course.

              (Unless you’re lucky/pro and get your hands on some super cheap used high end hardware…)

        • mapumbaa@lemmy.zipOP
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          1
          ·
          10 days ago

          There is nothing “optimized” that will get you better inference performance of medium/large models at $2000.