• Teanut@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    10 months ago

    nVidia’s new Digits workstation, while expensive from a consumer standpoint, should be a great tool for local inferencing research. $3000 for 128GB isn’t a crazy amount for a university or other researcher to spend, especially when you look at the price of the 5090.

    • brucethemoose@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      10 months ago

      Dense models that would fit in 100-ish GB like mistral large would be really slow on that box, and there isn’t a SOTA MoE for that size yet.

      So, unless you need tons of batching/parallel requests, its… kinda neither here nor there?

      As someone else said, the calculus changes with cheaper Strix Halo boxes (assuming those mini PCs are under $3K).