• spaduf@slrpnk.net
    link
    fedilink
    English
    arrow-up
    50
    ·
    2 days ago

    Honestly good for them. US tech CEOs deserve to have their lunch eaten for ducking the industry into stagnation with their short sighted greed.

  • wewbull@feddit.uk
    link
    fedilink
    English
    arrow-up
    20
    arrow-down
    3
    ·
    edit-2
    2 days ago

    In one story they’re using PTX on Nvidia H800s. In another they’re on Huawei chips.

    Which is it? Are we all just hypothesising?

  • taytay@lemmings.world
    link
    fedilink
    English
    arrow-up
    13
    arrow-down
    5
    ·
    2 days ago

    An unknown quantization of R1 is running on the 3rd iteration of outdated 7nm hardware taken from Sophgo’s work with TSMC last year?

    Is this meant to be impressive or alarming? Because I’m neither.

  • aaron@lemm.ee
    link
    fedilink
    English
    arrow-up
    1
    arrow-down
    42
    ·
    2 days ago

    I’m not going to parse this shit article. What does interference mean here? Please and thank you.

    • filister@lemmy.worldOP
      link
      fedilink
      English
      arrow-up
      39
      ·
      2 days ago

      That’s a very toxic attitude.

      Inference is in principle the process of generation of the AI response. So when you run locally and LLM you are using your GPU only for inference.

      • aaron@lemm.ee
        link
        fedilink
        English
        arrow-up
        23
        arrow-down
        7
        ·
        2 days ago

        Yeah, I misread because I’m stupid. Thanks for replying, non-toxic man.