• theunknownmuncher@lemmy.world
    link
    fedilink
    English
    arrow-up
    21
    arrow-down
    2
    ·
    edit-2
    4 days ago

    It literally defeats NVIDIA’s entire business model of “I shit golden eggs and I’m the only one that does and I can charge any price I want for them because you need my golden eggs”

    Turns out no one actually even needs a golden egg anyway.

    And… same goes for OpenAI, who were already losing money on every subscription. Now they’ve lost the ability to charge a premium for their service (anyone can train a GPT4 equivalent model cheaply, or use DeepSeek’s existing open models) and subscription prices will need to come down, so they’ll be losing money even faster

    • Justin@lemmy.jlh.name
      link
      fedilink
      English
      arrow-up
      10
      ·
      edit-2
      4 days ago

      Nvidia cards were the only GPUs used to train DeepSeek v3 and R1. So, that narrative still superficially holds. Other stocks like TSMC, ASML, and AMD are also down in pre-market.

        • Justin@lemmy.jlh.name
          link
          fedilink
          English
          arrow-up
          8
          ·
          4 days ago

          Ah, fair. I guess it makes sense that Wall Street is questioning the need for these expensive blackwell gpus when the hopper gpus are already so good?

          • legion02@lemmy.world
            link
            fedilink
            English
            arrow-up
            5
            ·
            4 days ago

            It’s more that the newer models are going to need less compute to train and run them.

            • frezik@midwest.social
              link
              fedilink
              English
              arrow-up
              8
              ·
              4 days ago

              Right. There’s indications of 10x to 100x less compute power needed to train the models to an equivalent level. Not a small thing at all.

              • NuXCOM_90Percent@lemmy.zip
                link
                fedilink
                English
                arrow-up
                5
                ·
                edit-2
                4 days ago

                Not small but… smaller than you would expect.

                Most companies aren’t, and shouldn’t be, training their own models. Especially with stuff like RAG where you can use the highly trained model with your proprietary offline data with only a minimal performance hit.

                What matters is inference and accuracy/validity. Inference being ridiculously cheap (the reason why AI/ML got so popular) and the latter being a whole different can of worms that industry and researchers don’t want you to think about (in part because “correct” might still be blatant lies because it is based on human data which is often blatant lies but…).

                And for the companies that ARE going to train their own models? They make enough bank that ordering the latest Box from Jensen is a drop in the bucket.


                That said, this DOES open the door back up for tiered training and the like where someone might use a cheaper commodity GPU to enhance an off the shelf model with local data or preferences. But it is unclear how much industry cares about that.