• spongebue@lemmy.world
      link
      fedilink
      English
      arrow-up
      43
      arrow-down
      1
      ·
      4 months ago

      First it was GPUs because crypto, then this. Wonder what useless thing the tech bros will cover up with in a few years!

      • lividweasel@lemmy.world
        link
        fedilink
        English
        arrow-up
        93
        ·
        4 months ago

        Article in 2027:

        Keyboard prices soared this month, as tech giants pivoted from failed AI projects to employing hordes of monkeys typing randomly. One CEO was quoted as saying, “Just a few trillion more dollars, and I think our random typing model could reproduce the lost contents of the Library of Alexandria.”

        • DaddleDew@lemmy.world
          link
          fedilink
          English
          arrow-up
          41
          ·
          edit-2
          4 months ago

          When in a gold rush, be the one selling shovels.

          I’m off to buy stocks in bananas.

    • givesomefucks@lemmy.world
      link
      fedilink
      English
      arrow-up
      28
      ·
      4 months ago

      Good luck…

      Even when the bubble bursts, they’re going to have an insane amount of computing power just sitting there, it will get sold off in bankruptcy proceedings, and some company will gobble it up and operate at a loss while continuing to secure future supply contracts.

      There’s a very real chance that we’re witnessing the slow death of home computing.

      The way things shake out it might end up being prohibitively expensive compared to cloud computing, and once that’s the norm they price gouge like Walmart did to destroy small businesses.

      Instead of dropping a couple grand for a PC every couple years, we’ll have steady contracts paying for month at a time indefinitely.

      • jollyrogue@lemmy.ml
        link
        fedilink
        English
        arrow-up
        3
        ·
        4 months ago

        Nah. Web devs will create even more bloated web pages to keep home computing in business.

        For real though, most people don’t need that much computing power, and we reached the plateau 12 years ago. That’s why we’re seeing crypto and AI grifts happen. They recentralize decentralized systems. The elites are striking back.

        You know the saying“information wants to be free; information wants to be expensive”? This is the expensive part where people try to horde knowledge by making it inaccessible to everyday people.

          • jollyrogue@lemmy.ml
            link
            fedilink
            English
            arrow-up
            1
            ·
            4 months ago

            How is this applicable to the comment? Companies never figured out how to charge rent for those.

            Devs see home computers as a free resource, and the burden is on the consumer to buy a computer which runs their software.

      • Fluffy Kitty Cat@slrpnk.net
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        1
        ·
        4 months ago

        Those GPUs fry themselves in a year or two, and utility prices will put pressure on governments to concept datacenters

    • ZoteTheMighty@lemmy.zip
      link
      fedilink
      English
      arrow-up
      8
      arrow-down
      1
      ·
      4 months ago

      I think it’s a lost cause. Essentially both crypto and AI were big because someone figured out how to offload shit to a GPU efficiently. There’s probably a ton of other appllications for GPUs we haven’t even tapped.

      • hark@lemmy.world
        link
        fedilink
        English
        arrow-up
        13
        arrow-down
        1
        ·
        4 months ago

        I’ve got this crazy idea where we can use GPUs to render 3D scenes efficiently.

      • tal@lemmy.today
        link
        fedilink
        English
        arrow-up
        6
        ·
        4 months ago

        Serial compute isn’t doing the double-every-18-months-in-speed since something like the early 2000s.

        Unlike with serial compute, not all problems can be solved, run faster, with parallel compute. But at some point, unless we figure out some sort of new way to play with physics, we pretty much have to move to parallel compute where we can if we want much more performance.

    • TheEighthDoctor@lemmy.zip
      link
      fedilink
      English
      arrow-up
      4
      ·
      4 months ago

      If the AI bubble bursts most of the western world will be thrown into deep recession again and I hope we don’t get a repeat of 2008 just for cheap RAM or GPUs

      • wewbull@feddit.uk
        link
        fedilink
        English
        arrow-up
        3
        ·
        4 months ago

        It’s already there for most of us. It only a few rich companies getting richer that’s making the line go up.