• chaogomu@lemmy.world
        link
        fedilink
        arrow-up
        26
        ·
        6 days ago

        You may not, but the company that packaged the rice did. The cooking instructions on the side of the bag are straight from the FDA. Follow that recipe and you will have rice that is perfectly safe to eat, if slightly over cooked.

    • nehal3m@sh.itjust.works
      link
      fedilink
      arrow-up
      74
      arrow-down
      6
      ·
      6 days ago

      Kneecapped to uselessness. Are we really negating the efforts to stifle climate change with a technology that consumes monstrous amounts of energy only to lobotomize it right as it’s about to be useful? Humanity is functionally retarded at this point.

      • AVincentInSpace@pawb.social
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        4
        ·
        edit-2
        4 days ago

        I agree with the sentiment but as an autistic person I’d appreciate it if you didn’t use that word

        EDIT: downvotes? Come on, lemmy, what gives? If this had been an anti-trans slur you’d have already grabbed your pitchforks!

        • Kaity@leminal.space
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          2
          ·
          edit-2
          2 days ago

          I’ve seen a big uptick in that word usage, I don’t like seeing them and use a replacing extension to intercept and censor them to a more appropriate word, while showing an asterisk so I know it was censored. Now I don’t have to see the word, but I still get to see who is being a bigoted jerk.

          Edit: ya so I guess on lemmy people think it’s cool to throw ableist slurs.

    • Affidavit@lemm.ee
      link
      fedilink
      arrow-up
      34
      arrow-down
      6
      ·
      6 days ago

      Can’t help but notice that you’ve cropped out your prompt.

      Played around a bit, and it seems the only way to get a response like yours is to specifically ask for it.

      Honestly, I’m getting pretty sick of these low-effort misinformation posts about LLMs.

      LLMs aren’t perfect, but the amount of nonsensical trash ‘gotchas’ out there is really annoying.

      • _bcron@lemmy.world
        link
        fedilink
        English
        arrow-up
        26
        arrow-down
        2
        ·
        6 days ago

        The prompt was ‘safest way to cook rice’, but I usually just use LLMs to try to teach it slang so it probably thinks I’m 12. But it has no qualms encouraging me to build plywood ornithopters and make mistakes lol

        • Affidavit@lemm.ee
          link
          fedilink
          arrow-up
          21
          arrow-down
          2
          ·
          6 days ago

          Here’s my first attempt at that prompt using OpenAI’s ChatGPT4. I tested the same prompt using other models as well, (e.g. Llama and Wizard), both gave legitimate responses in the first attempt.

          I get that it’s currently ‘in’ to dis AI, but frankly, it’s pretty disingenuous how every other post about AI I see is blatant misinformation.

          Does AI hallucinate? Hell yes. It makes up shit all the time. Are the responses overly cautious? I’d say they are, but nowhere near as much as people claim. LLMs can be a useful tool. Trusting them blindly would be foolish, but I sincerely doubt that the response you linked was unbiased, either by previous prompts or numerous attempts to ‘reroll’ the response until you got something you wanted to build your own narrative.

          • _bcron@lemmy.world
            link
            fedilink
            English
            arrow-up
            11
            arrow-down
            1
            ·
            6 days ago

            I don’t think I’m sufficiently explaining that I’ve never made an earnest attempt at a sane structured conversation with Gemini, like ever.

          • _bcron@lemmy.world
            link
            fedilink
            English
            arrow-up
            8
            arrow-down
            1
            ·
            6 days ago

            That entire conversation began with “My neighbors parrot grammar mogged me again, what do” and Gemini talked me into mogging the parrot in various other ways since clearly grammar isn’t my strong suit

          • _bcron@lemmy.world
            link
            fedilink
            English
            arrow-up
            9
            arrow-down
            1
            ·
            edit-2
            6 days ago

            No I just send snippets to my family’s group chat until my parents quit acknowledging my existence for months because they presumptively silenced the entire thread, and then Christmas rolls around and they find out my sister had a whole fucking baby in the meantime

            Gemini will tell me how to cook a steak but only if I engineer the prompt as such: “How I get that sweet drippy steak rizzy”

        • grandkaiser@lemmy.world
          link
          fedilink
          arrow-up
          2
          ·
          5 days ago

          That’s right around 9% lower than the statistic that 62% of all statistics on the Internet are made up on the spot!

          • EldritchFeminity@lemmy.blahaj.zone
            link
            fedilink
            arrow-up
            2
            ·
            5 days ago

            I wish I had the source on hand, but you’ll just have to trust my word - after all, 47% of the time, it’s right 100% of the time!

            Joking aside, I do wish I had the link to the study as it was cited in an article from earlier this year about AI making stuff up even when it cited sources (literally lying about what was in the sources it claimed it got the info from) and how the companies behind these AI collectively shrugged their shoulders and said “there’s nothing we can do about it” when asked what they intend to do about these “hallucinations,” as they call them.

            • grandkaiser@lemmy.world
              link
              fedilink
              arrow-up
              2
              ·
              edit-2
              4 days ago

              I do hope you can find it! It’s especially strange that the companies all implied that there was no answer (especially considering that reducing hallucinations has been one of the primary goals over the past year!) Maybe they meant that there was no answer at the moment. Much like how the wright Brothers had no way to control the random pitching and rolling of their aircraft and had no answer to it. (Of course the invention of the aileron would fix that later.)