• GenesisJones@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      8 months ago

      This reminds me of an NPR podcast from 5 or 6 years ago about the people who get paid by Facebook to moderate the worst of the worst. They had a former employee giving an interview about the manual review of images that were CP andrape related shit iirc. Terrible stuff

    • Clbull@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      8 months ago

      So they paid Kenyan workers $2 an hour to sift through some of the darkest shit on the internet.

      Ugh.

        • SacrificedBeans@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          8 months ago

          I’m sure there’s some loophole there, maybe between countries’ laws. And if there isn’t, Hey! We’ll make one!

        • smooth_tea@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          8 months ago

          I really find this a bit alarmist and exaggerated. Consider the motive and the alternative. You really think companies like that have any other options than to deal with those things?

        • Clbull@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          8 months ago

          Isn’t CSAM classed as images and videos which depict child sexual abuse? Last time I checked written descriptions alone did not count, unless they were being forced to look at AI generated image prompts of such acts?

          • Strawberry@lemmy.blahaj.zone
            link
            fedilink
            English
            arrow-up
            0
            ·
            8 months ago

            That month, Sama began pilot work for a separate project for OpenAI: collecting sexual and violent images—some of them illegal under U.S. law—to deliver to OpenAI. The work of labeling images appears to be unrelated to ChatGPT.

            This is the quote in question. They’re talking about images

  • Pohl@lemmy.world
    link
    fedilink
    English
    arrow-up
    1
    ·
    8 months ago

    If you ever needed a lesson in the difference between power and authority, this is a good one.

    The leaders of this coup read the rules and saw that they could use the board to remove Altman, they had the authority to make the move and “win” the game.

    It seems that they, like many fools mistook authority for power. The “rules” said they could do it! Alas they did not have the power to execute the coup. All the rules in the world cannot make the organization follow you.

    Power comes from people who grant it to you. Authority comes from paper. Authority is the guidelines for the use of power, without power, it is pointless.

  • ribboo@lemm.ee
    link
    fedilink
    English
    arrow-up
    1
    ·
    8 months ago

    It’s rather interesting here that the board, consisting of a fairly strong scientific presence, and not so much a commercial one, is getting such hate.

    People are quick to jump on for profit companies that do everything in their power to earn a buck. Well, here you have a company that fires their CEO for going too much in the direction of earning money.

    Yet every one is all up in arms over it. We can’t have the cake and eat it folks.

    • TurtleJoe@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 months ago

      It’s my opinion that every single person in the upper levels is this organization is a maniac. They are all a bunch of so-called “rationalist” tech-right AnCaps that justify their immense incomes through the lens of Effective Altruism, the same ideology that Sam Bankman-fried used to justify his theft of billions from his customers.

      Anybody with the urge to pick a “side” here ought to think about taking a step back and reconsider; they are all bad people.

  • Even_Adder@lemmy.dbzer0.com
    link
    fedilink
    English
    arrow-up
    0
    ·
    8 months ago

    You’re not going to develop AI for the benefit of humanity at Microsoft. If they go there, we’ll know "Open"AI’s mission was all a lie.

    • Gork@lemm.ee
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 months ago

      Yeah Microsoft is definitely not going to be benevolent. But I saw this as a foregone conclusion since AI is so disruptive that heavy commercialization is inevitable.

      We likely won’t have free access like we do now and it will be enshittified like everything else now and we’ll need to pay yet another subscription to even access it.

  • helenslunch@feddit.nl
    link
    fedilink
    English
    arrow-up
    0
    ·
    edit-2
    8 months ago

    :grabs popcorn:

    Nothing more entertaining than employees standing up against management.