• otp@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    4
    ·
    2 months ago

    I’m curious to know what happens if you ask ChatGPT to make you a text adventure based on that prompt.

    Not curious enough to try it and play it myself, though.

    • kat_angstrom@lemmy.world
      link
      fedilink
      English
      arrow-up
      12
      ·
      2 months ago

      It lacks cohesion the longer it goes on, not so much “hallucinating” as it is losing the thread, losing the plot. Internal consistency goes out the window, previously-made declarations are ignored, and established canon gets trounced upon.

      But that’s cuz it’s not AI, it’s just LLM all the way down.

        • kat_angstrom@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          ·
          2 months ago

          Depends on complexity and the number of elements to keep track of, and varies between models and people. Try it out for yourself to see! :)

        • CheeseNoodle@lemmy.world
          link
          fedilink
          English
          arrow-up
          3
          ·
          2 months ago

          Its kind of an exponential falloff, for a few lines it can follow concrete mathematical rules, for a few paragraphs it can remember basic story beats, for a few pages it can just about remember your name.

    • cadekat@pawb.social
      link
      fedilink
      English
      arrow-up
      7
      ·
      2 months ago

      It works okay for a while, but eventually it loses the plot. The storylines are usually pretty generic and washed out.

      • otp@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 months ago

        My god… they’ve reached PS1-era JRPG level in terms of video game storytelling…