Ok, you have a moderately complex math problem you needed to solve. You gave the problem to 6 LLMS all paid versions. All 6 get the same numbers. Would you trust the answer?

  • Mark with a Z@suppo.fi
    link
    fedilink
    arrow-up
    26
    arrow-down
    2
    ·
    edit-2
    20 hours ago

    LLMs don’t and can’t do math. They don’t calculate anything, that’s just not how they work. Instead, they do this:

    2 + 2 = ? What comes after that? Oh, I remember! It’s ‘4’!

    It could be right, it could be wrong. If there’s enough pattern in the training data, it could remember the correct answer. Otherwise it’ll just place a plausible looking value there (behavior known as AI hallucination). So, you can not “trust” it.

    • NewNewAugustEast@lemmy.zip
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      12 hours ago

      A good one will interpret what you are asking and then write code, often python I notice, and then let that do the math and return the answer. A math problem should use a math engine and that’s how it gets around it.

      But really why bother, go ask wolfram alpha or just write the math problem in code yourself.

      • CanadaPlus@lemmy.sdf.org
        link
        fedilink
        arrow-up
        7
        ·
        20 hours ago

        Some are just realistic to the point of being correct. It frightens me how many users have no idea about any of that.

    • Greg Clarke@lemmy.ca
      link
      fedilink
      English
      arrow-up
      1
      ·
      14 hours ago

      They don’t calculate anything

      They calculate the statistical probability of the next token in an array of previous tokens