Ok, you have a moderately complex math problem you needed to solve. You gave the problem to 6 LLMS all paid versions. All 6 get the same numbers. Would you trust the answer?
Ok, you have a moderately complex math problem you needed to solve. You gave the problem to 6 LLMS all paid versions. All 6 get the same numbers. Would you trust the answer?
LLMs don’t and can’t do math. They don’t calculate anything, that’s just not how they work. Instead, they do this:
It could be right, it could be wrong. If there’s enough pattern in the training data, it could remember the correct answer. Otherwise it’ll just place a plausible looking value there (behavior known as AI hallucination). So, you can not “trust” it.
A good one will interpret what you are asking and then write code, often python I notice, and then let that do the math and return the answer. A math problem should use a math engine and that’s how it gets around it.
But really why bother, go ask wolfram alpha or just write the math problem in code yourself.
Every LLM answer is a hallucination.
Some are just realistic to the point of being correct. It frightens me how many users have no idea about any of that.
They calculate the statistical probability of the next token in an array of previous tokens