Comment on ChatGPT's 'hallucination' problem hit with another privacy complaint in EU
Technus@lemmy.zip 7 months ago
This is an inherent, likely unfixable issue with LLMs because they simply don’t know right from wrong, or truth from fiction. All they do is output words that are likely to go together.
It’s literally just the Predictive Text game, or the “type <some prompt> and let your keyboard finish the sentence” meme. It’s not the same algorithms (autocorrect is much less sophisticated) but they’re surprisingly similar in how they actually function.
You can try to control what an LLM outputs by changing the prompt or adjust the model with negative feedback for certain combinations of words or phrases, but you can’t just tell it “don’t make up lies about people” and expect that to work.
Technus@lemmy.zip 7 months ago
By the way, for anyone interested in how ChatGPT works, the channel 3blue1brown recently put out a very good video on it.