Cover article to the Poison article
AI Halucination is ineviteble as users prefer certainty of answers over correctness - and so the models are trained that way. In addition, models are statistical engines, where errors accumulate with length and complexity.
Model which uses reinforcement learning.