The same processes that often lead to impressive results are also what cause generative models to hallucinate: when they output incorrect or misleading information. Hallucinations or confabulations happen when models perceive patterns in their training data sets that do not exist in the real world.
For example, prompting ChatGPT-4o mini with the question “In which episode of Seinfeld does Kramer go to Bangladesh?” returned this answer:
Kramer travels to Bangladesh in the episode titled “The Little Jerry,” which is the 145th episode of Seinfeld. It is the 145th [sic] episode of the show, and it originally aired during Season 8. In this episode, Kramer becomes involved in a scheme to import a toy from Bangladesh.
While The Little Jerry is the 145th episode of Seinfeld and aired during Season 8, it features Kramer buying and training a rooster—the eponymous Little Jerry. Prompting the chatbot again with the same question returned a second, different hallucination:
Kramer goes to Bangladesh in the episode titled "The Maldives," which is the 18th episode of Season 9. In this episode, he tries to start a new life after his trip, and the storyline revolves around the idea of him becoming a "Bangladeshi" fisherman.
In reality, episode 18 of season 9 of Seinfeld is The Frogger, and in it, Kramer only goes so far as the police station. There is no episode of Seinfeld called The Maldives, and at no point during the show’s entire run does Kramer go to Bangladesh.
Just as autoregressive models such as GPT are known to hallucinate with made-up facts, other types of models can hallucinate in their own respective ways. Diffusion models used for image generation sometimes depict real-world objects in ways that do not match reality.