Basic Concepts

AI Hallucination — What is it?

The tendency of AI models to generate non-real, fabricated, or incorrect information/content. It manifests differently in text, image, and video generation.

Detailed Explanation of AI Hallucination

AI hallucination is the tendency of AI models to generate content that doesn't correspond to reality, is fabricated, or incorrect. This concept manifests in different forms in both language models (LLMs) and image/video generation models, and is one of the most significant limitations of AI systems.

Examples of hallucination in image generation: extra or missing fingers, inconsistent shadows, unreadable or nonsensical text, anatomical errors, physics-defying elements, and repeating or merging objects. In video generation, temporal inconsistencies (objects suddenly disappearing), motion artifacts, and morphological changes are hallucination symptoms.

Causes of hallucination include gaps in the model's training data, ambiguities in latent space, low-quality or vague prompts, and limits of the model's generalization capacity. For example, hand and finger rendering is particularly challenging for AI models because training data contains hand images in highly varied angles and positions.

Methods to reduce hallucination: writing detailed and specific prompts, using negative prompts to specify unwanted elements, employing structural control tools like [ControlNet](https://tasarim.ai/sozluk/controlnet), preferring low denoising values, and post-processing corrections.

Latest-generation models like [Midjourney](https://tasarim.ai/kesfet/ai-gorsel-uretimi/midjourney) v6 and [Flux](https://tasarim.ai/kesfet/ai-gorsel-uretimi/flux) have significantly reduced hallucination rates. [DALL-E 3](https://tasarim.ai/kesfet/ai-gorsel-uretimi/dall-e-3) has largely solved text rendering hallucinations. [Ideogram](https://tasarim.ai/kesfet/ai-gorsel-uretimi/ideogram) is a pioneer in minimizing typography hallucinations.

In video generation, [Sora](https://tasarim.ai/kesfet/ai-video-uretimi/sora) has taken important steps in reducing physics simulation hallucinations, while [Runway](https://tasarim.ai/kesfet/ai-video-uretimi/runway) Gen-4 has made major progress in temporal consistency.

Practical tip: Always carefully review generated content to deal with hallucinations. Focus especially on critical areas like hands, text, faces, and physical interactions. You can fix problematic areas with [inpainting](https://tasarim.ai/sozluk/inpainting) or try regeneration with different seeds. Making your prompt more specific reduces hallucination risk.

More Basic Concepts Terms