DALL-E and GPT Image Gen: The Built-In Image Tool

OpenAI has shipped not one but two distinct image generation systems inside ChatGPT, and the interface doesn't make the distinction clear. DALL-E 3 — the older, more controlled system — and GPT-4o native image generation — the newer, more flexible, less predictable one — coexist in the same product. When you ask ChatGPT to make an image, the system decides which one to use based on factors that aren't fully transparent. The result is a tool that sometimes produces remarkable work and sometimes produces baffling output, with limited user control over why.

This matters because image generation is the AI feature with the highest gap between marketing and reality. Every launch demo shows cherry-picked outputs. Every product page shows the best 1% of generations. The actual hit rate — the percentage of generations that are usable without regeneration or heavy editing — is lower than any other AI product category. Understanding what these tools actually produce, consistently, is more valuable than knowing what they can produce in ideal conditions.

What The Docs Say

OpenAI's documentation describes two systems. DALL-E 3 is the text-to-image model that launched in late 2023, integrated into ChatGPT as a tool the model can call. You describe what you want, ChatGPT reformulates your prompt into a detailed DALL-E 3 prompt (often significantly longer and more specific than what you wrote), and DALL-E 3 generates the image. The reformulation is intentional — OpenAI's research showed that detailed prompts produce better results, so ChatGPT acts as a prompt expander.

GPT-4o native image generation is the newer system, announced in early 2025 [VERIFY exact date]. Unlike DALL-E 3, which is a separate model called as a tool, GPT-4o generates images as a native output modality — the same model that processes your text can also produce images. OpenAI's documentation emphasizes that this enables better prompt understanding, more coherent text rendering in images, and the ability to edit images in conversation by referring to what you want changed.

Both systems support text-to-image generation. GPT-4o image gen additionally supports image editing — you upload an image, describe the changes you want, and the model produces a modified version. OpenAI's content policy governs both systems, restricting generation of photorealistic faces of real people, explicit content, and several other categories.

What Actually Happens

I generated over 200 images across both systems over a month-long testing period, spanning illustration, photography-style, graphic design, text-heavy images, and image editing tasks.

GPT-4o's text rendering is the real breakthrough. Every previous AI image generator — DALL-E 3 included — struggled catastrophically with text. "Happy Birthday" would come out as "Hpapy Brithday." Logos were gibberish. Signs were abstract art. GPT-4o native image gen mostly solves this. It can render clean, readable text in images — signs, labels, headers, book covers, UI mockups. Not perfectly every time, but the success rate jumped from maybe 10% with DALL-E 3 to somewhere around 70-80% with GPT-4o. This is a genuine capability unlock for anyone who needs text in images — social media graphics, presentation slides, marketing materials, memes. It works.

Style consistency within a session is good. If you generate one image and then ask for another "in the same style," GPT-4o does a reasonable job maintaining consistency. Color palettes carry over. Illustration styles persist. Character designs stay roughly similar across two or three generations. This breaks down over longer sessions or when you introduce complex modifications, but for a set of 3-5 related images, it's functional. DALL-E 3 had no concept of style consistency between generations — every image was essentially independent.

Hands are better but not solved. The "AI hands" problem — extra fingers, impossible anatomy, fingers merging into each other — has improved significantly with GPT-4o image gen. Most generations produce correct hand anatomy. But "most" is not "all," and the failures still happen, especially in complex poses or when hands are interacting with objects. If your image prominently features hands, plan to regenerate at least once.

Specific faces don't work. Neither system will generate recognizable real people, and the content policy blocks attempts aggressively. This is a deliberate choice by OpenAI, not a technical limitation. If you need an image of a specific person, these tools are not an option. Stock photography or commissioned art remain the answer. For generic faces — "a woman in her 30s with dark hair" — both systems produce plausible results, though there's a noticeable homogeneity to the faces. AI-generated people have a specific look that becomes recognizable after you've seen enough of them.

Prompt specificity is everything. "A dog" gives you a generic, flat, uninteresting image of a dog. "A golden retriever sitting on a weathered wooden park bench, late afternoon light filtering through maple trees, shallow depth of field, slightly desaturated warm tones" gives you something you might actually use. The gap between a lazy prompt and a specific prompt is larger for image generation than for any other AI task. This isn't unique to OpenAI's tools — Midjourney and Stable Diffusion have the same dynamic — but it's worth emphasizing because most users' first experience with image generation is a vague prompt followed by disappointment.

Image editing is hit-or-miss. GPT-4o's ability to modify uploaded images is genuinely useful for certain tasks — style transfer (make this photo look like a watercolor), adding elements (put a hat on this person), and background changes (replace the background with a beach). Where it falls apart: precise modifications (move this object two inches to the left), clean removal (take out this person from the group photo), and detail preservation (change the shirt color but keep everything else exactly the same). The model treats the uploaded image as a reference, not a canvas. It generates a new image that's influenced by your original, not a surgical edit of your original. This distinction matters enormously in practice.

The Content Policy Reality

OpenAI's content policy is the most aggressive in the image generation space, and it has practical consequences for legitimate use cases. The guardrails block: photorealistic depictions of real public figures, violence beyond a certain threshold, explicit or sexual content, and several other categories. The implementation is broad rather than precise, which means it catches legitimate requests in the crossfire.

Examples from my testing: a request for a "dramatic war photograph in the style of Robert Capa" was blocked. A request for a medical illustration showing a surgical procedure was blocked. A request for a "noir film poster with a woman in a revealing dress" was blocked. None of these are unreasonable creative requests. All hit the content policy. The workarounds — rephrasing, abstracting, adding disclaimers — work sometimes and don't work other times, with no clear pattern to what triggers a block versus what slides through.

If your work regularly involves mature themes, medical imagery, historical violence, or anything in the gray zone of content policy — OpenAI's image tools will frustrate you. Midjourney is somewhat more permissive. Stable Diffusion (self-hosted) has no content policy at all. The trade-off is access versus control.

Where It Fits in the Landscape

The image generation landscape in 2026 has distinct tiers, and OpenAI's tools occupy a specific niche.

For convenience: OpenAI wins. If you're already in ChatGPT and you need an image, the integrated experience — describe what you want in the same conversation where you're working on the broader project — is unmatched. No separate tool, no separate account, no workflow switching. For quick social media graphics, presentation images, and visual brainstorming, this convenience factor is the entire value proposition.

For quality: Midjourney remains ahead for aesthetic output, particularly for artistic, stylized, and photography-adjacent work. Midjourney's default aesthetic — the "Midjourney look" — is more polished and more visually striking than what either DALL-E 3 or GPT-4o produces by default [VERIFY if this gap has closed with latest GPT-4o image gen updates]. The gap narrows with careful prompting, but Midjourney requires less prompt engineering to produce visually impressive results.

For control: Stable Diffusion (and its derivatives — ComfyUI, Forge, etc.) gives you parameter-level control that neither OpenAI nor Midjourney offers. Inpainting, outpainting, ControlNet, LoRA fine-tuning — if you need precise control over the generation process, open-source tools are the answer. The cost is setup complexity and hardware requirements.

For text in images: OpenAI's GPT-4o is currently the best option. This is a clear win and a real differentiator.

For image editing: None of the AI tools are reliable enough to replace Photoshop for precise editing work. They're useful for rough conceptual edits, style exploration, and quick mockups. For production work, the traditional tools still win.

When To Use This

Use OpenAI's image generation when you need a quick visual and you're already in ChatGPT. Use it when your image needs readable text — this is the clearest competitive advantage. Use it for brainstorming visual concepts before investing in professional illustration or photography. Use it for social media graphics, blog headers, presentation slides, and other contexts where "good enough, right now" beats "perfect, in three days."

Use it when convenience matters more than quality ceiling. The best image from a Midjourney session will beat the best image from a ChatGPT session for most artistic use cases. But the average ChatGPT image workflow — from need to finished asset — takes two minutes. The average Midjourney workflow takes ten to fifteen. For high-volume, low-stakes visual content, that speed advantage compounds.

When To Skip This

Skip it when you need consistent, branded visual output. Neither system gives you enough control over style to maintain brand consistency across dozens of images. You'll get close sometimes and way off other times, with no reliable knobs to turn.

Skip it when you need photorealistic humans in specific scenarios. The content policy, the face homogeneity, and the occasional anatomical failures make AI-generated people unreliable for professional use cases like advertising, editorial, or marketing materials. Stock photography is boring but consistent. AI-generated people are interesting but unpredictable.

Skip it when precision matters. If the image needs to be exactly what you specified — exact layout, exact colors, exact composition — you'll spend more time regenerating and adjusting than you would have spent creating it manually or briefing a designer. AI image generation is a suggestion engine, not a specification engine.

Skip it when the content policy will be a problem. If your work involves anything that might trigger OpenAI's guardrails — and the guardrails are broad enough that this covers more creative work than you'd expect — use a tool with more permissive policies or no policies at all.

The bottom line: OpenAI's image tools are the most convenient and the most capable for text-in-image tasks. For everything else, they're competitive but not best-in-class. The right tool depends on whether you're optimizing for speed, quality, control, or creative freedom — and no single tool wins on all four.


This is part of CustomClanker's GPT Deep Cuts series — what OpenAI's features actually do in practice.