← All news
Model News22 April 2026

ChatGPT Images 2.0: OpenAI's gpt-image-2 Model Explained

By Stephen Grindley

OpenAI has announced ChatGPT Images 2.0, a new image generation system powered by the gpt-image-2 model. Announced on 21 April 2026, the update introduces native reasoning for image generation, 2K resolution output, multi-image consistency, and significantly improved text rendering across more than a dozen languages.

For businesses that use AI for marketing creative, product mock-ups, training material, or internal documentation, this is the first major update to OpenAI's image stack in over a year. It also marks the end of the DALL-E line: DALL-E 2 and DALL-E 3 are scheduled to be retired on 12 May 2026, roughly three weeks from the launch of ChatGPT Images 2.0, with gpt-image-2 replacing them across every surface OpenAI controls.

ChatGPT Images 2.0 is not available in the Owlpen platform. More on that below. The remainder of this article focuses on what the new generative AI system actually does and what it means in practice for UK businesses.

What is gpt-image-2

gpt-image-2 is OpenAI's new native multimodal image model. It replaces gpt-image-1.5 and the remaining DALL-E variants across ChatGPT, Codex, and the OpenAI API. Unlike the DALL-E series, which was a standalone diffusion model called out from ChatGPT via tool use, gpt-image-2 is a foundation model built to generate text and images in one unified system.

The model has a knowledge cutoff of December 2025, which means it can reference current brand conventions, recent products, and up-to-date visual references without needing extensive prompt guidance. When paired with ChatGPT's thinking or pro tiers, it can also pull live information from the web before producing an image.

What has changed

The shift from gpt-image-1.5 to gpt-image-2 introduces five capability upgrades that are relevant to business use.

Instant mode and thinking mode

gpt-image-2 offers two modes. Instant mode produces fast outputs without deliberation (the behaviour users expect from DALL-E). Thinking mode applies explicit reasoning to the structure of the image before rendering: OpenAI describes this as the model acting as a "visual thought partner" rather than a rendering tool. In practice, thinking mode allows the model to self-verify layout, typography, and object placement before the image is finalised, which reduces the need for iterative re-prompting.

2K resolution and flexible aspect ratios

Output resolution is now up to 2K via the API, a meaningful step up for print-ready marketing assets and technical documentation. Aspect ratios are no longer locked to square or 16:9 presets: gpt-image-2 supports a continuous range from 3:1 wide to 1:3 tall, covering everything from social banners to vertical editorial layouts.

Multi-image consistency

The model can now generate up to eight coherent images from a single prompt, with character and object continuity maintained across every frame. This addresses one of the hardest problems in production image generation: producing a sequence where the same person, product, or scene looks consistent across multiple outputs. For use cases like storyboarding, campaign variants, or product photography, this removes a workflow that previously required substantial human stitching.

Text rendering and multilingual support

Text inside generated images has been one of the persistent weaknesses of diffusion-based image models. gpt-image-2 improves accuracy for in-image text in English and adds stronger support for Japanese, Korean, Chinese, Hindi, and Bengali. For agencies and in-house marketing teams producing localised collateral, this reduces dependence on post-production text replacement.

Style coverage

The model covers photorealism, cinematic rendering, manga, pixel art, and editorial illustration within a single system. OpenAI positions this as "magazine-grade" design, meaning the model is capable of composing full layouts (headlines, body text, hero imagery, and pull quotes) from a single prompt rather than requiring separate asset generation and layout work.

DALL-E retirement in three weeks

OpenAI has confirmed that DALL-E 2 and DALL-E 3 will be withdrawn on 12 May 2026, 20 days after the launch of ChatGPT Images 2.0. Businesses with internal tools or automation pipelines that call DALL-E endpoints directly should plan their migration to gpt-image-2 now: calls to the legacy endpoints will fail from that date.

Who can use it and at what price

Access is tiered. Basic gpt-image-2 generation (instant mode) is available to all ChatGPT users, including the free tier, and to Codex users from 21 April 2026. Thinking mode and the advanced reasoning features are restricted to ChatGPT Plus, Pro, and Business subscribers. Enterprise rollout is staged and expected shortly after launch.

For developers, gpt-image-2 is accessible through the OpenAI API. Pricing varies by output quality and resolution rather than a flat per-image rate, which means cost modelling for production workloads requires care: a 2K thinking-mode output with multi-image consistency will cost substantially more than a single instant-mode 1024-pixel image. We recommend benchmarking a representative sample of expected prompts before committing to production volume.

Limitations to be aware of

OpenAI is explicit that gpt-image-2 is not a finished product. Three limitations are worth flagging for anyone planning to deploy it in production workflows.

First, physical reasoning remains weak. The model struggles with objects that require coherent spatial mechanics (origami, Rubik's cubes, angled reflective surfaces). Outputs in these categories may be visually convincing but physically impossible, and human review is essential where accuracy matters.

Second, fine repetitive detail (individual grains of sand, dense foliage, detailed circuit diagrams) can exceed the fidelity limits of the model. For technical documentation that requires precise diagrammatic accuracy, traditional tools remain more reliable.

Third, iterative editing shows diminishing returns. The first one or two revisions of an image typically improve quality; further revisions tend to drift from the original intent. Teams should plan to accept or reject outputs within a small number of iterations rather than converging on a result through extended prompt refinement.

Governance and content controls

For regulated sectors, image generation carries different governance considerations than text. OpenAI has not published specific details on watermarking or C2PA provenance metadata for gpt-image-2 at launch, though the company's existing content moderation policies continue to apply. Businesses operating under UK GDPR or sector-specific rules (financial services, healthcare, legal) should treat generated images as they would any other third-party content: verify the license position, retain provenance records, and ensure any personally identifiable or trademarked material in outputs is handled through a formal responsible AI process.

Owlpen and ChatGPT Images 2.0

ChatGPT Images 2.0 is not available in the Owlpen platform and there are no current plans to integrate it. Owlpen is built around cost reduction, business intelligence, and document analysis workflows, which are primarily text and structured-data tasks. Image generation is not part of the platform's remit.

Clients with image generation requirements are better served by using ChatGPT, the OpenAI API, or a dedicated creative tool alongside Owlpen, rather than expecting image generation to be part of the Owlpen workflow. If you have a specific use case that sits at the boundary (for example, generating annotated diagrams from extracted invoice data), we are happy to discuss whether a bespoke integration makes sense.

Not in Owlpen

ChatGPT Images 2.0 and the gpt-image-2 model are not available inside the Owlpen platform. Owlpen focuses on text and structured-data workloads for cost reduction and business intelligence, and image generation is outside the platform's scope. Clients requiring image generation should use OpenAI's products directly.

If you would like to discuss how generative AI or the Owlpen platform could support your business, contact us at enquiries@coaleypeak.co.uk or read more about the Owlpen platform.

Disclaimer. This article is published by Coaley Peak Ltd for general informational purposes only. The views expressed are those of the author, Stephen Grindley, and do not constitute legal, regulatory, financial, or technical advice. Nothing in this article should be relied upon when making procurement, investment, compliance, or technology decisions. References to third-party products, platforms, and companies are for informational purposes only and do not constitute endorsement. Capability claims, availability dates, and pricing details cited are those reported by OpenAI and contemporaneous technology press, and have not been independently verified by Coaley Peak. Readers should seek independent professional advice appropriate to their specific circumstances. Information was accurate to the best of the author's knowledge at the date of publication. Coaley Peak Ltd and Stephen Grindley accept no liability for any loss or damage arising from reliance on the contents of this article.