OpenAI · New in 2026 · up to 16 references

New OpenAI model GPT Image 2

Pass up to 16 reference images in one request — GPT Image 2 blends character, style, background, and composition into a single result. Plus a native edit endpoint, legible text on images, and photorealism. No VPN, no ChatGPT Plus.

up to 16references
~10 seclow/medium
from $0.07per image
GPT Image 2 — Hubery.ai gpt-image-2
Compose a ramen shop using 3 references: interior from the first photo, neon sign from the second, Ghibli mood from the third
Ramen shop — GPT Image 2
1024×1024 · 10 sec

Done. Composed the scene from all 3 references: interior, sign, and mood. Want to add more references or edit?

Via edit: add a "RAMEN" sign to the facade, leave everything else
Result with text — GPT Image 2

Edit applied. The "RAMEN" sign was added to the facade; the rest of the composition is preserved pixel-for-pixel.

Describe an image...
// features

What GPT Image 2 Can Do

Six capabilities you won't find in earlier OpenAI image models

01

Up to 16 references at once

Pass character, style, background, and composition references in a single request — GPT Image 2 fuses them into one coherent result.

you 16 product shots + clothing ref + 3 backdrops = marketplace listing
gpt + All 20 references respected in a single coherent composition
02

Native edit endpoint

In-place edits — change a background, remove an object, swap a label — without regenerating the rest of the image.

you Edit: swap the background on this photo to studio, keep the face
gpt + Edit applied, face preserved pixel-for-pixel
03

Text on images

Renders legible text, logos, and captions in English, Russian, and other languages — no garbled lettering.

you Movie poster with the title "The Last Dawn" in noir style
gpt + Title "The Last Dawn" rendered legibly, no text artifacts
04

Quality tiers

Low, Medium, High, Auto — pick the speed/cost trade-off you need. From $0.07 for fast iteration to high-detail final assets.

you High quality: photorealistic water drop on a rose petal
gpt + High quality in 24s, macro-level detail
05

Photorealism + art

From product photography to painting and 3D — one model covers the whole visual spectrum.

you Surreal space scene, floating islands, oil painting
gpt + Prompt handled end-to-end, no lost detail
06

Speed

Results in 8–15 seconds at low/medium, ~25 seconds at high. Parallel generations don't block the UI.

you Generate 4 variants in low, then upscale the chosen one
gpt + Low — 4 variants in 12s, total $0.30
// prompt lab

Prompt Lab

36 ready-to-use prompts in 6 categories — click and create

// get started

Three Steps to a Masterpiece

01

Open Hubery.ai

No VPN, no OpenAI account needed. Select GPT Image 2 from the model list and attach up to 16 reference images.

02

Describe the image

In any language. Describe the goal and the role of each reference — character, background, style, composition. Pick a quality tier: low, medium, high, or auto.

03

Get your result

8–15 seconds for low/medium, ~25 seconds for high. Download the result or call the edit endpoint to tweak the image in place — the original is preserved.

// comparison

Hubery vs ChatGPT Plus

Why pay $20/mo when there's Hubery

Up to 16 reference images
- ChatGPT Plus + Hubery.ai
No VPN required
- ChatGPT Plus + Hubery.ai
No ChatGPT Plus subscription
- ChatGPT Plus + Hubery.ai
Native edit endpoint
+ ChatGPT Plus + Hubery.ai
Multilingual prompts
+ ChatGPT Plus + Hubery.ai
Pay-as-you-go pricing
- ChatGPT Plus + Hubery.ai
From $0.07 per image
- ChatGPT Plus + Hubery.ai
50+ other models in one account
- ChatGPT Plus + Hubery.ai
// questions

GPT Image 2 FAQ

What is GPT Image 2?

GPT Image 2 is the next-generation image model from OpenAI (2026), succeeding GPT Image 1.5. The key new capability is multi-reference image-to-image: it accepts up to 16 input images in a single request and composes the result with all of them in mind. It also adds a native edit endpoint for in-place editing while keeping the family's strengths: accurate text rendering, photorealism, and long-prompt instruction following.

How is GPT Image 2 different from GPT Image 1.5?

Four core upgrades: (1) up to 16 reference images per request — combine character, style, background, and composition in a single shot; (2) a dedicated edit endpoint that preserves untouched pixels of the source; (3) explicit quality tiers (low / medium / high / auto) that trade speed and price for fidelity; (4) better adherence to long, structured prompts.

What does "16 reference images" actually mean?

You can attach up to 16 images as visual context: a person, a clothing reference, a background, a color palette, a composition reference. GPT Image 2 reads all of them and produces a new image that respects each. Useful for product photography, character consistency without LoRA training, brand-style transfer, mood boards, and collage workflows.

How much does one GPT Image 2 generation cost?

On Hubery.ai it ranges from about $0.07 per image at low quality to roughly $0.40 at high quality, depending on the chosen tier and resolution. Free tokens are granted on sign-up so you can try the model before paying. No subscriptions, no ChatGPT Plus required — pay-as-you-go.

Do I need a VPN or ChatGPT Plus?

No. GPT Image 2 is available directly on Hubery.ai with no VPN and no ChatGPT Plus subscription. It works in English and dozens of other languages, including Russian.

What formats and resolutions does GPT Image 2 support?

GPT Image 2 supports 1024×1024 (square), 1536×1024 (landscape), and 1024×1536 (portrait). For image-to-image it accepts JPG, PNG, and WebP. The dedicated edit mode keeps the unchanged regions of the source image pixel-accurate.

Can GPT Image 2 render text on images?

Yes. GPT Image 2 keeps the family's signature strength of legible text rendering. It produces clean, readable text on banners, logos, posters, menus, and infographics — in English, Russian, and other languages.

Can I use the results commercially?

Yes. Images generated with GPT Image 2 on Hubery.ai can be used for commercial work — marketing, social media, websites, marketplaces, and client deliverables.

Updated: April 23, 2026

GPT Image 2 — New OpenAI Model with 16-Reference Image-to-Image

GPT Image 2 is OpenAI's next-generation image model, the successor to GPT Image 1.5. The headline upgrade is multi-reference image-to-image: the model accepts up to 16 input images in a single request and composes the output while respecting every one of them. It also ships a dedicated edit endpoint for in-place changes — edits are applied to the source image without regenerating untouched pixels. GPT Image 2 keeps the family's signature strengths: legible text rendering, photorealism, and long structured-prompt adherence.

What's new in GPT Image 2 vs. 1.5

GPT Image 2 builds on the same multimodal backbone as GPT Image 1.5 but adds four capabilities that no prior OpenAI image model supported end-to-end:

  • Multi-reference (up to 16 images). You can attach a character reference, a style sample, a color palette, a background, and a composition reference — and the model fuses them into one coherent output. Previously this required chaining multiple generations or external LoRAs.
  • Native edit endpoint. Upload a finished image and describe the change; the model treats the request like a mask, leaving the rest of the pixels alone. Ideal for poster tweaks, product shots, and banner iteration without losing the source.
  • Quality tiers (low / medium / high / auto). Pick fast, cheap iteration for exploration or maximum fidelity for final assets. Price and latency scale with the tier.
  • Better long-prompt adherence. GPT Image 2 handles multi-step instructions over 1,500+ characters more reliably — lists of requirements, technical parameters, and composition constraints are respected more often.

Multi-reference: what 16 references are actually for

The 16-reference capability is the main reason to move from GPT Image 1.5 to GPT Image 2. Practical workflows:

  • Marketplace product cards. Product shot + 3 backgrounds + lighting reference = a consistent series of listings.
  • Character consistency. 4–8 portraits from different angles + clothing reference + scene = new frames of the same person, no LoRA training required.
  • Brand-style transfer. Logo + brand colors + examples of existing creative = new assets that land inside your brand guidelines.
  • Mood boards and collages. Assemble a visual idea from 10–16 sources and get a single image that unifies them stylistically.
  • Style transfer. Source photo + artistic style reference = the same subject rendered in the chosen manner with original details preserved.

GPT Image 2 without ChatGPT Plus — how it works on Hubery.ai

On Hubery.ai, GPT Image 2 is available through a clean web UI — no ChatGPT Plus subscription ($20/mo), no VPN. Attach 1 to 16 reference images, describe the task, pick a quality tier, and you get a result in 8–25 seconds. Pricing ranges from about $0.07 at low quality to $0.40 at high quality. Free tokens are granted on sign-up.

Hubery.ai also gives you access to 50+ AI models inside a single account: GPT Image 2, GPT Image 1.5, DALL-E 3, Midjourney, Stable Diffusion, Flux, Seedream, NanoBanana, Ideogram, plus dozens of text, code, and audio models. No need to pay each vendor separately — everything runs off one balance.

GPT Image 2 vs GPT Image 1.5

GPT Image 1.5 is still a great choice for one-shot text-to-image generations: it's cheaper, faster, and for a typical prompt the output is comparable. GPT Image 1.5 remains available on Hubery.ai. Move to GPT Image 2 when you need image-to-image with multiple references, in-place editing without regenerating, explicit quality control, or reliable execution of a long brief. For commercial assets with character or brand consistency, GPT Image 2 saves several iteration cycles compared to 1.5.

GPT Image 2 vs DALL-E 3 and Midjourney

DALL-E 3 is built on a prior-generation diffusion pipeline and supports neither multi-reference nor a native edit endpoint. For collages, product photos with character consistency, and branded series, GPT Image 2 stays noticeably closer to the brief.

Midjourney v7 still leads on artistic atmosphere, but it runs through Discord, has no multi-image API, is weaker with on-image text, and can't edit a specific region of a finished image. GPT Image 2 wins where precision, text, edit, and predictable iteration matter.

Both DALL-E 3 and Midjourney are also available on Hubery.ai — pick the tool that fits the task without switching platforms.

Prompt examples for GPT Image 2

Ready-to-use prompts you can paste into Hubery.ai. For multi-reference prompts, attach the images first and then paste the text:

[Multi-ref: 1=product photo, 2=studio backdrop, 3=lighting reference]
Marketplace card: product from ref1 on backdrop from ref2,
match the lighting and shadows of ref3, add a "NEW" badge
in the top-right corner, 1024×1024, high quality.
[Multi-ref: 1=portrait of the person, 2=outfit sample, 3=scene]
Same person as ref1 wearing the outfit from ref2, in the
location shown in ref3. Natural light, full-length portrait,
photorealistic.
[Edit endpoint: a finished movie poster uploaded]
Replace the tagline with "Premieres May 1" in the same font,
leave everything else untouched. Preserve composition pixel-for-pixel.
[Text-to-image, high quality]
Coffee shop logo "North" — minimal, black on white,
a geometric pine tree built from two triangles, wordmark below
in a monospaced font. No gradients.

Professional use cases

GPT Image 2 covers jobs that previously required a designer or multiple specialist models:

  • E-commerce. Series of product cards with consistent background, lighting, and style — built from one product shot and a handful of references.
  • Marketing and advertising. Social creatives, banners with accurate copy and pricing, A/B variants in low quality, finals in high.
  • Branding. Logo concepts, identity mockups, and creative that stays inside an existing brand guideline via multi-reference.
  • Product prototyping. UI mockups and landing-page illustrations that match the visual language of the product.
  • Editorial and social content. Original covers and illustrations with clean, legible copy — instead of stock imagery.
  • Post-production. The edit endpoint replaces routine Photoshop chores: swap a background, remove an object, update a caption, all without re-rendering the whole image.

Try GPT Image 2

Up to 16 references · Edit endpoint · No VPN · From $0.07 per image

Create for free