OPENAI PUB_DATE: 2026.01.22

OPENAI GPT-IMAGE-1-MINI: CHEAPER IMAGE GENERATION WITH TEXT+IMAGE INPUT

OpenAI released gpt-image-1-mini, a cost-efficient image model that accepts text and image inputs and returns images. Pricing is low per image ($0.005 at 1024x1...

OpenAI gpt-image-1-mini: cheaper image generation with text+image input

OpenAI released gpt-image-1-mini, a cost-efficient image model that accepts text and image inputs and returns images. Pricing is low per image ($0.005 at 1024x1024 low quality; $0.011 medium; $0.036 high) with token-based rates for inputs/outputs and discounted cached inputs. It offers snapshots for version stability, defined rate limits (TPM/IPM by tier), and access via Images, Responses, Assistants, and Batch endpoints.

[ WHY_IT_MATTERS ]
01.

Lower per-image costs make scalable asset generation feasible without custom model hosting.

02.

Snapshots and clear limits reduce production risk and help plan capacity and reproducibility.

[ WHAT_TO_TEST ]
  • terminal

    Benchmark quality tiers and sizes against your acceptance criteria and measure latency since the model is the slowest tier.

  • terminal

    Load test TPM/IPM limits, implement retries/backoff, and validate savings from cached inputs in real workloads.

[ BROWNFIELD_PERSPECTIVE ]

Legacy codebase integration strategies...

  • 01.

    If migrating from DALL·E or gpt-image-1, align payloads and size/quality params, update cost calculators, and lock snapshots for reproducibility.

  • 02.

    Add rate-limit aware queuing and per-tenant budgeting to avoid regressions in throughput and spend.

[ GREENFIELD_PERSPECTIVE ]

Fresh architecture paradigms...

  • 01.

    Standardize on the Images or Responses API with snapshots, and use Batch for bulk/offline generation to control cost.

  • 02.

    Design for idempotent jobs, request deduping, and observability on cost, latency, and quality acceptance rates.