Skip to content

Create image variation

client.images.createVariation(ImageCreateVariationParams { image, model, n, 3 more } body, RequestOptionsoptions?): ImagesResponse { created, background, data, 4 more }
POST/images/variations

Creates a variation of a given image. This endpoint only supports dall-e-2.

ParametersExpand Collapse
body: ImageCreateVariationParams { image, model, n, 3 more }
image: Uploadable

The image to use as the basis for the variation(s). Must be a valid PNG file, less than 4MB, and square.

model?: (string & {}) | ImageModel | null

The model to use for image generation. Only dall-e-2 is supported at this time.

Accepts one of the following:
(string & {})
ImageModel = "gpt-image-1.5" | "dall-e-2" | "dall-e-3" | 2 more
Accepts one of the following:
"gpt-image-1.5"
"dall-e-2"
"dall-e-3"
"gpt-image-1"
"gpt-image-1-mini"
n?: number | null

The number of images to generate. Must be between 1 and 10.

minimum1
maximum10
response_format?: "url" | "b64_json" | null

The format in which the generated images are returned. Must be one of url or b64_json. URLs are only valid for 60 minutes after the image has been generated.

Accepts one of the following:
"url"
"b64_json"
size?: "256x256" | "512x512" | "1024x1024" | null

The size of the generated images. Must be one of 256x256, 512x512, or 1024x1024.

Accepts one of the following:
"256x256"
"512x512"
"1024x1024"
user?: string

A unique identifier representing your end-user, which can help OpenAI to monitor and detect abuse. Learn more.

ReturnsExpand Collapse
ImagesResponse { created, background, data, 4 more }

The response from the image generation endpoint.

created: number

The Unix timestamp (in seconds) of when the image was created.

background?: "transparent" | "opaque"

The background parameter used for the image generation. Either transparent or opaque.

Accepts one of the following:
"transparent"
"opaque"
data?: Array<Image { b64_json, revised_prompt, url } >

The list of generated images.

b64_json?: string

The base64-encoded JSON of the generated image. Returned by default for the GPT image models, and only present if response_format is set to b64_json for dall-e-2 and dall-e-3.

revised_prompt?: string

For dall-e-3 only, the revised prompt that was used to generate the image.

url?: string

When using dall-e-2 or dall-e-3, the URL of the generated image if response_format is set to url (default value). Unsupported for the GPT image models.

output_format?: "png" | "webp" | "jpeg"

The output format of the image generation. Either png, webp, or jpeg.

Accepts one of the following:
"png"
"webp"
"jpeg"
quality?: "low" | "medium" | "high"

The quality of the image generated. Either low, medium, or high.

Accepts one of the following:
"low"
"medium"
"high"
size?: "1024x1024" | "1024x1536" | "1536x1024"

The size of the image generated. Either 1024x1024, 1024x1536, or 1536x1024.

Accepts one of the following:
"1024x1024"
"1024x1536"
"1536x1024"
usage?: Usage { input_tokens, input_tokens_details, output_tokens, 2 more }

For gpt-image-1 only, the token usage information for the image generation.

input_tokens: number

The number of tokens (images and text) in the input prompt.

input_tokens_details: InputTokensDetails { image_tokens, text_tokens }

The input tokens detailed information for the image generation.

image_tokens: number

The number of image tokens in the input prompt.

text_tokens: number

The number of text tokens in the input prompt.

output_tokens: number

The number of output tokens generated by the model.

total_tokens: number

The total number of tokens (images and text) used for the image generation.

output_tokens_details?: OutputTokensDetails { image_tokens, text_tokens }

The output token details for the image generation.

image_tokens: number

The number of image output tokens generated by the model.

text_tokens: number

The number of text output tokens generated by the model.

Create image variation

import OpenAI from 'openai';

const client = new OpenAI({
  apiKey: process.env['OPENAI_API_KEY'], // This is the default and can be omitted
});

const imagesResponse = await client.images.createVariation({
  image: fs.createReadStream('otter.png'),
});

console.log(imagesResponse.created);
{
  "created": 0,
  "background": "transparent",
  "data": [
    {
      "b64_json": "b64_json",
      "revised_prompt": "revised_prompt",
      "url": "url"
    }
  ],
  "output_format": "png",
  "quality": "low",
  "size": "1024x1024",
  "usage": {
    "input_tokens": 0,
    "input_tokens_details": {
      "image_tokens": 0,
      "text_tokens": 0
    },
    "output_tokens": 0,
    "total_tokens": 0,
    "output_tokens_details": {
      "image_tokens": 0,
      "text_tokens": 0
    }
  }
}
Returns Examples
{
  "created": 0,
  "background": "transparent",
  "data": [
    {
      "b64_json": "b64_json",
      "revised_prompt": "revised_prompt",
      "url": "url"
    }
  ],
  "output_format": "png",
  "quality": "low",
  "size": "1024x1024",
  "usage": {
    "input_tokens": 0,
    "input_tokens_details": {
      "image_tokens": 0,
      "text_tokens": 0
    },
    "output_tokens": 0,
    "total_tokens": 0,
    "output_tokens_details": {
      "image_tokens": 0,
      "text_tokens": 0
    }
  }
}