Images
Create image
Create image edit
Create image variation
ModelsExpand Collapse
Image = object { b64_json, revised_prompt, url } Represents the content or the URL of an image generated by the OpenAI API.
Represents the content or the URL of an image generated by the OpenAI API.
The base64-encoded JSON of the generated image. Returned by default for the GPT image models, and only present if response_format is set to b64_json for dall-e-2 and dall-e-3.
For dall-e-3 only, the revised prompt that was used to generate the image.
When using dall-e-2 or dall-e-3, the URL of the generated image if response_format is set to url (default value). Unsupported for the GPT image models.
ImageEditCompletedEvent = object { b64_json, background, created_at, 5 more } Emitted when image editing has completed and the final image is available.
Emitted when image editing has completed and the final image is available.
Base64-encoded final edited image data, suitable for rendering as an image.
background: "transparent" or "opaque" or "auto"The background setting for the edited image.
The background setting for the edited image.
The Unix timestamp when the event was created.
output_format: "png" or "webp" or "jpeg"The output format for the edited image.
The output format for the edited image.
quality: "low" or "medium" or "high" or "auto"The quality setting for the edited image.
The quality setting for the edited image.
size: "1024x1024" or "1024x1536" or "1536x1024" or "auto"The size of the edited image.
The size of the edited image.
The type of the event. Always image_edit.completed.
usage: object { input_tokens, input_tokens_details, output_tokens, total_tokens } For the GPT image models only, the token usage information for the image generation.
For the GPT image models only, the token usage information for the image generation.
The number of tokens (images and text) in the input prompt.
input_tokens_details: object { image_tokens, text_tokens } The input tokens detailed information for the image generation.
The input tokens detailed information for the image generation.
The number of image tokens in the input prompt.
The number of text tokens in the input prompt.
The number of image tokens in the output image.
The total number of tokens (images and text) used for the image generation.
ImageEditPartialImageEvent = object { b64_json, background, created_at, 5 more } Emitted when a partial image is available during image editing streaming.
Emitted when a partial image is available during image editing streaming.
Base64-encoded partial image data, suitable for rendering as an image.
background: "transparent" or "opaque" or "auto"The background setting for the requested edited image.
The background setting for the requested edited image.
The Unix timestamp when the event was created.
output_format: "png" or "webp" or "jpeg"The output format for the requested edited image.
The output format for the requested edited image.
0-based index for the partial image (streaming).
quality: "low" or "medium" or "high" or "auto"The quality setting for the requested edited image.
The quality setting for the requested edited image.
size: "1024x1024" or "1024x1536" or "1536x1024" or "auto"The size of the requested edited image.
The size of the requested edited image.
The type of the event. Always image_edit.partial_image.
ImageEditStreamEvent = ImageEditPartialImageEvent { b64_json, background, created_at, 5 more } or ImageEditCompletedEvent { b64_json, background, created_at, 5 more } Emitted when a partial image is available during image editing streaming.
Emitted when a partial image is available during image editing streaming.
ImageEditPartialImageEvent = object { b64_json, background, created_at, 5 more } Emitted when a partial image is available during image editing streaming.
Emitted when a partial image is available during image editing streaming.
Base64-encoded partial image data, suitable for rendering as an image.
background: "transparent" or "opaque" or "auto"The background setting for the requested edited image.
The background setting for the requested edited image.
The Unix timestamp when the event was created.
output_format: "png" or "webp" or "jpeg"The output format for the requested edited image.
The output format for the requested edited image.
0-based index for the partial image (streaming).
quality: "low" or "medium" or "high" or "auto"The quality setting for the requested edited image.
The quality setting for the requested edited image.
size: "1024x1024" or "1024x1536" or "1536x1024" or "auto"The size of the requested edited image.
The size of the requested edited image.
The type of the event. Always image_edit.partial_image.
ImageEditCompletedEvent = object { b64_json, background, created_at, 5 more } Emitted when image editing has completed and the final image is available.
Emitted when image editing has completed and the final image is available.
Base64-encoded final edited image data, suitable for rendering as an image.
background: "transparent" or "opaque" or "auto"The background setting for the edited image.
The background setting for the edited image.
The Unix timestamp when the event was created.
output_format: "png" or "webp" or "jpeg"The output format for the edited image.
The output format for the edited image.
quality: "low" or "medium" or "high" or "auto"The quality setting for the edited image.
The quality setting for the edited image.
size: "1024x1024" or "1024x1536" or "1536x1024" or "auto"The size of the edited image.
The size of the edited image.
The type of the event. Always image_edit.completed.
usage: object { input_tokens, input_tokens_details, output_tokens, total_tokens } For the GPT image models only, the token usage information for the image generation.
For the GPT image models only, the token usage information for the image generation.
The number of tokens (images and text) in the input prompt.
input_tokens_details: object { image_tokens, text_tokens } The input tokens detailed information for the image generation.
The input tokens detailed information for the image generation.
The number of image tokens in the input prompt.
The number of text tokens in the input prompt.
The number of image tokens in the output image.
The total number of tokens (images and text) used for the image generation.
ImageGenCompletedEvent = object { b64_json, background, created_at, 5 more } Emitted when image generation has completed and the final image is available.
Emitted when image generation has completed and the final image is available.
Base64-encoded image data, suitable for rendering as an image.
background: "transparent" or "opaque" or "auto"The background setting for the generated image.
The background setting for the generated image.
The Unix timestamp when the event was created.
output_format: "png" or "webp" or "jpeg"The output format for the generated image.
The output format for the generated image.
quality: "low" or "medium" or "high" or "auto"The quality setting for the generated image.
The quality setting for the generated image.
size: "1024x1024" or "1024x1536" or "1536x1024" or "auto"The size of the generated image.
The size of the generated image.
The type of the event. Always image_generation.completed.
usage: object { input_tokens, input_tokens_details, output_tokens, total_tokens } For the GPT image models only, the token usage information for the image generation.
For the GPT image models only, the token usage information for the image generation.
The number of tokens (images and text) in the input prompt.
input_tokens_details: object { image_tokens, text_tokens } The input tokens detailed information for the image generation.
The input tokens detailed information for the image generation.
The number of image tokens in the input prompt.
The number of text tokens in the input prompt.
The number of image tokens in the output image.
The total number of tokens (images and text) used for the image generation.
ImageGenPartialImageEvent = object { b64_json, background, created_at, 5 more } Emitted when a partial image is available during image generation streaming.
Emitted when a partial image is available during image generation streaming.
Base64-encoded partial image data, suitable for rendering as an image.
background: "transparent" or "opaque" or "auto"The background setting for the requested image.
The background setting for the requested image.
The Unix timestamp when the event was created.
output_format: "png" or "webp" or "jpeg"The output format for the requested image.
The output format for the requested image.
0-based index for the partial image (streaming).
quality: "low" or "medium" or "high" or "auto"The quality setting for the requested image.
The quality setting for the requested image.
size: "1024x1024" or "1024x1536" or "1536x1024" or "auto"The size of the requested image.
The size of the requested image.
The type of the event. Always image_generation.partial_image.
ImageGenStreamEvent = ImageGenPartialImageEvent { b64_json, background, created_at, 5 more } or ImageGenCompletedEvent { b64_json, background, created_at, 5 more } Emitted when a partial image is available during image generation streaming.
Emitted when a partial image is available during image generation streaming.
ImageGenPartialImageEvent = object { b64_json, background, created_at, 5 more } Emitted when a partial image is available during image generation streaming.
Emitted when a partial image is available during image generation streaming.
Base64-encoded partial image data, suitable for rendering as an image.
background: "transparent" or "opaque" or "auto"The background setting for the requested image.
The background setting for the requested image.
The Unix timestamp when the event was created.
output_format: "png" or "webp" or "jpeg"The output format for the requested image.
The output format for the requested image.
0-based index for the partial image (streaming).
quality: "low" or "medium" or "high" or "auto"The quality setting for the requested image.
The quality setting for the requested image.
size: "1024x1024" or "1024x1536" or "1536x1024" or "auto"The size of the requested image.
The size of the requested image.
The type of the event. Always image_generation.partial_image.
ImageGenCompletedEvent = object { b64_json, background, created_at, 5 more } Emitted when image generation has completed and the final image is available.
Emitted when image generation has completed and the final image is available.
Base64-encoded image data, suitable for rendering as an image.
background: "transparent" or "opaque" or "auto"The background setting for the generated image.
The background setting for the generated image.
The Unix timestamp when the event was created.
output_format: "png" or "webp" or "jpeg"The output format for the generated image.
The output format for the generated image.
quality: "low" or "medium" or "high" or "auto"The quality setting for the generated image.
The quality setting for the generated image.
size: "1024x1024" or "1024x1536" or "1536x1024" or "auto"The size of the generated image.
The size of the generated image.
The type of the event. Always image_generation.completed.
usage: object { input_tokens, input_tokens_details, output_tokens, total_tokens } For the GPT image models only, the token usage information for the image generation.
For the GPT image models only, the token usage information for the image generation.
The number of tokens (images and text) in the input prompt.
input_tokens_details: object { image_tokens, text_tokens } The input tokens detailed information for the image generation.
The input tokens detailed information for the image generation.
The number of image tokens in the input prompt.
The number of text tokens in the input prompt.
The number of image tokens in the output image.
The total number of tokens (images and text) used for the image generation.
ImageModel = "gpt-image-1.5" or "dall-e-2" or "dall-e-3" or 2 more
ImagesResponse = object { created, background, data, 4 more } The response from the image generation endpoint.
The response from the image generation endpoint.
The Unix timestamp (in seconds) of when the image was created.
background: optional "transparent" or "opaque"The background parameter used for the image generation. Either transparent or opaque.
The background parameter used for the image generation. Either transparent or opaque.
The list of generated images.
The list of generated images.
The base64-encoded JSON of the generated image. Returned by default for the GPT image models, and only present if response_format is set to b64_json for dall-e-2 and dall-e-3.
For dall-e-3 only, the revised prompt that was used to generate the image.
When using dall-e-2 or dall-e-3, the URL of the generated image if response_format is set to url (default value). Unsupported for the GPT image models.
output_format: optional "png" or "webp" or "jpeg"The output format of the image generation. Either png, webp, or jpeg.
The output format of the image generation. Either png, webp, or jpeg.
quality: optional "low" or "medium" or "high"The quality of the image generated. Either low, medium, or high.
The quality of the image generated. Either low, medium, or high.
size: optional "1024x1024" or "1024x1536" or "1536x1024"The size of the image generated. Either 1024x1024, 1024x1536, or 1536x1024.
The size of the image generated. Either 1024x1024, 1024x1536, or 1536x1024.
usage: optional object { input_tokens, input_tokens_details, output_tokens, 2 more } For gpt-image-1 only, the token usage information for the image generation.
For gpt-image-1 only, the token usage information for the image generation.
The number of tokens (images and text) in the input prompt.
input_tokens_details: object { image_tokens, text_tokens } The input tokens detailed information for the image generation.
The input tokens detailed information for the image generation.
The number of image tokens in the input prompt.
The number of text tokens in the input prompt.
The number of output tokens generated by the model.
The total number of tokens (images and text) used for the image generation.
output_tokens_details: optional object { image_tokens, text_tokens } The output token details for the image generation.
The output token details for the image generation.
The number of image output tokens generated by the model.
The number of text output tokens generated by the model.