Announced Open Responses: an open-source spec for building multi-provider, interoperable LLM interfaces built on top of the original OpenAI Responses API. Feature gpt-5.2-codex v1/responses Released Update gpt-realtime-mini gpt-audio-mini Updated the gpt-realtime-mini and gpt-audio-mini slugs to point to the 2025-12-15 snapshots. If you need the previous model snapshots, use Updated the sora-2 slug to point to Update gpt-4o-mini-tts gpt-4o-mini-transcribe Updated the Fix gpt-image-1.5 chatgpt-image-latest Fixed an issue where Update gpt-image-1.5 chatgpt-image-latest Added Feature gpt-image-1.5 chatgpt-image-latest Feature gpt-realtime-mini gpt-audio-mini gpt-4o-mini-transcribe gpt-4o-mini-tts Released four new dated audio snapshots. These updates deliver reliability, quality, and voice fidelity improvements for real-time, voice-driven applications. Read more here. This launch also includes support for Custom voices for eligible customers. Feature gpt-5.2 gpt-5.2-chat-latest v1/responses v1/chat/completions Released GPT-5.2, the newest flagship model in the GPT-5 model family. GPT-5.2 shows improvements over the previous GPT-5.1 in: What's new in 5.2 is a new xhigh reasoning effort level, concise reasoning summaries, and new context management using compaction. Feature v1/responses/compact Released client-side compaction. For long-running conversations with the Responses API, you can use the Feature gpt-5.1-codex-max v1/responses Released Added support for DTMF key presses in the Realtime API. You can now receive DTMF events while using a Realtime sideband connection. See docs here for more information. Feature gpt-5.1 gpt-5.1-codex gpt-5.1-chat-latest gpt-5.1-codex-mini v1/responses v1/chat/completions Released GPT-5.1, the newest flagship model in the GPT-5 model family. GPT-5.1 is trained to be especially proficient in: Note that GPT-5.1 defaults to a new Feature gpt-5.1-codex gpt-5.1-codex-mini v1/responses Released Released extended prompt cache retention. Extended prompt cache retention keeps cached prefixes active for longer, up to a maximum of 24 hours. Extended Prompt Caching works by offloading the key/value tensors to GPU-local storage when memory is full, significantly increasing the storage capacity available for caching. Feature gpt-oss-safeguard-120b gpt-oss-safeguard-20b gpt-oss-safeguard-120b and gpt-oss-safeguard-20b are safety reasoning models built-upon gpt-oss. Read more here. Released Enterprise Key Management (EKM). Enterprise Key Management (EKM) allows you to encrypt your customer content at OpenAI using keys managed by your own external Key Management System (KMS). Feature gpt-5-pro gpt-realtime-mini gpt-audio-mini gpt-image-1-mini sora-2 sora-2-pro v1/responses v1/batch v1/chat/completions v1/videos v1/realtime v1/images/generations Released IP allowlist. IP allowlisting restricts API access to only the IP addresses or ranges you specify. Feature gpt-5-codex v1/responses Launched special-purpose model gpt-5-codex, built and optimized for use with the Codex CLI. Added support for connectors to the Responses API. Connectors are OpenAI-maintained MCP wrappers for popular services like Google apps, Dropbox, and more that can be used to give model read access to data stored in those services. Feature v1/conversations v1/responses v1/assistants Released the Conversations API, which allows you to create and manage long-running conversations with the Responses API. See the migration guide to see a side-by-side comparison and learn how to migrate from an Assistants API integration to Responses and Conversations. Feature v1/chat/completions v1/responses Released GPT-5 family of models in the API, including Introduced the Introduced Launched support for Priority processing. Priority processing delivers significantly lower and more consistent latency compared to Standard processing while keeping pay-as-you-go flexibility. Feature o3-deep-research o3-deep-research-2025-06-26 o4-mini-deep-research o4-mini-deep-research-2025-06-26 v1/responses New reusable prompts are now available in the dashboard and Responses API. Via API, you can now reference templates created in the dashboard via the Feature o3-pro v1/responses v1/batch Released o3-pro, a version of the o3 reasoning model that uses more compute to answer hard problems with better reasoning and consistency. Prices for the o3 model have also been reduced for all API requests, including batch and flex processing. Added fine-tuning support with direct preference optimization for the models Feature v1/chat/completions v1/realtime Feature v1/responses v1/chat/completions Added support for using Feature codex-mini-latest v1/responses v1/chat/completions Feature v1/fine-tuning v1/responses v1/chat/completions Feature v1/images/generations v1/images/edits Added a new image generation model, Updated the Image Generation and Edit endpoints to support new parameters specific to the Feature v1/chat/completions v1/responses Added two new o-series reasoning models, Launched Codex, our code generation CLI tool. Feature gpt-4.1 gpt-4.1-mini gpt-4.1-nano v1/responses v1/chat/completions v1/fine_tuning Added Added Feature o1-pro v1/responses v1/batch Released o1-pro, a version of the o1 reasoning model that uses more compute to answer hard problems with better reasoning and consistency. Feature gpt-4o-search-preview gpt-4o-mini-search-preview computer-use-preview v1/chat/completions v1/assistants v1/responses Released several new models and tools and a new API for agentic workflows: Feature v1/fine_tuning/jobs Added Feature GPT-4.5 v1/chat/completions v1/assistants v1/batch Released a research preview of GPT-4.5—our largest and most capable chat model yet. GPT-4.5's high "EQ" and understanding of user intent make it better at creative tasks and agentic planning. Launched the API Usage Dashboard Update. This update addresses requests for additional data filters, such as project selection, date picker, and fine-grained intervals. There’s also better support for viewing usage across different products and service tiers. Introducing data residency in Europe. Read more here. Feature o3-mini o3-mini-2025-01-31 v1/chat/completions Launched o3-mini, a new small reasoning model that is optimized for science, math, and coding tasks. Expanded access to o1 model. The o1 series of models are trained with reinforcement learning to perform complex reasoning. Launched Admin API Key Rotations, enabling customers to programmatically rotate their admin api keys. Updated Admin API Invites, enabling customers to programmatically invite users to projects at the same time they are invited to organizations. Feature o1 gpt-4o gpt-4o-mini v1/fine_tuning v1/chat/completions v1/realtime Launched Usage API, enabling customers to programmatically query activities and spending across OpenAI APIs. Update v1/chat/completions Feature v1/chat/completions Released Predicted Outputs, which greatly reduces latency for model responses where much of the response is known ahead of time. This is most common when regenerating the content of documents and code files with only minor changes. Feature gpt-4o-realtime-preview gpt-4o-audio-preview v1/chat/completions Feature gpt-4o-audio-preview v1/chat/completions Feature v1/realtime v1/chat/completions v1/fine_tuning Released several new features at OpenAI DevDay in San Francisco: Realtime API: Build fast speech-to-speech experiences into your applications using a WebSockets interface. Model distillation: Platform for fine-tuning cost-efficient models with your outputs from a large frontier model. Image fine-tuning: Fine-tune GPT-4o with images and text to improve vision capabilities. Evals: Create and run custom evaluations to measure model performance on specific tasks. Prompt caching: Discounts and faster processing times on recently seen input tokens. Generate in playground: Easily generate prompts, function definitions, and structured output schemas in the playground using the Generate button. Feature omni-moderation-latest v1/moderations Released new Feature o1-preview o1-mini v1/chat/completions Released o1-preview and o1-mini, new large language models trained with reinforcement learning to perform complex reasoning tasks. Feature gpt-4o v1/fine_tuning Update gpt-4o v1/chat/completions Launched Admin and Audit Log APIs, allowing customers to programmatically administer their organization and monitor changes using the audit logs. Audit logging must be enabled within settings. Launched self-serve SSO configuration, allowing Enterprise customers on custom and unlimited billing to set up authentication against their desired IDP. Released GPT-4o mini, our affordable an intelligent small model for fast, lightweight tasks. Released Uploads to upload large files in multiple parts. Released GPT-4o in the API. GPT-4o is our fastest and most affordable flagship model. Added a new endpoint to delete a message from a thread in the Assistants API. Introduced project based hierarchy for organizing work by projects, including the ability to create API keys and manage rate and cost limits on a per-project basis (cost limits available only for Enterprise customers). Added support for streaming in the Assistants API Released embedding V3 models and an updated GPT-4 Turbo preview Added January, 2026
gpt-5.2-codex to the Responses API. GPT-5.2-Codex is a version of GPT-5.2 optimized for agentic coding tasks in Codex or similar environments. Read more here.gpt-realtime-mini-2025-10-06 and gpt-audio-mini-2025-10-06.sora-2-2025-12-08. If you need the previous model snapshot, use sora-2-2025-10-06.gpt-4o-mini-tts and gpt-4o-mini-transcribe slugs to point to the 2025-12-15 snapshots. If you need the previous model snapshots, use gpt-4o-mini-tts-2025-03-20 and gpt-4o-mini-transcribe-2025-03-20. We currently recomend using gpt-4o-mini-transcribe over gpt-4o-transcribe for the best results.gpt-image-1.5 and chatgpt-image-latest were incorrectly using high fidelity for image edits through /v1/images/edits, even when fidelity was explicitly set to low (the default).December, 2025
gpt-image-1.5 and chatgpt-image-latest to the Responses API image generation tool.
/responses/compact endpoint to shrink the context you send with each turn.gpt-5.1-codex-max to the Responses API. GPT-5.1-Codex is our most intelligent coding model optimized for long-horizon, agentic coding tasks. Read more here.November, 2025
none reasoning setting for faster responses when less thinking's required—different from the previous medium default setting in GPT-5.gpt-5.1-codex and gpt-5.1-codex-mini to the Responses API. GPT-5.1-Codex is a version of GPT-5.1 optimized for agentic coding tasks in Codex or similar environments. Read more here.October, 2025
September, 2025
August, 2025
gpt-5, gpt-5-mini, and gpt-5-nano.minimal reasoning effort value to optimize for fast responses in GPT-5 models (which support reasoning).custom tool call type, which allows for freeform inputs to and outputs from the model when tool calling.June, 2025
prompt parameter (with a prompt id, optional version) and supply dynamic variables that can include strings, images, or file inputs. Reusable prompts are not available in Chat Completions. Learn more.gpt-4.1-2025-04-14, gpt-4.1-mini-2025-04-14, and gpt-4.1-nano-2025-04-14.May, 2025
strict mode for tool schemas when using parallel tool calling with non-fine-tuned models.
Added new schema features, including string validation for email and other patterns and specifying ranges for numbers and arrays.April, 2025
gpt-image-1. This model sets a new standard for image generation, with improved quality and instruction following.gpt-image-1 model.o3 and o4-mini. They set a new standard for math, science, and coding, visual reasoning tasks, and technical writing.gpt-4.1, gpt-4.1-mini, and gpt-4.1-nano models to the API. These new models feature improved instruction following, coding, and a larger context window (up to 1M tokens). gpt-4.1 and gpt-4.1-mini are available for supervised fine-tuning. Announced deprecation of gpt-4.5-preview.March, 2025
gpt-4o-mini-tts, gpt-4o-transcribe, gpt-4o-mini-transcribe, and whisper-1 models to the Audio API.
gpt-4o-search-preview, gpt-4o-mini-search-preview, computer-use-preview.metadata field support to fine-tuning jobs.February, 2025
January, 2025
December, 2024
November, 2024
October, 2024
September, 2024
omni-moderation-latest moderation model, which supports both images and text (for some categories), supports two new text-only harm categories, and has more accurate scores.August, 2024
July, 2024
June, 2024
May, 2024
April, 2024
March, 2024
February, 2024
January, 2024
dimensions parameter to the Embeddings APIDecember, 2023
November, 2023
October, 2023