Feature gpt-image-1.5 chatgpt-image-latest Feature gpt-realtime-mini gpt-audio-mini gpt-4o-mini-transcribe gpt-4o-mini-tts Released four new dated audio snapshots. These updates deliver reliability, quality, and voice fidelity improvements for real-time, voice-driven applications:
gpt-realtime-mini-2025-12-15
gpt-audio-mini-2025-12-15
gpt-4o-mini-transcribe-2025-12-15
gpt-4o-mini-tts-2025-12-15 Read more here. Feature gpt-5.2 gpt-5.2-chat-latest v1/responses v1/chat/completions Released GPT-5.2, the newest flagship model in the GPT-5 model family. GPT-5.2 shows improvements over the previous GPT-5.1 in: What's new in 5.2 is a new xhigh reasoning effort level, concise reasoning summaries, and new context management using compaction. Feature gpt-5.1 gpt-5.1-codex gpt-5.1-chat-latest gpt-5.1-codex-mini v1/responses v1/chat/completions Released GPT-5.1, the newest flagship model in the GPT-5 model family. GPT-5.1 is trained to be especially proficient in: Note that GPT-5.1 defaults to a new Feature gpt-5-pro gpt-realtime-mini gpt-audio-mini gpt-image-1-mini sora-2 sora-2-pro v1/responses v1/batch v1/chat/completions v1/videos v1/realtime v1/images/generations Feature gpt-5-codex v1/responses Launched special-purpose model gpt-5-codex, built and optimized for use with the Codex CLI. Added support for connectors to the Responses API. Connectors are OpenAI-maintained MCP wrappers for popular services like Google apps, Dropbox, and more that can be used to give model read access to data stored in those services. Feature v1/conversations v1/responses v1/assistants Released the Conversations API, which allows you to create and manage long-running conversations with the Responses API. See the migration guide to see a side-by-side comparison and learn how to migrate from an Assistants API integration to Responses and Conversations. Feature v1/chat/completions v1/responses Released GPT-5 family of models in the API, including Introduced the Introduced Feature o3-deep-research o3-deep-research-2025-06-26 o4-mini-deep-research o4-mini-deep-research-2025-06-26 v1/responses New reusable prompts are now available in the dashboard and Responses API. Via API, you can now reference templates created in the dashboard via the Feature o3-pro v1/responses v1/batch Released o3-pro, a version of the o3 reasoning model that uses more compute to answer hard problems with better reasoning and consistency. Prices for the o3 model have also been reduced for all API requests, including batch and flex processing. Added fine-tuning support with direct preference optimization for the models Feature v1/chat/completions v1/realtime Feature v1/responses v1/chat/completions Added support for using Feature codex-mini-latest v1/responses v1/chat/completions Feature v1/fine-tuning v1/responses v1/chat/completions Feature v1/images/generations v1/images/edits Added a new image generation model, Updated the Image Generation and Edit endpoints to support new parameters specific to the Feature v1/chat/completions v1/responses Added two new o-series reasoning models, Launched Codex, our code generation CLI tool. Feature gpt-4.1 gpt-4.1-mini gpt-4.1-nano v1/responses v1/chat/completions v1/fine_tuning Added Added Feature o1-pro v1/responses v1/batch Released o1-pro, a version of the o1 reasoning model that uses more compute to answer hard problems with better reasoning and consistency. Feature gpt-4o-search-preview gpt-4o-mini-search-preview computer-use-preview v1/chat/completions v1/assistants v1/responses Released several new models and tools and a new API for agentic workflows: Feature v1/fine_tuning/jobs Added Feature GPT-4.5 v1/chat/completions v1/assistants v1/batch Released a research preview of GPT-4.5—our largest and most capable chat model yet. GPT-4.5's high "EQ" and understanding of user intent make it better at creative tasks and agentic planning. Feature o3-mini o3-mini-2025-01-31 v1/chat/completions Launched o3-mini, a new small reasoning model that is optimized for science, math, and coding tasks. Launched Admin API Key Rotations, enabling customers to programmatically rotate their admin api keys. Updated Admin API Invites, enabling customers to programmatically invite users to projects at the same time they are invited to organizations. Feature o1 gpt-4o gpt-4o-mini v1/fine_tuning v1/chat/completions v1/realtime Launched Usage API, enabling customers to programmatically query activities and spending across OpenAI APIs. Update v1/chat/completions Feature v1/chat/completions Released Predicted Outputs, which greatly reduces latency for model responses where much of the response is known ahead of time. This is most common when regenerating the content of documents and code files with only minor changes. Feature gpt-4o-realtime-preview gpt-4o-audio-preview v1/chat/completions Feature gpt-4o-audio-preview v1/chat/completions Feature v1/realtime v1/chat/completions v1/fine_tuning Released several new features at OpenAI DevDay in San Francisco: Realtime API: Build fast speech-to-speech experiences into your applications using a WebSockets interface. Model distillation: Platform for fine-tuning cost-efficient models with your outputs from a large frontier model. Image fine-tuning: Fine-tune GPT-4o with images and text to improve vision capabilities. Evals: Create and run custom evaluations to measure model performance on specific tasks. Prompt caching: Discounts and faster processing times on recently seen input tokens. Generate in playground: Easily generate prompts, function definitions, and structured output schemas in the playground using the Generate button. Feature omni-moderation-latest v1/moderations Released new Feature o1-preview o1-mini v1/chat/completions Released o1-preview and o1-mini, new large language models trained with reinforcement learning to perform complex reasoning tasks. Feature gpt-4o v1/fine_tuning Update gpt-4o v1/chat/completions Launched Admin and Audit Log APIs, allowing customers to programmatically administer their organization and monitor changes using the audit logs. Audit logging must be enabled within settings. Launched self-serve SSO configuration, allowing Enterprise customers on custom and unlimited billing to set up authentication against their desired IDP. Released GPT-4o mini, our affordable an intelligent small model for fast, lightweight tasks. Released Uploads to upload large files in multiple parts. Released GPT-4o in the API. GPT-4o is our fastest and most affordable flagship model. Added a new endpoint to delete a message from a thread in the Assistants API. Introduced project based hierarchy for organizing work by projects, including the ability to create API keys and manage rate and cost limits on a per-project basis (cost limits available only for Enterprise customers). Added support for streaming in the Assistants API Released embedding V3 models and an updated GPT-4 Turbo preview Added December, 2025
November, 2025
none reasoning setting for faster responses when less thinking's required—different from the previous medium default setting in GPT-5.October, 2025
September, 2025
August, 2025
gpt-5, gpt-5-mini, and gpt-5-nano.minimal reasoning effort value to optimize for fast responses in GPT-5 models (which support reasoning).custom tool call type, which allows for freeform inputs to and outputs from the model when tool calling.June, 2025
prompt parameter (with a prompt id, optional version) and supply dynamic variables that can include strings, images, or file inputs. Reusable prompts are not available in Chat Completions. Learn more.gpt-4.1-2025-04-14, gpt-4.1-mini-2025-04-14, and gpt-4.1-nano-2025-04-14.May, 2025
strict mode for tool schemas when using parallel tool calling with non-fine-tuned models.
Added new schema features, including string validation for email and other patterns and specifying ranges for numbers and arrays.April, 2025
gpt-image-1. This model sets a new standard for image generation, with improved quality and instruction following.gpt-image-1 model.o3 and o4-mini. They set a new standard for math, science, and coding, visual reasoning tasks, and technical writing.gpt-4.1, gpt-4.1-mini, and gpt-4.1-nano models to the API. These new models feature improved instruction following, coding, and a larger context window (up to 1M tokens). gpt-4.1 and gpt-4.1-mini are available for supervised fine-tuning. Announced deprecation of gpt-4.5-preview.March, 2025
gpt-4o-mini-tts, gpt-4o-transcribe, gpt-4o-mini-transcribe, and whisper-1 models to the Audio API.
gpt-4o-search-preview, gpt-4o-mini-search-preview, computer-use-preview.metadata field support to fine-tuning jobs.February, 2025
January, 2025
December, 2024
November, 2024
October, 2024
September, 2024
omni-moderation-latest moderation model, which supports both images and text (for some categories), supports two new text-only harm categories, and has more accurate scores.August, 2024
July, 2024
June, 2024
May, 2024
April, 2024
March, 2024
February, 2024
January, 2024
dimensions parameter to the Embeddings APIDecember, 2023
November, 2023
October, 2023