Back

developers.openai.com

OpenAI for Developers in 2025

Updated 3/30/2026
https://developers.openai.com/blog/openai-for-developers-2025

- **Reasoning became a core dial** and increasingly converged with general-purpose chat models. - **Multimodality (docs, audio, images, video)** became a first-class citizen in the API. - **Agent building blocks** (Responses API, Agents SDK, AgentKit) made multi-step workflows easier to ship and operate. … ## TL;DR - The big shift was **agent-native APIs** plus **better models** that can perform more complex tasks, requiring reasoning and tool use. - Codex matured across both models and tooling, pairing GPT-5.2-Codex’s repo-scale reasoning with a production-ready CLI, web, and IDE workflows for long-horizon coding tasks. … ### PDFs and documents - **PDF inputs** enabled document-heavy workflows directly in the API. - **PDF-by-URL** reduced friction by referencing documents without upload. **Why it matters:** you can now rely on the OpenAI platform for not only text & vision but also your image and video generation workflows as well as speech-to-speech use cases. … ... Beyond the CLI, Codex expanded support for longer sessions and iterative problem solving across the **web + cloud** and the **IDE extension**, tightening the loop between conversational reasoning and concrete code changes. Teams could also automate parts of the workflow with **Codex Autofix** in CI. **Why it matters:** by the end of 2025, Codex functioned less as “a model you prompt” and more as a coding surface–combining reasoning-capable models with tools developers already use. ## Platform shift: Responses API and agentic building blocks One of the most important platform changes in 2025 was the move toward **agent-native APIs**. The **Responses API** made it easier to build for the new generation of models: - Support for multiple inputs and outputs, including different modalities - Support for reasoning controls and summaries - Better support for tool calling, including during reasoning … ## Run and scale: async, events, and cost controls Once agents moved from “single request” to “multi-step jobs,” production teams needed primitives for cost, latency, and reliability. - **Prompt caching** reduced latency and input costs when prompts share long, repeated prefixes (system prompts, tools, schemas). - **Background mode** enabled long-running responses without holding a client connection open. - **Webhooks** turned “polling everything” into event-driven systems (batch completion, background completion, fine-tuning completion). - **Rate limits** and workload optimization guidance matured as usage tiers and model families expanded. … ## Evaluation, tuning, and shipping safely - **Evals API** for eval-driven development. - **Reinforcement fine-tuning (RFT)** using programmable graders. - **Supervised fine-tuning / distillation** for pushing quality down into smaller, cheaper models once you’ve validated a task with a larger one. - **Graders** and the **Prompt optimizer** helped teams run a tighter “eval → improve → re-eval” loop. ## Wrapping up Throughout 2025, we focused on a few consistent themes aimed at making it easier for developers to build and ship on our platform: - Scaled, controllable reasoning as a core capability - A unified, agent-native API surface - Open building blocks and emerging interoperability standards - Deep multimodal support across text, images, audio, video, and documents - Stronger production tooling for evaluation, tuning, and deployment