OpenAI equips the Responses API with shell, containers, and compaction for production agents
Original: From model to agent: Equipping the Responses API with a computer environment View original →
OpenAI's March 11 engineering post explains how the company is turning the Responses API from a model interface into an execution environment for agents. The core idea is to combine the API with a shell tool and hosted containers so a model can propose actions, inspect results, and continue a tool loop without developers building a custom orchestration layer from scratch.
According to the post, the setup addresses practical agent problems that do not fit neatly inside prompt-only systems: handling intermediate files, querying structured data, running services, making API requests, and managing retries or timeouts. OpenAI says the platform runs model-proposed commands in an isolated workspace with a filesystem, optional SQLite storage, and restricted network access. Models GPT-5.2 and later are trained to propose shell commands for this flow.
The operational detail matters. The Responses API can stream shell output back to the model in near real time, execute multiple shell sessions concurrently, and cap overly large outputs so long tool traces do not consume the whole context window. OpenAI also added native compaction, which produces a token-efficient summary when long-running tasks approach context limits, letting workflows continue across many steps without custom summarization logic from the developer.
Security and workflow controls
- Hosted containers use a sidecar egress proxy so outbound requests go through allowlists and access controls.
- Secrets are injected at egress on an approved-domain basis, meaning raw credentials stay outside model-visible context.
- OpenAI recommends using container files and databases, rather than stuffing large tables directly into prompts.
The post matters because it shows where OpenAI thinks agent infrastructure is heading: not only better models, but a managed runtime that can keep state, execute tools, and stay inside security boundaries. For teams building agentic products, the Responses API is being positioned as a higher-level operating layer rather than a simple text completion interface.
Related Articles
OpenAI Developers said on March 21, 2026 that container startup for skills, hosted shell, and code interpreter was about 10x faster via a new container pool in the Responses API. Updated OpenAI shell docs show hosted shell can create containers automatically, reuse active containers by reference, and keep them alive for 20 minutes of inactivity.
OpenAI Developers published a March 11, 2026 engineering write-up explaining how the Responses API uses a hosted computer environment for long-running agent workflows. The post centers on shell execution, hosted containers, controlled network access, reusable skills, and native compaction for context management.
Perplexity said on March 11, 2026 that its new Agent API combines search, tool execution, and multi-model orchestration behind one managed runtime. The launch positions Perplexity less as a single-answer interface and more as infrastructure for production agent workflows.
Comments (0)
No comments yet. Be the first to comment!