{"data":{"id":"74e952d4-f6dd-483f-b91d-432ebb10f504","title":"The next evolution of the Agents SDK","summary":"OpenAI introduced new capabilities to the Agents SDK, a toolkit for developers building AI agents that can work with files and run commands on computers. The update includes a model-native harness (a framework optimized for OpenAI models) and native sandbox execution (a controlled, isolated computer environment where agents can safely run code and access files). The SDK aims to bridge the gap between flexibility and production-readiness by providing developers with standardized infrastructure that keeps agents aligned with how frontier models (the most advanced AI models available) work best.","solution":"The Agents SDK includes several built-in protections: 'Separating harness and compute helps keep credentials out of environments where model-generated code executes.' The SDK also supports 'built-in snapshotting and rehydration' so 'the Agents SDK can restore the agent's state in a fresh container and continue from the last checkpoint if the original environment fails or expires.' Additionally, developers can configure sandbox execution with 'Blaxel, Cloudflare, Daytona, E2B, Modal, Runloop, and Vercel' providers, and the SDK provides a 'Manifest abstraction for describing the agent's workspace' to control access to files and data.","labels":["industry"],"sourceUrl":"https://openai.com/index/the-next-evolution-of-the-agents-sdk","publishedAt":"2026-04-15T10:00:00.000Z","cveId":null,"cweIds":null,"cvssScore":null,"cvssSeverity":null,"severity":"info","attackType":[],"issueType":"news","affectedPackages":null,"affectedVendors":["OpenAI"],"affectedVendorsRaw":["OpenAI","Agents SDK"],"classifierModel":"claude-haiku-4-5-20251001","classifierPromptVersion":"v3","cvssVector":null,"attackVector":null,"attackComplexity":null,"privilegesRequired":null,"userInteraction":null,"exploitMaturity":null,"epssScore":null,"patchAvailable":null,"disclosureDate":"2026-04-15T10:00:00.000Z","capecIds":null,"crossRefCount":0,"attackSophistication":"moderate","impactType":["safety","integrity"],"aiComponentTargeted":"agent","llmSpecific":true,"classifierConfidence":0.92,"researchCategory":null,"atlasIds":null}}