Back to blog

OpenAI and Amazon announce strategic partnership

Sebastian Relard

Sebastian Relard

OpenAI and Amazon announce strategic partnership
OpenAIAmazonKI
relard.dev

The new AWS-OpenAI partnership brings OpenAI’s Frontier platform and a stateful runtime environment to Amazon Bedrock. For German companies, this means production-grade AI agents with memory, governance, and scalability directly within the existing AWS landscape. I’ll show how we use this pragmatically at Relard.dev and get to production in 90 days.

What’s changing and why it matters to you

OpenAI and Amazon have announced a multi-year collaboration that brings together three things I’ve long wanted as an integrator:

  • Frontier models via AWS: AWS will be the exclusive third-party provider for the cloud distribution of OpenAI Frontier. This gives enterprises official, integrated access to the most advanced OpenAI models inside their own AWS environment.
  • Stateful Runtime Environment in Bedrock: A stateful runtime, powered by OpenAI models, is coming to Amazon Bedrock and integrates with AgentCore as well as AWS infrastructure services. This runtime retains context and identity, accesses compute and tools, and works cleanly across data sources. It was explicitly built for ongoing projects and workflows.
  • Massive capacity and serious investment: OpenAI is using 2 gigawatts of Trainium capacity on AWS. Amazon is investing USD 50 billion in OpenAI, starting with 15 billion. To me, that’s a clear signal: capacity, roadmap, and enterprise focus are secured for the long term.

Concrete impact for German companies: We can build AI agents that are more than chatbots. Agents that work on cases over weeks, enforce permissions cleanly with IAM, keep data in the EU, talk to SAP or an on-prem ERP, and remember the entire history. And all that without a zoo of tinkered services, but natively in AWS plus the tools you already have.

In recent years, I’ve often run into exactly these limits with customers in DACH: session-less models, fragile memory hacks, compliance questions, latencies between cloud providers. The new combination of Bedrock AgentCore and a true stateful runtime removes many of these pains. At the same time, risks like vendor lock-in and cost blowouts remain. Both can be managed if architecture and governance are clean from day one.

Three use cases that now become practical

  1. Service agent with memory in B2B support
    Goal: An agent that triages tickets, asks follow-up questions, triggers appropriate steps in n8n or directly in AWS, and remains contextual across the entire customer relationship.

So I build it this way:

  • Data foundation: Customer records in an EU S3 bucket, metadata in Postgres or DynamoDB, knowledge indexing via OpenSearch or Amazon Kendra.
  • Agent: Bedrock AgentCore orchestrates capabilities such as ticket creation, RMA creation, knowledge search, and email reply. The new stateful runtime stores working context, roles, permissions, and open to-dos per case.
  • Security: Access via IAM roles, secrets in AWS Secrets Manager, VPC endpoints and PrivateLink for Bedrock. All actions are logged in CloudTrail and CloudWatch.
  • Automation: n8n bridges to legacy systems (on-prem ERP via OData, SMTP, SFTP), triggers workflows, and writes back status.
  • Human-in-the-loop: Approval steps for credits or goodwill cases remain with humans. The agent proposes, prepares documents, but finalizes only after approval.

Previously, I had to "re-prime" agents after every context loss. With the stateful environment, the case history is part of the runtime. That saves tokens, reduces errors, and makes interaction more natural.

  1. Technical sales assistant in mechanical engineering
    Goal: Quote configurations for complex products, bill-of-material checks, variant validation, order preparation.

So I build it this way:

  • RAG stack: S3 as the source of truth for datasheets, configuration rules as structured policies, vector search via OpenSearch.
  • Agent capabilities: Configure, plausibility check, fetch pricing logic, generate the quote as a PDF. Tool calls are strictly limited via IAM policies.
  • Process: The agent conducts the customer dialogue, creates a configuration, validates with domain rules, triggers ERP integration in n8n, and stores a draft quote.
  • Evaluation: Every new rule or product variant gets test prompts and regression checks. Cost control via Bedrock cost-center tags, AWS Budgets, and rate limits per sales team.

Here, Frontier via AWS especially helps with reasoning-intensive steps and multi-stage sequences. The runtime holds temporary intermediates instead of recomputing them each time.

  1. Document assistant for legal and compliance teams
    Goal: Classify contracts, compare clauses, flag risks, propose measures. Results remain audit-proof.

So I build it this way:

  • Pipeline: Upload to S3 with virus scan, PII redaction, OCR. Metadata in DynamoDB. Versioning on.
  • Agent: Uses Bedrock Guardrails for sensitive content, keeps case context in the stateful runtime, writes findings to an audit log.
  • Data residency: EU region, KMS-encrypted, access only from the corporate VPC. Optional Private CA for mTLS between services.
  • Governance: Every recommendation includes source references. Reviewer UI in a small web app (Next.js) with "Accept, Edit, Reject". Changes do not automatically train the model; they flow, curated, into rules or fine-tunes.

Especially here, explainable behavior is more important than mere accuracy. The combination of AgentCore, guardrails, and clean logging helps me build audit-ready processes.

Roadmap: From PoC to operations in 90 days

I use a clear, lean plan. The goal is a production-near core process, not a lab prototype.

  • Week 1 to 2: Discovery and security foundation
    • Define use-case scope, success criteria, and metrics.
    • Map data flows, mark PII and classified data.
    • Check AWS basics: accounts, VPC setup, IAM patterns, KMS keys, logging.
  • Week 3 to 5: Minimal agent on Bedrock
    • Start with a model available in Bedrock, later switch to Frontier via AWS as soon as it’s enabled.
    • RAG minimum: 200 to 500 representative documents, indexing, citation requirement.
    • Tooling: 2 to 3 core actions (e.g., create ticket, generate quote, write record).
  • Week 6 to 8: Integration and runtime
    • Connect to SAP or CRM, n8n flows for legacy.
    • Introduce the stateful runtime, store working context and identity, permission checks via IAM.
    • Guardrails, prompt contracts, telemetry with CloudWatch and structured logs.
  • Week 9 to 10: Evaluation, cost, hardening
    • Benchmarks on real cases, classify error types, automate regression tests.
    • Cost caps: budgets, alerts, token quotas, caching strategies.
    • Security review, light pen test, runbook and on-call.
  • Week 11 to 12: Pilot operations
    • Onboard pilot users, 2-week feedback cycle, go/no-go decision for rollout.
    • Scaling plan with Auto Scaling, maintainability, backups, disaster recovery.

In parallel, I prepare for using Frontier via AWS. Since AWS acts as the exclusive third-party provider, I design interfaces so that a model switch later requires only configuration.

Governance, data protection, and cost under control

  • Data protection and location: Data residency in EU regions, encryption with KMS, minimization principle via strict context filtering. For sensitive content I use Bedrock Guardrails and PII redaction before the model call.
  • Network and identities: PrivateLink for Bedrock, no open internet paths from production systems. Least-privilege roles per tool task, short-lived credentials, Secrets Manager for keys.
  • Audit and traceability: Every agent action is logged with inputs, tool calls, and results. For audits I export tamper-evident reports.
  • Cost control: Review architecture for cost before scaling. Measure unit economics per business case. Caching, batch processing, configurable response depth. AWS Budgets and Cost Explorer active, tags mandatory.
  • Lock-in risk: I encapsulate model calls behind an internal API and keep vector search and data storage neutral. That way a switch remains possible while I still leverage Bedrock integration.

My experience: The biggest cost traps are poor context strategy and missing rate limiting. Most quality issues stem from untested system prompts. I solve both with hard limits, evaluated prompt contracts, and an automated test set that runs on every change.

What the partnership specifically promises

I’m referring directly to the announcement: The stateful runtime environment is coming to Amazon Bedrock in the next few months and integrates with AgentCore. It lets models access compute, identity, tools, and data sources and keeps context across ongoing projects. AWS will be the exclusive third-party provider for the cloud distribution of OpenAI Frontier. OpenAI is booking 2 gigawatts of Trainium capacity on AWS. Amazon is investing USD 50 billion in OpenAI. For IT leaders, that’s a clear roadmap for enterprise AI in AWS. For me as an integrator, it’s an invitation to turn proofs into real products.

Conclusion

If you operate on AWS, now is the time to pilot AI agents with real responsibility. The combination of Frontier via AWS, Bedrock AgentCore, and the new stateful runtime delivers exactly the building blocks that have been missing in practice. My offer: We identify a value-creating use case, build a robust pilot in 90 days, and scale only when quality and cost line up.

Frequently Asked Questions

Share

AI Systems

Want AI in your app?

I build AI features with RAG, LLMs, and intelligent workflows. Production-ready, privacy-compliant, measurable.

Discuss AI Project