June was a big one.

We launched Model Catalog, one of our most requested features from enterprise users, and something we’ve spent the last few weeks building, testing, and refining. It’s now the central control plane for managing every model your teams use through Portkey.

Along with that, we rolled out several core updates across the platform and gateway including OpenAI Agents TS SDK support, circuit breaker, global endpoints for Vertex AI, expanded Azure coverage, and more.

June also saw increased GitHub momentum as we rolled out one of our largest infrastructure upgrades to date!

Here’s everything that went live in June.

Summary

AreaKey Highlights
Platform• Model Catalog launch
• OpenAI background mode
• Circuit breaker config
Gateway & Providers• Support for Sutra, o3-pro, Magistral, and Gemini 2.5 models
• Vertex AI global endpoints
• Anthropic Computer Use tool support
• Support for additional Azure OpenAI endpoints
• Bedrock Inference Profiles
• Prompt caching for tools (Anthropic)
Integrations• OpenAI Agent SDK (TypeScript)
• Langroid native support
• Strands SDK & ADK integration
• Cursor integration
• Gemini CLI support

🎉 Introducing the Model Catalog 🎉

Take control of which models your teams can use across every provider, from one place. With Model Catalog, you can:

  • Manage access to 1,600+ models across OpenAI, Anthropic, Azure, and more
  • Decide which models are available to which teams
  • Set rate limits and budgets per workspace
  • Automatically enable new models as they’re released

No more scattered configuration or manual provisioning. Just clean, centralized governance for all your AI usage.

To enable it in your organizattion, reach out to us at [email protected]

Circuit breaker

When a model starts failing or slowing down, Portkey can now temporarily route traffic to fallback targets — then automatically restore routing once things stabilize. It’s a smarter way to handle failovers without manual intervention. Available as a config option. Learn more ->

OpenAI Agent SDK (TypeScript)

With the OpenAI Agents SDK (TS), building agents in TypeScript just got easier - no Python, no context switching, just native tooling and agent workflows.

But turning those prototypes into production-ready systems reveals real gaps:

  • No logging or tracing
  • No retries or failover
  • No cost visibility or access control
  • No simple way to switch LLM providers

That’s where Portkey comes in. Route your calls through Portkey and unlock observability, guardrails, prompt versioning, and multi-provider support instantly. Learn more ->

Strands Agents SDK

Enterprises building on Strands Agents SDK now have a cleaner path to production — thanks to Portkey as the LLM gateway. With a single abstraction layer, you can:

  • Access 2,000+ LLMs with provider-agnostic logic
  • Standardize tool calling across providers
  • Add conditional routing for cost control
  • Enable retries, rate limits, and full observability

Huge thanks to Federico Kamelhar for leading the integration effort and bringing Portkey support to Strands. If you love Strands & Portkey contribute to this PR and help us stabilize this integration.

Gateway & Providers

AWS Bedrock inference profiles

You can now route Bedrock requests by inference profile, allowing access to multiple regions or compute configurations through a single virtual key. Learn more -> .

Vertex AI global endpoint support

Portkey helps you improve model availability and reduce 429 (rate limit) errors with support for Vertex AI’s global endpoints. You can now set region = global in your Vertex AI Virtual Key config to automatically access Google’s distributed infrastructure — no manual region selection needed.

OpenAI Background Mode

Reasoning models can take minutes to solve complex problems. With background mode, you can now run long-running tasks on models like o3-pro and o1-pro reliably, without worrying about timeouts or dropped connections.

Portkey now supports background mode for OpenAI requests. Simply pass background:True as a parameter, and Portkey will handle the rest.

Anthropic’s Computer Use

Experiment confidently with Anthropic’s new Computer Use tool by adding observability, fallback logic, and cost controls from day one. Learn more ->

Support for additional Azure OpenAI endpoints

Portkey now supports a wider range of Azure OpenAI endpoints including image generation, audio transcription, speech synthesis, file management, and batch operations.

Anthropic prompt caching (tools)

Anthropic tool-based interactions now benefit from prompt caching in Portkey, improving performance and reducing token use for repeated tool calls.

To keeping up the pace!

New models and providers

  • Sutra: Multilingual LLM with standout MMLU scores in Hindi & Gujarati
  • Magistral: Mistral’s first reasoning model, built for multilingual and domain-specific logic
  • o3-pro: OpenAI’s latest flagship model with strong reasoning and fast response times
  • Gemini 2.5 (Flash, Pro, Flash-Lite): Now GA and supported with full observability
  • Kluster AI: Claude-compatible, MCP-enabled models optimized for low latency and high availability.
  • Hyperbolic AI: OpenAI-compatible models focused on cost efficiency and speed for production-scale usage.
  • Featherless AI: Serverless access to Hugging Face models with a lightweight setup.
  • Groq: Support for service_tier flag added

Integrations

OpenAI Agent SDK (TypeScript)

Portkey now works with OpenAI’s new TypeScript Agents SDK. Add retries, observability, rate limits, and multi-provider support to your agent flows, without changing core logic.

Langroid

Langroid now supports Portkey natively — plug it in to get multi-provider routing, observability, fallback logic, and guardrails in your agentic Python apps.

Cursor

Bring visibility and governance to Cursor’s coding assistant with Portkey. Choose from 1600+ models, track requests, enforce rate limits, and log usage, making it easy for enterprises to govern Cursor across the org.

Agent Development Kit (ADK)

Portkey now integrates with Google’s Agent Development Kit (ADK), bringing production-grade features like retries, fallback logic, and observability to ADK-based agents.

Langflow

Add enterprise-grade features to your Langflow workflows with Portkey — including unified model access, full observability, usage governance, and security guardrails.

Partnerships

  • Prompt Security – Secure every prompt and response in real time by embedding Prompt Security directly into Portkey’s AI Gateway. Read more here

  • Lasso Security – Combine infra-level controls and real-time behavioral monitoring to secure the entire LLM lifecycle — from access to output. Read more here

  • FutureAGI – Use Portkey as the control layer and FutureAGI as the eval layer to automate output scoring across all model traffic. See how you can implement this

  • Arize AI – Connect Portkey’s routing and guardrails with Arize’s observability to monitor model drift, latency, cost, and quality in one flow. Read more here

Portkey Live!

In partnership with Pangea, we hosted a live webinar on how to build scalable, secure GenAI infrastructure. Catch the replay here!.

Improvements

  • Renamed Model Whitelist to Allowed Models for clarity and consistency
  • Improved error responses for webhook failures, making them easier to debug and handle programmatically

Teams love Portkey!

If you love Portkey, drop a ⭐ on GitHub

Resources

Docs are now open for contributions!

We’re opening up our documentation for contributions. If you’ve found parts that could be clearer, better explained, or just more complete, we’d truly appreciate your help. Every suggestion, edit, or fix helps make Portkey better for the whole community. See how you can contribute

Community Contributors

A special thanks to our community contributors this month:

Coming this month!

Struggling with unauthorized tool usage in MCP? Portkey is about to solve that. Stay tuned.

Support