Skip to main content
Portkey has been recognized as one of the 2025 Gartner® Cool Vendors™ in LLM Observability, highlighting our work in bringing visibility, reliability, and governance to production AI systems. Alongside this recognition, we shipped major platform updates including Terraform support, enhanced rate limiting and timeout controls, new security guardrails, and improved observability. We’re also preparing for a series of upcoming conversations and community sessions focused on building and governing AI responsibly in production. And as always, we continue to learn from how our customers use Portkey, helping us refine the platform for production AI everywhere.

Summary

AreaKey Highlights
Platform• Terraform Provider for Portkey
• Configurable request timeouts
• Enhanced rate limits for tokens and requests
Agents• OpenAI AgentKit with multi-provider support
Guardrails• Javelin AI Security integration
• Add Prefix guardrail
• Allowed Request Types guardrail
Gateway & Providers• Claude Haiku 4.5
• Expanded provider updates (OpenAI, Azure, Bedrock, Vertex AI)
Observability• Structured tool-call visibility
• Enhanced web search view for Anthropic

Portkey recognized as a 2025 Gartner® Cool Vendor™ in LLM Observability

2025 Gartner® Cool Vendor™ Portkey has been recognized as one of the 2025 Gartner® Cool Vendors™ in LLM Observability, in the report by Padraig Byrne, Tanmay Bisht, and Andre Bridges. This recognition reinforces our mission to help teams move from experimentation to production with confidence and observability built in. If you like what we are building, please drop us a review here, it’d mean a lot to us!

What industry leaders are telling about us!

anthropic

Platform

Terraform Provider for Portkey

Portkey now has a Terraform provider for managing workspaces, users, and organization resources through the Portkey Admin API. This enables you to manage:
  • Workspaces: Create and update workspaces for teams and projects
  • Members: Assign users to workspaces with defined roles
  • Access: Send user invites with organization and workspace access
  • Users: Query and manage existing users in your organization

Enhanced rate limits

You can now configure rate limits for both requests and tokens under each API key, giving teams precise control over workload, costs, and performance across large deployments. Enhanced rate limits
  • Request-based limits: Cap the number of requests per minute, hour, or day.
  • Token-based limits: Cap the number of tokens consumed per minute, hour, or day.

Configurable request timeouts

A new REQUEST_TIMEOUT environment variable lets you control how long the Gateway waits before timing out outbound LLM requests — helping fine-tune latency, retries, and reliability for large-scale workloads.

Customer love!

Enhanced rate limits

Guardrails

Javelin AI Security integration

You can now configure Javelin AI Security guardrails directly in the Portkey Gateway to evaluate every model interaction for:
  • Trust & Safety — detect and filter harmful or unsafe content
  • Prompt Injection Detection — identify attempts to manipulate model behavior
  • Language Detection — verify and filter language in user or model responses
These guardrails extend Portkey’s security layer, making it easier to enforce safe and compliant model use at scale. See how to set up Javelin guardrails here.

Add Prefix guardrail

Add Prefix guardrail The new Add Prefix guardrail lets you automatically prepend a configurable prefix to user inputs before sending them to the model, useful for enforcing consistent tone, context, or compliance instructions across all model interactions.

Allowed Request Types guardrail

Allowed Request Types guardrail With the Allowed Request Types guardrail, you can now define which request types (endpoints) are permitted through the Gateway. Use an allowlist or blocklist approach to control access at a granular level and restrict unwanted request types.

Gateway

New models and providers

  • Claude Haiku 4.5: Anthropic’s lightweight model for real-time use cases.
  • OpenAI & Azure OpenAI: Added support for streaming audio transcription.
  • Vertex AI: Added custom-model support for batching and input-audio parameter.
  • AWS Bedrock: Added global profiles, token counting, and batch pass-throughs.
  • Azure: Improved Entra caching and image-editing pricing support.
  • Anthropic Beta (Vertex AI): Added beta header support for compatibility.

Agents

OpenAI AgentKit with multi-provider support

Building agents is now easier with OpenAI’s AgentKit, Portkey takes it further. You can now connect your AgentKit agents to 1,600+ LLMs across providers through Portkey, while gaining end-to-end observability, guardrails, and cost tracking. See the integration guide here

Observability

Structured tool call visibility

Structured tool call visibility Tool calls made during model interactions are now automatically detected and displayed as separate, structured entries in the Portkey logs.
This makes it easier to trace the tools invoked, inspect parameters, and understand agent behavior across multi-step workflows.
Enhanced view for web search Web search results from Anthropic models now appear in a clearer, formatted view within the dashboard.
This update improves readability and makes it easy to see what information the model used to generate its response, enabling faster analysis and better transparency.

Customer Stories: From arm pain to AI gateway

After an unexpected setback forced Rahul Bansal to rethink how he worked, he built Dictation Daddy, an AI-powered dictation platform now used by professionals, doctors, and lawyers to write faster and with greater accuracy, powered by Portkey. Read the full story here

Community & Events

MCP Gateway for Higher Education

MCP Gateway for Higher Education We’re teaming up with Internet2 to host a session on how higher education institutions can implement the Model Context Protocol (MCP) securely and at scale. Join us to learn how MCP fits within existing campus IT frameworks, what governance models are emerging, and the practical steps universities can take to enable secure, compliant AI access across departments.

LibreChat in Production

Regex Replace Guardrail LibreChat makes it easy to build chat-based AI experiences but adding governance, access control, and observability is key to running it in production. Join us to see how Portkey brings budgets, RBAC, and usage visibility to LibreChat, while connecting to 1,600+ LLMs across providers.

Resources

Community Contributors

A special thanks to our contributors this month:TensorNull, miuosz, marsianin, and uc4w6c

Support