Product
October '24
🎃 🪔 Portkey in October
October was packed with treats (no tricks!) for Portkey. As we celebrate Halloween and Diwali, we’re lighting up your AI infrastructure with some exciting updates. Let’s dive in!
Highlights
- Guardrails are now GA: Our guardrails feature is now generally available, helping you enforce LLM behavior in realtime. (Docs)
- Enterprise Updates: Refreshed our enterprise offering and welcomed one of the world’s largest tech companies to the Portkey family.
- Featured in Media: Check out our why we’re building DevOps for AI in the People+AI Newsletter and our CEO’s interview on Pulse2.
- Anthropic Prompt Caching in Playground: For Anthropic models, you can now enable any message to be cached right on the prompt playground.
- Portkey Tops Agent Ops Tooling Benchmark: Portkey provides 11 critical capabilities to put agents in production. Link
Features
- AWS Auth: AWS Assume Role Support for Bedrock. (Docs)
- Resend User Invite API: Launched a new API to re-invite a user to your org. (Docs)
- New API Specs: We also added API specs for: Prompt Completions API, Prompt Render API, and Inserting Logs API
- .NET (C#) Support: See how you can integrate Portkey in your .NET app easily using the OpenAI library and get advanced monitoring, routing, and enterprise features. (Docs)
- New OpenAI Param: OpenAI’s
max_completion_tokens
is now supported - Calculating Cached Requests:Updated cost calculations for cached responses for both OpenAI & Azure OpenAI.
- JSON Mode: We now support JSON mode for Gemini models.
- Controlled Generations: Gemini Controlled Generations is now supported on Portkey (alnog with support for Pydantic!)
- Bedrock Converse API: We’ve integrated Bedrock’s Converse API for all
/chat/completions
requests
Providers
We added 6 new providers on the Gateway this month!
Models
In true multimodal fashion, we added the all new SDv3 model across a bunch of providers, as well as Google’s Imagen model, and more!
Guardrails
As we continued to develop more Guardrails
Lowercase Detection
Check if the given string is lowercase or not.
Custom Webhooks
Along with the Webhook information, you can now send any custom metadata along with your request
LLM-based Guardrails
Portkey’s LLM-based Guardrails are now updated. You can do:
- PII Detection
- Language Detection
- Moderation
- Gibberish Detection
- PII Detection
- Language Detection
- Moderation
- Gibberish Detection
Integrations
And also pushed out some really important integrations that help everyone from IT Admins to SDEs, to Indie hackers!
LibreChat for Portkey
Tim wrote up a way to send unique user IDs from LibreChat back to Portkey. Very useful if you’re a system admin, and you’re looking to track the costs/user on a centralized instance of LibreChat.
MindsDB
Connect your databases, vector stores, and apps to 250+ LLMs with enterprise-grade monitoring and reliability built-in.
OpenWebUI
Portkey is the only plugin you’ll need for your model management, cost tracking, observability, metadata logging, and more for your Open WebUI instance.
ToolJet
Add AI-powered capabilities such as chat completions and automations into your ToolJet apps easily
Resources
2-Min Guides
- Guide to Prompt Caching link
- Building Prod-Ready Apps with Vercel link
- OpenAI Swarm Cheat Sheet link
Longer Guides
- How to Build Multi-Agent AI Systems with OpenAI Swarm & Secure Them Using Portkey link
- We modified Anthropic’s RAG Cookbook to have Observability and unified API link
- End-to-End Guide for Using Vercel with Portkey link
- What is Automated Prompt Engineering? link
- OpenAI’s Prompt Caching: A Deep Dive link
- The Complete Guide to Prompt Engineering link
- Multi-Agent AI Systems: OpenAI Swarm link
- The Developer’s Guide to Opentelemetry link
More awesome content here.
Fixes
- Enhanced streaming transformer for Perplexity
- Fixed response transformation for Ollama
- ⭐️ Added missing logprob mapping for Azure OpenAI (Community Contribution by Avishkar
- Token counting is now fixed for Vertex embeddings (we now count tokens instead of characters)
- Added default models for Gemini, Together AI, Fireworks AI.
- Fireworks:
accounts/fireworks/models/llama-v3p1-405b-instruct
- Together AI:
meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo
- Gemini:
gemini-1.5-pro
- Fireworks:
- Support for cross-region model IDs for Bedrock including cost calculations - https://github.com/Portkey-AI/gateway/pull/641 - Pricing for Bedrock cross-region model IDs- https://github.com/Portkey-AI/Winky/pull/278/files
- Fixed response transform for Ollama
anthropic-beta
andanthropic-version
headers are now supported- Fix for sending media files for Vertex AI & Gemini
- Support for additional headers from providers - https://github.com/Portkey-AI/portkey-python-sdk/pull/222/files
- API key is optional when using self-hosted Gateway with the Portkey SDK
Events
- TED: We co-sponsored the TED AI Hackathon!
- LLMs in Prod Dinner: Singapore - We are organising a closed-door dinner with some of the leading tech executives in Singapore, on the preset of OpenAI Dev Day. Link to Register
News
- OpenAI users were hitting usage limits earlier this month. Though, if you were on Portkey, you should have been fine.
Found a bug or have a feature request? Open an issue on our GitHub repository.
Was this page helpful?