What is Portkey?

Teams use Portkey to improve the cost, performance, and accuracy of their Gen AI apps.

It takes <2 mins to integrate and with that, it already starts monitoring all of your LLM requests and also makes your app resilient, secure, performant, and more accurate at the same time.

Here's a product walkthrough (3 mins):

Integrate in 3 Lines of Code

# pip install portkey-ai

from openai import OpenAI
from portkey_ai import PORTKEY_GATEWAY_URL, createHeaders

client = OpenAI(
    base_url=PORTKEY_GATEWAY_URL,
    default_headers=createHeaders(
        provider="openai", 
        api_key="PORTKEY_API_KEY"
    )
)

chat_complete = client.chat.completions.create(
    model="gpt-4",
    messages=[{"role": "user", "content": "Say this is a test"}],
)

print(chat_complete.choices[0].message.content)

🚀 Quickstart

Setting up Portkey takes less than 2 mintues.

🤝 Integrations

Find the best integration for you with 0+ models across LLM providers and multiple frameworks.

✨ Product Features

Jump to the product section to learn more about the Portkey modules and the use cases they solve.

📔 API Reference

Head to the API reference and code samples for all Portkey functionality available through REST APIs and SDKs

Languages Supported

AI Providers Supported

Portkey is multimodal by default - along with chat and text models, we also support audio, vision, and image generation models.

AI ProviderStatus
fully supportedpublic
fully supportedpublic
fully supportedpublic
fully supportedpublic
fully supportedpublic
fully supportedpublic
fully supportedpublic
fully supportedpublic
fully supportedpublic
fully supportedpublic
fully supportedpublic
fully supportedpublic

AI21

fully supportedpublic
fully supportedpublic
fully supportedpublic

AzureML

partially supported
fully supportedpublic
fully supportedpublic
fully supportedpublic

View all the supported integration guides.

Frameworks Supported

FrameworkStatus
nativepythontypescript
nativepythontypescript
nativepython
nativetypescript

Last updated