Portkey Blog Portkey Blog
  • Home
  • Production Guides
  • New Releases
  • Talks
  • Upcoming Events
  • Portkey Docs
Sign in Subscribe

Cost reduction

LLM cost attribution: Tracking and optimizing spend for GenAI apps

Learn how to track and optimize LLM costs across teams and use cases. This blog covers challenges, best practices, and how LLMOps platforms like Portkey enable cost attribution at scale.
Drishti Shah 29 Apr 2025
Our AI overlords

⭐ Semantic Cache for Large Language Models

Learn how semantic caching for large language models reduces cost, improves latency, and stabilizes high-volume AI applications by reusing responses based on intent, not just text.
Vrushank Vyas 11 Jul 2023

Subscribe to Portkey Blog

  • Blog Home
  • Portkey Website
Portkey Blog © 2026. Powered by Ghost