Portkey Blog Portkey Blog
  • Home
  • Production Guides
  • New Releases
  • Talks
  • Upcoming Events
  • Paper Summaries
  • Portkey Docs
  • Join Community
Sign in Subscribe

instruction tuning

LLMLingua: Compressing Prompts for Accelerated Inference of Large Language Models - Summary

LLMLingua: Compressing Prompts for Accelerated Inference of Large Language Models - Summary

This paper presents a method for compressing prompts in large language models (LLMs) to accelerate model inference and reduce cost. The method involves a budget controller, a token-level iterative compression algorithm, and an instruction tuning based method for distribution alignment. Experimental
The Quill Oct 14, 2023

HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in HuggingFace - Summary

The paper proposes HuggingGPT, a system that uses large language models (LLMs) like ChatGPT to connect various AI models in machine learning communities like HuggingFace to solve complicated AI tasks. The system leverages the strong language capability of ChatGPT and abundant AI models in HuggingFa
The Quill Apr 20, 2023

Subscribe to Portkey Blog

  • Portkey Blog
  • Portkey Website
Portkey Blog © 2025. Powered by Ghost