Show HN: GPTCache – Redis for LLMs https://ift.tt/4IiKYd6

Show HN: GPTCache – Redis for LLMs Hey folks, As much as we love GPT-4, it's expensive and can be slow at times. That's why we built GPTCache - a semantic cache for autoregressive LMs - atop the vector database Milvus and SQLite. GPTCache provides several benefits: 1) reduced expenses due to minimizing the number of requests and tokens sent to the LLM service 2) enhanced performance by fetching cached query results directly 3) improved scalability and availability by avoiding rate limits, and 4) a flexible development environment that allows developers to verify their application's features without connecting to LLM APIs or network. Come check it out! https://ift.tt/hrBTep1 https://ift.tt/zkfLmAo April 13, 2023 at 03:14AM

Comments

Popular Posts