🎉 GPTCache officially released the first version.
Introduction
GPTCache is a library for creating semantic cache to store responses from LLM queries.
What's Supported
- LLM Adapter
- Support OpenAI ChatGPT API
- Support langchain
- Embedding
- Disable embedding. This will turn GPTCache into a keyword-matching cache
- Support OpenAI embedding API
- Support ONNX with the GPTCache/paraphrase-albert-onnx model
- Support Hugging Face embedding API
- Support Cohere embedding API
- Support fastText embedding API
- Support SentenceTransformers embedding API
- Cache Storage
- Support SQLite
- Support PostgreSQL
- Support MySQL
- Support MariaDB
- Support SQL Server
- Support Oracle
- Vector Store
- Support Milvus
- Support Zilliz Cloud
- Support FAISS
- Similarity Evaluator
- The distance we obtain from the Vector Store
- A model-based similarity determined using the GPTCache/albert-duplicate-onnx model from ONNX
- Exact matches between the input request and the requests obtained from the Vector Store
- Distance represented by applying linalg.norm from numpy to the embeddings
Full Changelog: https://github.com/zilliztech/GPTCache/commits/0.1.5