77 skills found · Page 1 of 3
zilliztech / GPTCacheSemantic cache for LLMs. Fully integrated with LangChain and llama_index.
olimorris / Onedarkpro.nvim🎨 Atom's iconic One Dark theme. Cacheable, fully customisable, Tree-sitter and LSP semantic token support. Comes with variants
codefuse-ai / ModelCacheA LLM semantic caching system aiming to enhance user experience by reducing response time via cached query-result pairs.
redis-developer / ArXivChatGuruUse ArXiv ChatGuru to talk to research papers. This app uses LangChain, OpenAI, Streamlit, and Redis as a vector database/semantic cache.
edwinkys / OasysdbIn-memory vector store with efficient read and write performance for semantic caching and retrieval system. Redis for Semantic Caching.
thu-nics / C2C[ICLR'26] The official code implementation for "Cache-to-Cache: Direct Semantic Communication Between Large Language Models"
messkan / Prompt CacheCut LLM costs by up to 80% and unlock sub-millisecond responses with intelligent semantic caching.A drop-in, provider-agnostic LLM proxy written in Go with sub-millisecond response
aqstack / Mimirmimir is a drop-in proxy that caches LLM API responses using semantic similarity, reducing costs and latency for repeated or similar queries.
peva3 / SmarterRouterSmarterRouter: An intelligent LLM gateway and VRAM-aware router for Ollama, llama.cpp, and OpenAI. Features semantic caching, model profiling, and automatic failover for local AI labs.
colossus-lab / Openarg BackendAI-powered analysis engine for Argentine government open data. Multi-agent pipeline (LangGraph) with 10 data connectors, NL2SQL, semantic caching, and real-time streaming. Built with FastAPI, PostgreSQL + pgvector, Celery, and Gemini/Claude LLMs.
sensoris / SemcacheSemantic caching layer for your LLM applications. Reuse responses and reduce token usage.
vcache-project / VCacheReliable and Efficient Semantic Prompt Caching with vCache
marmeladema / Clru RsAn LRU cache implementation with constant time operations and weighted semantic.
shivendrasoni / Vector CacheA simple semantic cache implementaion. It caches responses from an LLM based on semantic similarity.
winter2020 / KleespectreKLEESpectre is a symbolic execution engine with speculation semantic and cache modelling
albertobadia / ZoocacheSemantic dependency based cache with high performance and concurrency in mind.
AzozzALFiras / Claude Context OptimizerMCP server that cuts Claude Code token usage by up to 98% — smart file caching, semantic read, log compression, task checkpoints, and context watchdog. Zero native dependencies.
databricks-industry-solutions / Semantic CachingThis project implements a caching system for Databricks, designed to improve response times and reduce cost for frequently asked questions or similar queries.
sjtu-zhao-lab / ClusterKVClusterKV: Manipulating LLM KV Cache in Semantic Space for Recallable Compression (DAC'25)
Canonical-AI-Inc / CanonicalContext-Aware Semantic Cache for Conversational AI