Use Cases
Use Infinispan in real-world applications
Boost Application Performance
Sub-millisecond reads and writes with embedded or client-server mode. Side-caching, near-caching, and session externalization.
High Availability & Elasticity
Distribute data across clusters with automatic rebalancing. No single point of failure, elastic scaling, and session externalization.
Backup Across Data Centers
Cross-site replication with active-active or active-passive topologies. Automatic conflict resolution and disaster recovery.
Caching
Side-caching boosts speed, reduces backend load, cuts costs, improves reliability during spikes or outages, and keeps data fresh with flexible expiration.
Semantic Caching for LLMs
Cache LLM responses by semantic similarity instead of exact key match. Save tokens, reduce latency, and cut API costs.
Vector Store for RAG
Store document embeddings, retrieve relevant chunks via kNN search, and feed them to LLMs as grounding context.
Redis drop-in replacement
Infinispan understands the RESP3 protocol and supports a large number of Redis commands. Connect any compatible Redis client without changing any configuration.
Memcached drop-in replacement
Infinispan understands both the text and binary protocols. Connect any compatible Memcached client without changing any configuration.
Agentic AI Caching
Cache tool results, API responses, and intermediate computations for AI agents. Reduce redundant calls and speed up agent workflows.
Conversation Memory Store
Persist multi-turn LLM conversation history across restarts, scale horizontally, and replicate globally with cross-site replication.
MCP Server for AI Agents
Built-in MCP server lets AI assistants and agents query, cache, and manage Infinispan data through the Model Context Protocol.


