Semantic caching for reducing LLM latency and LLM cost

Read more here: External Link