AI Dev 25 x NYC | Nitin Kanukolanu: Semantic Caching for LLM Applications
Автор: DeepLearningAI
Загружено: 2025-12-05
Просмотров: 161
Описание:
Nitin Kanukolanu, Applied AI Engineer at Redis, focused on semantic caching during AI Dev 25 x NYC.
As LLMs drive the next wave of applications, compute bottlenecks are becoming a critical challenge. Semantic caching has emerged as a practical strategy to cut costs, reduce latency, and improve consistency in agentic systems. This session covered real-world use cases, explained how semantic caching works, highlighted what to measure in production, and shared strategies for boosting both performance and quality at scale.
Take our course with Redis on this topic: https://www.deeplearning.ai/short-cou...
--------------
Join us at AI Dev 26 x San Francisco! Tickets: https://ai-dev.deeplearning.ai/
Повторяем попытку...
Доступные форматы для скачивания:
Скачать видео
-
Информация по загрузке: