Reducing LLM Calls with Vector Search Patterns - Raphael De Lio (Redis)
#1about 3 minutes
The hidden costs of large LLM context windows
Large context windows in models like GPT-5 seem to eliminate the need for RAG, but the high token cost makes this approach expensive and unscalable for every request.
#2about 3 minutes
A brief introduction to vectors and vector search
Text is converted into numerical vector embeddings that capture its semantic meaning, allowing computers to efficiently calculate the similarity between different phrases or documents.
#3about 9 minutes
How to classify text using a vector database
Instead of using a costly LLM for every classification task, you can use a vector database to match new text against pre-embedded reference examples for a specific label.
#4about 5 minutes
Using semantic routing for efficient tool calling
By matching user prompts against pre-defined reference phrases for each tool, you can directly trigger the correct function without an initial, expensive LLM call.
#5about 5 minutes
Reducing latency and cost with semantic caching
Semantic caching stores LLM responses and serves them for new, semantically similar prompts, which avoids re-computation and significantly reduces both cost and latency.
#6about 7 minutes
Strategies for optimizing vector search accuracy
Improve the accuracy of vector search patterns through techniques like self-improvement, a hybrid approach that falls back to an LLM, and chunking complex prompts into smaller clauses.
#7about 3 minutes
Addressing advanced challenges in semantic caching
Mitigate common caching pitfalls, like misinterpreting negative prompts, by using specialized embedding models and combining semantic routing with caching to avoid caching certain types of queries.
Related jobs
Jobs that call for the skills explored in this talk.
Matching moments
34:07 MIN
Reducing latency and cost with semantic caching
WeAreDevelopers LIVE - Vector Similarity Search Patterns for Efficiency and more
05:45 MIN
Solving LLM limitations with RAG and vector databases
Accelerating GenAI Development: Harnessing Astra DB Vector Store and Langflow for LLM-Powered Apps
20:14 MIN
Using semantic classification to categorize text
WeAreDevelopers LIVE - Vector Similarity Search Patterns for Efficiency and more
26:04 MIN
Exploring advanced RAG techniques and other applications
Build RAG from Scratch
12:27 MIN
Practical strategies for reducing token count
Prompt Engineering - an Art, a Science, or your next Job Title?
21:06 MIN
Advanced patterns for building sophisticated AI applications
Java Meets AI: Empowering Spring Developers to Build Intelligent Apps
00:53 MIN
Understanding LLMs, context windows, and RAG
Beyond Prompting: Building Scalable AI with Multi-Agent Systems and MCP
28:58 MIN
Implementing semantic routing for tool calling and guardrails
WeAreDevelopers LIVE - Vector Similarity Search Patterns for Efficiency and more
Featured Partners
Related Videos
WeAreDevelopers LIVE - Vector Similarity Search Patterns for Efficiency and more
Chris Heilmann, Daniel Cranney, Raphael De Lio & Developer Advocate at Redis
How to Avoid LLM Pitfalls - Mete Atamel and Guillaume Laforge
Meta Atamel & Guillaume Laforge
Carl Lapierre - Exploring Advanced Patterns in Retrieval-Augmented Generation
Carl Lapierre
Accelerating GenAI Development: Harnessing Astra DB Vector Store and Langflow for LLM-Powered Apps
Dieter Flick & Michel de Ru
Martin O'Hanlon - Make LLMs make sense with GraphRAG
Martin O'Hanlon
Semantic AI: Why Embeddings Might Matter More Than LLMs
Christian Weyer
Graphs and RAGs Everywhere... But What Are They? - Andreas Kollegger - Neo4j
Three years of putting LLMs into Software - Lessons learned
Simon A.T. Jiménez
Related Articles
View all articles



From learning to earning
Jobs that call for the skills explored in this talk.

AI Systems and MLOps Engineer for Earth Observation
Forschungszentrum Jülich GmbH
Jülich, Germany
Intermediate
Senior
Linux
Docker
AI Frameworks
Machine Learning

AI & Embedded ML Engineer (Real-Time Edge Optimization)
autonomous-teaming
Canton of Toulouse-5, France
Remote
C++
GIT
Linux
Python
+1

AI & Embedded ML Engineer (Real-Time Edge Optimization)
autonomous-teaming
München, Germany
Remote
C++
GIT
Linux
Python
+1

Machine Learning Engineer - Large Language Models (LLM) - Startup
Startup
Charing Cross, United Kingdom
PyTorch
Machine Learning

Manager of Machine Learning (LLM/NLP/Generative AI) - Visas Supported
European Tech Recruit
Municipality of Bilbao, Spain
Junior
GIT
Python
Docker
Computer Vision
Machine Learning
+2

Deep Learning Engineer For Language Technologies (Re3)
Barcelona Supercomputing Center
Barcelona, Spain
Docker
Ansible
Continuous Integration

Student project: Optimizing Open-set Recognition Methods for Reliable Real-world AI Systems
Imec
Azure
Python
PyTorch
TensorFlow
Computer Vision
+1

Machine Learning Engineer (LLM)
European Tech Recruit
Municipality of Madrid, Spain
Intermediate
Python
PyTorch
Computer Vision
Machine Learning

Deep Learning Engineer for Language Technologies (RE2)
Barcelona Supercomputing Center
Barcelona, Spain
Intermediate
Python
PyTorch
Machine Learning