Use the latest developments with agents to enhance chatbots.
More and more companies are building their own virtual assistants utilizing agents and Retrieval Augmented Generation (RAG) to enhance responses from Large Language Models (LLMs). This approach allows companies to enhance virtual assistants by grounding answers in fact while minimizing security and data leakage concerns. Many companies are in the exploratory phase, and architects and devs have questions about the best ways to structure virtual assistants and the flow of data. Building these apps for production requires weighing considerations such as performance, quality, flexibility, and cost. With Redis and LlamaIndex, customers can build faster, more accurate chatbots at scale while optimizing cost.
Join this session to learn best practices for:
- Architecting virtual assistant apps
- Accelerating document ingestion while minimizing cost
- Improving responses using AI agents
- Optimizing response time and cost with semantic caching
Event Speakers
Tyler Hutcherson
Senior Applied AI Engineer
Redis
Laurie Voss
VP of Developer Relations
LlamaIndex