Developing and Serving RAG-Based LLM Applications in Production

Описание к видео Developing and Serving RAG-Based LLM Applications in Production

There are a lot of different moving pieces when it comes to developing and serving LLM applications. This talk will provide a comprehensive guide for developing retrieval augmented generation (RAG) based LLM applications — with a focus on scale (embed, index, serve, etc.), evaluation (component-wise and overall) and production workflows. We’ll also explore more advanced topics such as hybrid routing to close the gap between OSS and closed LLMs.

Takeaways:

• Evaluating RAG-based LLM applications are crucial for identifying and productionizing the best configuration.

• Developing your LLM application with scalable workloads involves minimal changes to existing code.

• Mixture of Experts (MoE) routing allows you to close the gap between OSS and closed LLMs.

Find the slide deck here: https://drive.google.com/file/d/1ZnE9...


About Anyscale
---
Anyscale is the AI Application Platform for developing, running, and scaling AI.

https://www.anyscale.com/

If you're interested in a managed Ray service, check out:
https://www.anyscale.com/signup/

About Ray
---
Ray is the most popular open source framework for scaling and productionizing AI workloads. From Generative AI and LLMs to computer vision, Ray powers the world’s most ambitious AI workloads.
https://docs.ray.io/en/latest/


#llm #machinelearning #ray #deeplearning #distributedsystems #python #genai

Комментарии

Информация по комментариям в разработке