AI Agent Evaluation with RAGAS

Описание к видео AI Agent Evaluation with RAGAS

RAGAS (RAG ASsessment) is an evaluation framework for RAG pipelines. Here, we see how to use RAGAS for evaluating an AI agent built using LangChain and using Anthropic's Claude 3, Cohere's embedding models, and the Pinecone vector database.

📌 Code:
https://github.com/pinecone-io/exampl...

📕 Article:
https://www.pinecone.io/learn/series/...

🌲 Subscribe for Latest Articles and Videos:
https://www.pinecone.io/newsletter-si...

👋🏼 AI Consulting:
https://aurelio.ai

👾 Discord:
  / discord  

Twitter:   / jamescalam  
LinkedIn:   / jamescalam  

00:00 RAG Evaluation
00:39 Overview of LangChain RAG Agent
03:04 RAGAS Code Prerequisites
03:40 Agent Output for RAGAS
05:14 RAGAS Evaluation Format
08:04 RAGAS Metrics
08:56 Understanding RAGAS Metrics
09:16 Retrieval Metrics
11:55 RAGAS Context Recall
14:43 RAGAS Context Precision
15:52 Generation Metrics
16:05 RAGAS Faithfulness
17:16 RAGAS Answer Relevancy
18:40 Metrics Driven Development

#ai #artificialintelligence #nlp #chatbot #langchain

Комментарии

Информация по комментариям в разработке