Decode, Detect, Diagnose Your LLM: Troubleshooting User Behavior Changes & Performance Drift

Описание к видео Decode, Detect, Diagnose Your LLM: Troubleshooting User Behavior Changes & Performance Drift

Join this hands-on workshop to learn data-centric approaches to diagnose issues in LLM-powered applications using text quality metrics such as sentence structure, vocabulary choice, toxicity, and sentiment.

Large language models (LLMs) rarely provide consistent responses for the same prompts over time. It could be due to changes in your LLM model’s performance, but it can also be a result of changes in user behavior. Text quality metrics, when combined, can help to pinpoint and mitigate these issues without the need for expensive ground truth labeling.

This workshop will cover:

Different types of data drift common to LLM applications
Sentiment, toxicity, and vocabulary metrics useful for text applications
Combining text quality metrics to measure change in user behavior and model performance
Translating changes in text quality into actionable mitigation techniques

What you’ll need:
A free WhyLabs account (https://whylabs.ai/free)
A Google account (for saving a Google Colab)

Комментарии

Информация по комментариям в разработке