GPU vs CPU: Running Small Language Models with Ollama & C#

Описание к видео GPU vs CPU: Running Small Language Models with Ollama & C#

In this video, we'll explore the performance differences when running Large Language Models (LLMs) in Ollama using both the CPU and GPU. Watch as I demonstrate a live sample in C# using Microsoft.Extensions.AI to run Ollama inside a Docker container. Curious to see how these models perform locally? Let's dive in and compare the results!

Useful links:

.NET Video Analyzer repository: https://aka.ms/netaivideoanalyser
Ollama in Docker: https://ollama.com/blog/ollama-is-now...
.NET & AI Show:    • .NET AI Community Standup - Build a V...  

Комментарии

Информация по комментариям в разработке