Logo video2dn
  • Сохранить видео с ютуба
  • Категории
    • Музыка
    • Кино и Анимация
    • Автомобили
    • Животные
    • Спорт
    • Путешествия
    • Игры
    • Люди и Блоги
    • Юмор
    • Развлечения
    • Новости и Политика
    • Howto и Стиль
    • Diy своими руками
    • Образование
    • Наука и Технологии
    • Некоммерческие Организации
  • О сайте

Скачать или смотреть PagedAttention: Revolutionizing LLM Inference with Efficient Memory Management - DevConf.CZ 2025

  • DevConf
  • 2025-06-26
  • 105
PagedAttention: Revolutionizing LLM Inference with Efficient Memory Management - DevConf.CZ 2025
  • ok logo

Скачать PagedAttention: Revolutionizing LLM Inference with Efficient Memory Management - DevConf.CZ 2025 бесплатно в качестве 4к (2к / 1080p)

У нас вы можете скачать бесплатно PagedAttention: Revolutionizing LLM Inference with Efficient Memory Management - DevConf.CZ 2025 или посмотреть видео с ютуба в максимальном доступном качестве.

Для скачивания выберите вариант из формы ниже:

  • Информация по загрузке:

Cкачать музыку PagedAttention: Revolutionizing LLM Inference with Efficient Memory Management - DevConf.CZ 2025 бесплатно в формате MP3:

Если иконки загрузки не отобразились, ПОЖАЛУЙСТА, НАЖМИТЕ ЗДЕСЬ или обновите страницу
Если у вас возникли трудности с загрузкой, пожалуйста, свяжитесь с нами по контактам, указанным в нижней части страницы.
Спасибо за использование сервиса video2dn.com

Описание к видео PagedAttention: Revolutionizing LLM Inference with Efficient Memory Management - DevConf.CZ 2025

Speaker(s): Rahul Belokar, Sagar Jalindar Aivale

Large language models (LLMs) are pushing the boundaries of artificial intelligence, but their deployment is often hampered by memory bottlenecks arising from the ever-growing size of key-value (KV) caches. Traditional LLM serving systems struggle with inefficient memory utilization and limited scalability. Inspired by the concept of virtual memory paging in operating systems, PagedAttention offers a transformative solution. This novel technique partitions the KV cache into smaller, non-contiguous blocks, enabling dynamic allocation, efficient retrieval, and flexible reuse of memory. By decoupling the physical layout of the cache from the logical structure, PagedAttention minimizes memory fragmentation and overhead.

This approach, integrated within the vLLM framework, an open-source, high-performance LLM serving framework developed at UC Berkeley, yields significant performance gains.Designed to address memory bottlenecks in traditional LLM serving methods, vLLM leverages PagedAttention for efficient KV cache management, optimizing batch processing and eliminating redundant computations. As a result, PagedAttention achieves up to 30× higher throughput compared to traditional LLM serving methods like Hugging Face Transformers, Orca, and NVIDIA’s FasterTransformer. It also reduces KV cache waste to approximately 4%, ensuring near-optimal memory usage and enabling larger batch processing by minimizing memory overhead.

Furthermore, vLLM seamlessly supports advanced sampling techniques, including beam search, without compromising latency. While challenges such as the overhead of managing lookup tables and the potential for increased latency in certain scenarios exist, ongoing research is addressing these limitations. For example, optimized data structures and prefetching strategies can mitigate lookup overhead. Despite these challenges, PagedAttention represents a major advancement in LLM inference, unlocking the potential for scalable and efficient deployment, even on resource-constrained hardware. This breakthrough paves the way for wider adoption of LLMs and empowers researchers to explore even larger and more complex models.
---

Full schedule, including slides and other resources:
https://pretalx.devconf.info/devconf-...

Комментарии

Информация по комментариям в разработке

Похожие видео

  • О нас
  • Контакты
  • Отказ от ответственности - Disclaimer
  • Условия использования сайта - TOS
  • Политика конфиденциальности

video2dn Copyright © 2023 - 2025

Контакты для правообладателей [email protected]