Logo video2dn
  • Сохранить видео с ютуба
  • Категории
    • Музыка
    • Кино и Анимация
    • Автомобили
    • Животные
    • Спорт
    • Путешествия
    • Игры
    • Люди и Блоги
    • Юмор
    • Развлечения
    • Новости и Политика
    • Howto и Стиль
    • Diy своими руками
    • Образование
    • Наука и Технологии
    • Некоммерческие Организации
  • О сайте

Скачать или смотреть Avro, arrow, protobuf, parquet and why

  • StreamNative
  • 2025-06-05
  • 141
Avro, arrow, protobuf, parquet and why
Apache FlinkApache IcebergApache KafkaApache PulsarData StreamingData Streaming SummitStreaming Lakehouse
  • ok logo

Скачать Avro, arrow, protobuf, parquet and why бесплатно в качестве 4к (2к / 1080p)

У нас вы можете скачать бесплатно Avro, arrow, protobuf, parquet and why или посмотреть видео с ютуба в максимальном доступном качестве.

Для скачивания выберите вариант из формы ниже:

  • Информация по загрузке:

Cкачать музыку Avro, arrow, protobuf, parquet and why бесплатно в формате MP3:

Если иконки загрузки не отобразились, ПОЖАЛУЙСТА, НАЖМИТЕ ЗДЕСЬ или обновите страницу
Если у вас возникли трудности с загрузкой, пожалуйста, свяжитесь с нами по контактам, указанным в нижней части страницы.
Спасибо за использование сервиса video2dn.com

Описание к видео Avro, arrow, protobuf, parquet and why

When it comes to streaming data, there's a lot of data formats out there, and as much as we give JSON a pass because it's easy, it's an overhead we can't often afford, so let's dive in and see which ones matter and why! Come along and we'll run through the how and why of data serialisations, how to manage schemas and why we need rows and columns in streaming, and in data lakes. We'll talk through tools evolution, efficiency and portability. Some horror stories of data conflicts and some of the newer ways to take advantage of these

Speakers:
Ben Gamble (Field CTO, Ververica)

TL;DR
In the session "Avro, Arrow, Protobuf, Parquet and Why," Ben Gamble addresses the complexities of choosing data formats for streaming, emphasizing the shortcomings of JSON due to its overhead. He highlights Avro, Protobuf, and Parquet as key formats, explaining their roles in optimizing data processing efficiency. By selecting appropriate formats, practitioners can enhance performance, manage schema evolution, and improve data portability across systems.

Opening
Imagine managing a SaaS platform for same-day deliveries where a sudden influx of JSON requests overwhelms your system, causing data conflicts and service disruptions. This scenario illustrates the pressing need to optimize data serialization for scalability and efficiency. In data streaming, where speed and accuracy are paramount, understanding the right data format can make the difference between seamless operations and catastrophic failures.

What You'll Learn (Key Takeaways)

*Selecting the Right Data Format* – Discover the benefits of using Avro for its dynamic schema capabilities, Protobuf for strong schema guarantees in RPC, and Parquet for efficient long-term data storage.
*Managing Schema Evolution* – Learn how proper schema management can prevent data conflicts and ensure backward and forward compatibility, although it's generally safer to version topics with schema changes.
*Optimizing Data Processing* – Understand how Apache Arrow's columnar in-memory format can enhance computational efficiency, particularly in analytical streaming systems.
*Real-World Applications* – Explore how these formats apply to large-scale streaming systems, from Flink's role in managing petabytes of data to Parquet's use in data lake architectures for long-term storage.

Q&A Highlights
Q: How do large language models fit into the streaming data world dominated by binary formats?
A: Large language models excel with text and JSON, but struggle with schemas. Message Pack offers a middle ground by providing a binary, JSON-like format that is efficient for wire transmission.

Q: Does Parquet work well for high-frequency logging?
A: Yes, Parquet is effective for high-frequency logging; however, the lack of a variant type can be a limitation, which is why most logging systems do not use Parquet as their backend.

For further insights, connect with Ben Gamble on LinkedIn, where he shares his extensive knowledge on data streaming and more.

Комментарии

Информация по комментариям в разработке

Похожие видео

  • О нас
  • Контакты
  • Отказ от ответственности - Disclaimer
  • Условия использования сайта - TOS
  • Политика конфиденциальности

video2dn Copyright © 2023 - 2025

Контакты для правообладателей [email protected]