IN PERSON: Apache Kafka® x Apache Iceberg™ Meetup


Details
Join us for an Apache Kafka® Meetup on Monday, April 14th from 6:00pm hosted by Elaia!
Elaia is a full stack tech and deep tech investor. We partner with ambitious entrepreneurs from inception to leadership, helping them navigate the future and the unknown. For over twenty years, we have combined deep scientific and technological expertise with decades of operational experience to back those building tomorrow. From our offices in Paris, Barcelona and Tel Aviv, we have been active partners with over 100 startups including Criteo, Mirakl, Shift Technology, Aqemia and Alice & Bob.
📍Venue:
Elaia
21 Rue d'Uzès, 75002 Paris, France
🗓 Agenda:
- 6:00pm: Doors Open/Welcome, Drinks
- 6:15pm - 7:00pm: Roman Kolesnev, Principal Software Engineer, Streambased
- 7:00pm - 7:45pm: Viktor Gamov, Principal Developer Advocate, Confluent
- 7:45pm - 8:30pm: Food, Additional Q&A, Networking
💡 Speaker One:
Roman Kolesnev, Principal Software Engineer, Streambased
Talk:
Melting Icebergs: Enabling Analytical Access to Kafka Data through Iceberg Projections
Abstract:
An organisation's data has traditionally been split between the operational estate, for daily business operations, and the analytical estate for after-the-fact analysis and reporting. The journey from one side to the other is today a long and torturous one. But does it have to be?
In the modern data stack Apache Kafka is your defacto standard operational platform and Apache Iceberg has emerged as the champion of table formats to power analytical applications. Can we leverage the best of Iceberg and Kafka to create a powerful solution greater than the sum of its parts?
Yes you can and we did!
This isn't a typical story of connectors, ELT, and separate data stores. We've developed an advanced projection of Kafka data in an Iceberg-compatible format, allowing direct access from warehouses and analytical tools.
In this talk, we'll cover:
* How we presented Kafka data for Iceberg processors without moving or transforming data upfront—no hidden ETL!
* Integrating Kafka's ecosystem into Iceberg, leveraging Schema Registry, consumer groups, and more.
* Meeting Iceberg's performance and cost reduction expectations while sourcing data directly from Kafka.
Expect a technical deep dive into the protocols, formats, and services we used, all while staying true to our core principles:
* Kafka as the single source of truth—no separate stores.
* Analytical processors shouldn't need Kafka-specific adjustments.
* Operational performance must remain uncompromised.
* Kafka's mature ecosystem features, like ACLs and quotas, should be reused, not reinvented.
Join us for a thrilling account of the highs and lows of merging two data giants and stay tuned for the surprise twist at the end!
Bio:
Roman is a Principal Software Engineer at Streambased. His experience includes building business critical event streaming applications and distributed systems in the financial and technology sectors.
💡 Speaker Two:
Viktor Gamov, Principal Developer Advocate, Confluent
Talk:
One Does Not Simply Query a Stream
Abstract:
Streaming data with Apache Kafka® has become the backbone of modern day applications. While streams are ideal for continuous data flow, they lack built-in querying capability. Unlike databases with indexed lookups, Kafka's append-only logs are designed for high throughput processing, not for on-demand querying. This necessitates teams to build additional infrastructure to enable query capabilities for streaming data. Traditional methods replicate this data into external stores such as relational databases like PostgreSQL for operational workloads and object storage like S3 with Flink, Spark, or Trino for analytical use cases. While useful sometimes, these methods deepen the divide between operational and analytical estates, creating silos, complex ETL pipelines, and issues with schema mismatches, freshness, and failures.
In this session, we’ll explore and see live demos of some solutions to unify the operational and analytical estates, eliminating data silos. We’ll start with stream processing using Kafka Streams, Apache Flink®, and SQL implementations, then cover integration of relational databases with real-time analytics databases such as Apache Pinot® and ClickHouse. Finally, we’ll dive into modern approaches like Apache Iceberg® with Tableflow, which simplifies data preparation by seamlessly representing Kafka topics and associated schemas as Iceberg or Delta tables in a few clicks. While there's no single right answer to this problem, as responsible system builders, we must understand our options and trade-offs to build robust architectures.
Bio:
Viktor Gamov is a Principal Developer Advocate at Confluent, founded by the original creators of Apache Kafka®. . With a rich background in implementing and advocating for distributed systems and cloud-native architectures, Viktor excels in open-source technologies. He is passionate about assisting architects, developers, and operators in crafting systems that are not only low in latency and scalable but also highly available.
As a Java Champion and an esteemed speaker, Viktor is known for his insightful presentations at top industry events like JavaOne, Devoxx, Kafka Summit, and QCon. His expertise spans distributed systems, real-time data streaming, JVM, and DevOps.
Viktor has co-authored "Enterprise Web Development" from O'Reilly and "Apache Kafka® in Action" from Manning.
Follow Viktor on X - @gamussa to stay updated with Viktor's latest thoughts on technology, his gym and food adventures, and insights into open-source and developer advocacy.
***
DISCLAIMER
We cannot cater to those under the age of 18.
If you would like to speak at / host a future meetup, please reach out to community@confluent.io

Sponsors
IN PERSON: Apache Kafka® x Apache Iceberg™ Meetup