September 4, 2019 | Episode 51

Ask Confluent #15: Attack of the Zombie Controller

  • Notes

Gwen Shapira (Core Kafka Software Engineer, Confluent) sits down to answer the questions you've had about event streaming, Apache Kafka®, Confluent, and everything in between. This includes creating tables in nested JSON topics, how to balance ordering, latency and reliability, building event-based systems, and how to navigate the tricky endOffsets API. She talks about the hardships of fencing Zombie requests, some of the talks given at previous Kafka Summits, and an important question from Ask Confluent #3

Continue Listening

Episode 52September 9, 2019 | 54 min

Connecting to Apache Kafka with Neo4j

Michael Hunger and David Allen discuss Neo4j basics and major features introduced in Neo4j 3.4.15. They'll cover the history of the integration and features in relation to Apache Kafka®, change data capture (CDC), using Neo4j to put graph operations into an event streaming application, and how GraphQL fits in with event streaming and GRANDstack.

Episode 53September 12, 2019 | 48 min

Jay Kreps on the Last 10 Years of Apache Kafka and Event Streaming

Jay Kreps to talk about stream processing, his early coding days at LinkedIn, starting Confluent, the highs, the lows, and everything in between.

Episode 54September 16, 2019 | 29 min

Should You Run Apache Kafka on Kubernetes? ft. Balthazar Rouberol

What are the maturing stages of Kubernetes adoption? How did Datadog experience these stages? Balthazar Rouberol explains what to think about before hopping on Kubernetes hype train.

Got questions?

If there's something you want to know about Apache Kafka, Confluent or event streaming, please send us an email with your question and we'll hope to answer it on the next episode of Ask Confluent.

Email Us

Never miss an episode!

Be the first to get updates and new content

We will only share developer content and updates, including notifications when new content is added. We will never send you sales emails. 🙂 By subscribing, you understand we will process your personal information in accordance with our Privacy Statement.