March 11, 2019 | Episode 23

Containerized Apache Kafka On Kubernetes with Viktor Gamov

  • Notes

Kubernetes provides all the building blocks needed to run stateful workloads, but creating a truly enterprise-grade Apache Kafka® platform that can be used in production is not always intuitive. In this episode, Tim Berglund and Viktor Gamov address some of the challenges and pitfalls of managing Kafka on Kubernetes at scale. They also share lessons learned from the development of the Confluent Operator for Kubernetes, and answer questions like:
-What is Kubernetes?
-What are stateful workloads?
-Why are they hard?
-Will Confluent Operator make it easier?

EPISODE LINKS

Continue Listening

Episode 24March 18, 2019 | 36 min

It’s Time for Streaming to Have a Maturity Model ft. Nick Dearden

Nick Dearden explains the five stages of streaming maturity, from the first streaming project you ever build all the way to a state where an entire organization is transformed to think in terms of real-time, event-driven systems.

Episode 25March 26, 2019 | 14 min

Ask Confluent #11: More Services, More Metrics, More Fun

Do metrics for detecting clients from old versions actually exist? Or is Gwen making features up? This and more useful advice is coming up on today's episode of Ask Confluent.

Episode 26April 3, 2019 | 46 min

Magnus Edenhill on librdkafka 1.0

librdkafka has finally reached 1.0! Several important new features include the idempotent producer, sparse broker connections, support for the vaunted KIP-62 and a complete makeover for the C#/.NET client.

Got questions?

If there's something you want to know about Apache Kafka, Confluent or event streaming, please send us an email with your question and we'll hope to answer it on the next episode of Ask Confluent.

Email Us

Never miss an episode!

Be the first to get updates and new content

We will only share developer content and updates, including notifications when new content is added. We will never send you sales emails. 🙂 By subscribing, you understand we will process your personal information in accordance with our Privacy Statement.