Use CCLOUD50 to get an additional $50 of free Confluent Cloud- (details)

Streaming Audio Podcast

Listen to Tim Berglund, Gwen Shapira, and guests unpack a variety of topics surrounding Apache Kafka®, Confluent, real-time data streaming, and the cloud.

All Topics
  • All Topics
  • Apache Kafka
  • Architecture
  • Ask Confluent
  • Big Ideas
  • Clients
  • Cloud
  • Confluent Platform
  • Distributed Systems Engineering
  • Event Streaming
  • Integration
  • Kafka
  • Kafka 101
  • Kafka Connect
  • Kafka Streams
  • ksqlDB
  • Kubernetes
  • Microservices
  • Schema Registry
  • Spring
  • Stream Processing
  • Streaming Audio Special
  • Use Cases
Subscribe
Episode 36June 12, 2019 | 46 min

Why Kafka Connect? ft. Robin Moffatt

Tim and Robin cover the motivating factors for Kafka Connect, why people end up reinventing the wheel when they're not aware of it and Kafka Connect's capabilities. They also talk about the importance of schemas in Apache Kafka® pipelines and programs, and how the Confluent Schema Registry can help.

Episode 35June 3, 2019 | 41 min

Schema Registry Made Simple by Confluent Cloud ft. Magesh Nandakumar

Tim Berglund and Magesh Nandakumar discuss why schemas matter for building systems on Apache Kafka®, and how Confluent Schema Registry helps with the problem. They talk about how Schema Registry works, how you can collaborate around schema change through `avsc` files, and what it means for this to be available in Confluent Cloud today.

Episode 34May 29, 2019 | 45 min

Why is Stream Processing Hard? ft. Michael Drogalis

Tim Berglund and Michael Drogalis talk about all things stream processing: why it’s complex, how it has evolved, and what’s on the horizon to make it simpler.

Episode 33May 20, 2019 | 42 min

Testing Kafka Streams Applications with Viktor Gamov

Tim Berglund is joined by Viktor Gamov (Developer Advocate, Confluent) to discuss various approaches to testing Kafka Streams applications.

Episode 32May 16, 2019 | 50 min

Chris Riccomini on the History of Apache Kafka and Stream Processing

Chris Riccomini tells us how Apache Kafka® and the stream processing framework Samza came about, and also what he’s doing these days at WePay—building systems that use Kafka as a primary datastore.

Page 20
  • Page 1
  • Page 2
  • Page 3
  • Page 4
  • Page 5
  • Page 6
  • Page 7
  • Page 8
  • Page 9
  • Page 10
  • Page 11
  • Page 12
  • Page 13
  • Page 14
  • Page 15
  • Page 16
  • Page 17
  • Page 18
  • Page 19
  • Page 20
  • Page 21
  • Page 22
  • Page 23
  • Page 24
  • Page 25
  • Page 26

Meet your hosts

Tim Berglund

Tim Berglund is a teacher, author, and technology leader with Confluent, where he serves as the senior director of developer advocacy. He can frequently be found at speaking at conferences in the U.S. and all over the world. Tim is the co-presenter of various O'Reilly training videos on topics ranging from Git to distributed systems, and he is the author of "Gradle Beyond the Basics." He lives in Littleton, CO, U.S., with the wife of his youth.

Gwen Shapira

Gwen Shapira is an engineering leader at Confluent. She has over 15 years of experience working with code and customers to build scalable data architectures, integrating relational and big data technologies. Gwen is the author of "Kafka: The Definitive Guide" and "Hadoop Application Architectures." Gwen is a frequent presenter at industry conferences, a PMC member on the Apache Kafka project, and a committer on Apache Sqoop™. When Gwen isn't building data pipelines or thinking up new features, you can find her pedaling on her bike exploring the roads and trails of California, and beyond.

Got questions?

If there's something you want to know about Apache Kafka, Confluent or event streaming, please send us an email with your question and we'll hope to answer it on the next episode of Ask Confluent.

Email Us

Never miss an episode!

Be the first to get updates and new content

We will only share developer content and updates, including notifications when new content is added. We will never send you sales emails. 🙂 By subscribing, you understand we will process your personal information in accordance with our Privacy Statement.