Business events occur all the time in the world and Apache Kafka is the leading technology to store and process these events in real time.
In this series of courses, we will guide you through the data in motion universe, starting with what Apache Kafka is, which pieces it comprises, and how to use them effectively to build real-time, event streaming applications.
Why is Flink useful, and how does it work? This course uses Flink SQL to introduce the core concepts of stream processing with Flink and shows how they fit together to solve real problems.
Creating and storing events is useful, but being able to react to those events and process them is what is truly transformational to how businesses operate today.
When you want the full control over the application logic, Kafka Streams is your go-to stream processing framework.
In the event-driven world, schemas are essential as they can enforce the shape of the data and act as contracts between producers and consumers. The Schema Registry can manage those contracts and allow their evolution in a safe way. Let’s see how it works.
In large organizations, applications usually don’t live in a vacuum. You often need to pull data from an extensive range of data stores across the company and also push some data into other systems.
Fortunately, thanks to the strong community behind Kafka, hundreds of connectors are readily available for you to choose from so you don’t have to reinvent the ‘data integration’ wheel.
If you want to get instant insights from your data streams but don’t want to go through the hassle of writing a full-fledged application and worry about the infrastructure, then ksqlDB is for you. Just write some good old SQL code, press the button, and voila, it’s processing your business events in real-time.
How do you move from a simple proof of concept to a bulletproof data streaming system that is ready for production deployment? In this course, you will learn how to avoid pitfalls when scaling your data streaming platform. Additionally, you'll delve deep into the GitOps framework, allowing you to deliver changes swiftly and securely, not only to your platform but also to the streaming applications built on it.
Learn to build Apache Flink Jobs in Java through video lectures and hands-on exercises, including the creation of a set of Flink jobs that interact with Apache Kafka.
How about we get our hands dirty and write some code now?
Get ready to learn how to develop a robust event streaming application using one of the most widely used Java frameworks. Spring Boot meets Kafka Streams, let the magic begin!
Apache Kafka is a great fit for .NET applications due to its powerful real-time data processing features. Learn how to use Kafka to build a streaming application with .NET.
This course aims to get developers started writing simple Python applications that stream events to and from a Kafka cluster.
Build a scalable, streaming data pipeline in under 20 minutes using Kafka and Confluent. Operationalize data in motion using real-time event streams and change data capture.
You learned earlier that events are immutable and are an exact record of what happened in your system. Storing data as events rather than in relational databases working with traditional rows and columns brings many advantages, like recovering your system quickly by replaying events, a free audit log, or the ability to get better business insights. In the next course, you’ll learn about how events can be used as the storage model for your applications, what event sourcing is, how it works, and how it relates to approaches like CQRS and CDC.
Earlier, you learned that events are the very foundation of real-time event-streaming applications. This course gives you design tips to craft top-notch events that are uniquely tailored to your business needs.
Event Modeling is a simple visual method for designing event-driven systems. In this course, learn how to build an event model and implement the modeled system.
Imagine if anyone had access to the key of your house? Well, it’s the same if you don’t secure your IT infrastructure and services. Without proper security measures, you risk going out of business in the event of a data leak or hacking attack.
Kafka takes security seriously and has incorporated robust security features into all components. If you use Confluent Cloud, check out our Confluent Cloud Security course.
How to integrate Confluent Cloud with on-prem, public, and private cloud data streaming applications to meet connectivity, privacy, and security requirements.
From Apache Kafka security, authentication, and RBAC, to cloud data security and monitoring, learn how to use Confluent Cloud's security features to meet all your security and compliance needs.
A hybrid cloud is now an essential element of overall data architecture. This course explores its advantages and prepares learners for disaster recovery, cloud migration, and data sharing.
In this course, we look at some of the specific Stream Governance features in Confluent Cloud. You will have an opportunity to try out many of these features in the hands-on exercises.
Data mesh is a framework for decentralized domain-driven architectures, with data as a product, that is self-service, and with strong governance models. Learn the benefits of data mesh and how it works.
We believe it’s important to know how a piece of technology works under the hood to understand its consistency, performance, security, and scalability characteristics.
This course describes the architecture of Apache Kafka with lectures from Jun Rao, one of its original authors and a Confluent co-founder. Over to you, Jun!
ksqlDB looks like magic when you first use it, so let’s take a tour and discover how it’s been built with ease-of-use in mind but also security, scalability and performance.