You can use the kafka-topics.sh –describe option to know which is the leader for a Kafka topic and the broker instances acting as replicas for the topic, along with the number of partitions of a Kafka Topic that has been created with.
Welcome to Apache Kafka Cheatsheet! This is the most complete Apache Kafka cheatsheet. Enjoy it!
Every message your producers send to a Kafka partition has an offset—a sequential index number that identifies each message. To keep track of which messages have already been processed, your consumer needs to commit the offsets of the messages that were processed.
Kafka topics can be consumed using a single-threaded Consumer or using a multi-threaded Consumer. In this tutorial we will learn the differences between them.
In this tutorial we will learn how to get started quickly with Apache Kafka on OpenShift. The same concepts apply if you want to run Apache Kafka on any Kubernates compliant environment
In order to kickstart Apache Kafka on OpenShift, we will install the Strimzi Operator. The project Strimzi (https://strimzi.io/) simplifies the process of running Apache Kafka in a Kubernetes cluster in various deployment configurations.
Let’s create a new OpenShift project named ‘kafka-demo’
In this tutorial we will learn how access an Apache Kafka cluster running on OpenShift / Kubernetes using the Kafka Bridge.
The Apache Kafka Bridge includes a RESTful interface that let HTTP-based clients to interact with a Kafka cluster. This can simplify accessing the Kafka cluster which might be running on a Cloud Environment such as Kubernates or OpenShift.
In order to complete this tutorial, install the Strimzi Operator, the Kafka cluster and a Topic as discussed in this tutorial: Apache Kafka on OpenShift quickstart guide
In this tutorial we will learn how to connect to a Kafka cluster from a Spring Boot REST Controller. As a proof of concept, we will set up a basic Web application which produces and consumes messages that will be streamed to Kafka.
This is our first tutorial about Apache Kafka . We will start learning what Apache Kafka and why we need in our Enterprise Integration layer.