In this article we will learn how to run Apache Kafka as Docker image using Docker Compose, which allows to manage multiple Container Images. Firstly, we will show how to run Kafka with Zookeeper and Docker. Then, we will show how to run Apache Kafka KRaft with Docker.
This tutorial will teach you how to run Apache Kafka in KRaft mode which does not require a Zookeeper server to manage the cluster capablities. We will learn at first how does the Kafka KRaft mode work and then we will learn how to create a cluster without using Zookeeper.
This article discusses how you can actively monitor Apache Kafka using JMX and the JmxTool that is available in your installation. At the end of this tutorial, you will be able to collect key metrics to troubleshoot and monitor your Kafka cluster.
Apache Kafka and Zookeeper are important components of the Apache Kafka ecosystem that play a crucial role in the processing and management of large-scale distributed data.
To monitor and manage these processes effectively, it is important to know their Process ID (PID), which is a unique identifier assigned to each running process on a system. In this article, we will discuss how to find the PID of Apache Kafka and Zookeeper processes.
Clustering Apache ZooKeeper is a way to set up multiple ZooKeeper servers to work together as a single logical unit. Clustering is useful when you need to scale your system or provide high availability. Here is a tutorial on how to set up a ZooKeeper cluster.
Apache ZooKeeper is a distributed, open-source coordination service for distributed systems. It helps to manage large distributed systems by providing a centralized service that allows nodes to communicate and exchange information.
In this tutorial, we will learn how to use ZooKeeper to coordinate the actions of distributed nodes.
In this tutorial we will learn how to get started with Apache Kafka. We will create a simple set up with a Single Broker Kafka Cluster and produce/consume messages with it.
In this tutorial we will learn how to start a Single Node with Mutiple Broker Kafka Cluster and produce/consume messages with it.
In this tutorial we will learn how to set up a Maven project to run a Kafka Java Consumer and Producer.
this is the second tutorial about creating a Java Producer an Consumer with Apache Kafka. In the first tutorial we have learnt how to set up a Maven project to run a Kafka Java Consumer and Producer (Kafka Tutorial: Creating a Java Producer and Consumer ) Now we will code a more advanced use case, when custom Java types are used in messaging.