Category "apache-kafka"

Make kafka consumer group inactive

My biggest question is that do we have a concept in kafka called as consumer group to be inactive/active. I am not sure how to achieve this in my local kafka

Spark streamming take long time read from kafka

I build a cluster use CDH5.14.2, includes 5 nodes, each node has 130G momery and 40 cpu cores. I builded the spark streamming application to read from multiple

Kafka Producer Thread, huge amound of threads even when no message is send

i currently profiled my kafka producer spring boot application and found many "kafka-producer-network-thread"s running (47 in total). Which would never stop run

Apache Kafka - DNS resolution failed for url in bootstrap.servers (Failed to construct kafka producer)

I have a problem to construct kafka producer with hostname. Here's what I did. 1) I set the host.name, port config to server.properties listeners=PLAINTEXT://

Where is the error log file destination for Zookeeper distributed with Kafka?

I'm a bit confused on how to configure where the error log files are written for Zookeeper when running ZK and Kafka on the same node. I understand it uses log

Error sending fetch request (sessionId=1175648978, epoch=189) to node 53: org.apache.kafka.common.errors.DisconnectException

We have a topic with 100 partitions and the load is millions of records per hour. We ran into the problem whenever we deploy a new version of stream-processor

Synchronising transactions between database and Kafka producer

We have a micro-services architecture, with Kafka used as the communication mechanism between the services. Some of the services have their own databases. Say

Unable to connect to Kafka using p12 file in .net core application

I have a .net core application that needs to connect to Kafka using SSL protocol. I have a p12 file which I need to use for authentication with the broker but I

Message loss/missing from same topic, read with different consumer group in Kafka

I have been encountering a weird issue with Kafka and Confluent Sink Connector which I am using in my setup. I have a system where in I have two kafka connect s

Locust Python: Kafka Consumer in a separate thread

I'm using Locust for load testing. I want to register a Kafka consumer in separate thread to measure the time of message processing. Here is what I got now: def

What should I use instead deprecated FlinkKafkaConsumer? Scala Flink

I try to get data from Kafka to Flink, I use FlinkKafkaConsumer but Intellij shows me that it is depricated and also ssh console in Google Cloud shows me this e

Is there a way to manually store Kafka offset so a consumer never misses messages?

Using PHP Laravel Framework to consume kafka messages with the help of the mateusjunges/laravel-kafka laravel package. Is it possible to save the offset by cons

Writing JUnit tests for Kafka Consumer

I have a kafka consumer which is subscribing on a topic. Implementation is working fine. But when trying to implement unit tests for that, there's a problem bec

Not able to create kafka topics with desired partitions using Kafka Admin Client API

I'm using Kafka Admin client API's to create the topic. The topic is getting created, however the topic is getting created with 1 partition by default. The API

Kafka Headless Service Issue

Headless service does not bind to new IPs. A pod restart is required. When both zookeepers and brokers restarted at same time brokers are not able to connect wi

"The $changeStream stage is only supported on replica sets" error while using mongodb-source-connect

I get an error when running kafka-mongodb-source-connect I was trying to run connect-standalone with connect-avro-standalone.properties and MongoSourceConnector

Spring Kafka : Record listener vs Batch listener

With spring-kafka, there is two types of Kafka listeners. Record Listeners : @KafkaListener(groupId = "group1", topics = {"my.topic"}) public void listenSingl

NestJS: How to retry when Kafka consumer fails?

There might be a situation where even though the message is received from Kafka, but due to some reason (Database is down, webhook is offline or ...) still the

Allowed packages in custom header of Kafka-Message

In spring-kafka, how do I add classes from a package to be trusted as a custom header field? The message is being sent like this: @Autowired private KafkaTemp

Exponential backoff with message order guarantee using spring-kafka

I'm trying to implement a Spring Boot-based Kafka consumer that has some very strong message delivery guarentees, even in a case of an error. messages from a pa