I'm trying to seek the offset to some specific offset. It seems I need to use seekToTimestamp and there is no seekToOffset or something. I've research a lot but
auto.register.schemas=false doesn't work as I expect. If I read the documentation it's suppose to counter the producer to regsiter new schemas. https://docs.c
I am trying to access a bitnami/kafka cluster deployed using helm. I need to be able to send messages to kafka from my local machine and then have my pods proc
I use spring-cloud-stream-binder-kafka to connect to the kafka broker I updated spring-cloud-dependencies version to 2021.0.1 I went from this [INFO] +- org.spr
How can I integrate Kafka producer with spark stateful streaming which uses checkpoint along with StreamingContext.getOrCreate. I read this post: How to write s
I'm implementing a Django web service, which is about to have different platform apps, Reactjs for computers, a swift app for ios, and Kotlin for android device
Currently, I am having an issue where one of the consumer functions throws an error which makes Kafka retry the records again and again. @Bean public Consumer&l
Im new to pyflink. Im tryig to write a python program to read data from kafka topic and prints data to stdout. I followed the link Flink Python Datastream API K
We have a need to push the kafka topic JSON records to postGresSql database. the JSON are compliant to https://json-schema.org/draft-07/json-schema-release-note
Let's say we are using Kafka with manual commits. We are processing the incoming message. But if there is a failure in processing for any reason, we want to rer
I'm trying to send bytes with a JSON object using a Java POJO and an Avro schema, with schema registry. So my question is, how should my code send "bytes" type
I am using AWS MSK for our production workload and we have been noticing some not very clear log messages in cloudwatch. The messages are about the internal com
I have used KafkaJS library and configured an consumer code to fetch messages from a topic of one of my streaming API. I have created a consumer.js file and I'm
I am writing sample applications in .Net core to interact with Kafka. I have downloaded Kafka and Zookeeper official docker images to my machine. I am using Con
Thos code gives the error "no module named Kafka" import os from datetime import datetime import time import threading import json from kafka import KafkaProduc
I am trying to consume messages from Kafka and publish them to Google PubSub. We have 4 Kafka concurrent consumer threads and I injected the Google pubsub clien
I am trying to consume messages from Kafka and publish them to Google PubSub. We have 4 Kafka concurrent consumer threads and I injected the Google pubsub clien
I need to create an application which has to bridge multiple Queue solutions like Rabbitmq, Kafka, Azure eventhub. I am thinking to use Spring cloud Stream so t
I use istio-ingress gateway and virtualservice to expose different microservices. So far all of them have been http services, so it was straight-forward to foll
I am using version 2.13. Kafka is able to produce messages but not able to consume messages after investigation I found that consumer groups are not getting cre