Hi,
Flink works very well with Kafka if you wish to stream data. Following is how I am streaming data with Kafka and Flink.
FlinkKafkaConsumer08<Event>
kafkaConsumer =
new FlinkKafkaConsumer08<>(KAFKA_AVRO_TOPIC,
avroSchema,
properties);
DataStream<Event>
messageStream =
env.addSource(kafkaConsumer);
Is there a way to do a micro batch operation on the data coming from Kafka? What I want to do is to
reduce or aggregate the events coming from Kafka. For instance I am getting 40000 events per second from Kafka and what I want is to group 2000 events into one and send it to my microservice for further processing. Can I use the
Flink DataSet API for this or should I go with Spark or some other framework?
Thanks & Regards
Zeeshan Alam