Same kafka partition being consumed by multiple task managers.

classic Classic list List threaded Threaded
1 message Options
Reply | Threaded
Open this post in threaded view
|

Same kafka partition being consumed by multiple task managers.

Deshpande, Omkar

Hello,

I am running a streaming Beam app with the Flink runner(java). 

  • Beam 2.19

  • Flink 1.9

Checkpoints and savepoints are configured to go to s3 and HA is enabled using Zookeeper.

I was running the app with 3 task managers. I took a savepoint and started the app with 6 task managers. My input topic has 12 partitions. With 3 pods, the partitions were distributed evenly. After restarting with the increased number of task managers, some partitions are being consumed by 2 task managers.
partition task manager
0 4
1 4
2 1
3 1
4 3
5 3
6 4, 0
7 4, 0
8 1, 2
9 1, 2
10 3, 5
11 3, 5
 Looks like there were 3 task managers [1,3,4] and they correctly distributed partitions between them.
Then 3 new task managers were added [0,2,5] and partitions were not properly re-distributed.

Where could this metadata be coming from?

Omkar