Hi lvan Yang,
#1. If a TaskManager crashed exceptionally and there are some running task on it, it
could not join back gracefully. Whether the full job will be restarted depends on the
failover strategies[1].
#2. Currently, when new TaskManagers join to the Flink cluster, the running Flink
job could not rescale automatically. You need to stop with a savepoint and restart
the job manually. The community is still working on this. And you could find more
information in this ticket[2].
Best,
Yang
Hi,
I have setup Filnk 1.9.1 on Kubernetes on AWS EKS with one job manager pod, 10 task manager pods, one pod per EC2 instance. Job runs fine. After a while, for some reason, one pod (task manager) crashed, then the pod restarted. After that, the job got into a bad state. All the parallelisms are showing different color (orange, purple) on the console. I had to basically stop the entire job. My question is should a task manager restart affect the entire cluster/job? Or should it join back gracefully?
Second question is regarding to auto scaling Flink cluster on kubernetes. If I add more nodes/pods (task manager containers) to the cluster, will a running Flink job redistribute load to the additional resources or I have to stop to a savepoint, and restart the job?
Thanks and regards.
Ivan