http://deprecated-apache-flink-user-mailing-list-archive.369.s1.nabble.com/Starting-a-job-that-does-not-use-checkpointing-from-a-savepoint-is-broken-tp17874.html
I ran into a rather annoying issue today while upgrading a flink jobs from flink 1.3.2 to 1.4.0
This particular job does not use checkpointing not state.
First created a savepoint, upgraded the cluster, then restarted the job from the savepoint.
This all went well until later a few hours later one of our kafka nodes dies.This triggered an exception in the job which was subsequently restarted.
Needless to say that creating a savepoint for a job without state or checkpoints does not make that much sense. But I would not expect a restart from a savepoint to completely break a job in the case of failure.
I created a repository that reproduces the scenario I encountered
Am I misunderstanding anything or should i file a bug for this ?