Hi, All We encountered some problems during the upgrade from Flink 1.5 to Flink 1.9. Flink's scheduling strategy has changed. Flink 1.9 prefers centralized scheduling, while Flink 1.5 prefers decentralized scheduling. This change has caused resources imbalance and blocked our upgrade plan. We have thousands of jobs that need to be upgraded. For example, There is a job with 10 sources and 100 sinks. Each source need 1 core and each sink need 0.1 core. Try to run this job on Yarn, configure the numberOfTaskSlots is 10, yarn.containers.vcores is 2. When using Flink-1.5: Each TaskManager will run 1 source and 9 sinks, they need 1.9 cores totally. So the job with this configuration works very well. The schedule results is shown in Figure 1. When using Flink-1.9: The 10 sources will be scheduled to one TaskManager and the 100 sinks will scheduled to other 10 TaskManagers. The schedule results is shown in Figure 2. In this scenario, the TaskManager which run sources need 10 cores, other TaskManagers need 1 cores. But TaskManager must be configured the same, So we need 11 TaskManager with 10 cores. This situation waste (10-2)*11 = 88 cores more than Flink 1.5. In addition to the waste of resources, we also encountered other problems caused by centralized scheduling strategy.
In summary, we think a decentralized scheduling strategy is necessary. Figure 1. Flink 1.5 schedule results Figure 2. Flink 1.9 schedule results Best Weihua Hu |
HI HuWeihua, I think your issue should resolve with 1.9.2 and 1.10 (not released but in progress). You can check the related Jira ticket [1]. Best, Andrey On Wed, Jan 15, 2020 at 10:08 AM HuWeihua <[hidden email]> wrote:
|
In reply to this post by HuWeihua
This is a known issue that's will be
fixed in 1.9.2/1.10.0; see
https://issues.apache.org/jira/browse/FLINK-12122 .
On 15/01/2020 10:07, HuWeihua wrote:
|
In reply to this post by Andrey Zagrebin-5
Hi, Andrey Thanks for your response. I have checked this Jira ticket and I think it can work in standalone mode which TaskManager has been started before scheduling tasks. But we are currently running flink on yarn in per-job cluster mode. I noticed that this issue has already been raised. I will keep watching this ticket. Thanks again. Best Weihua Hu
|
Thanks for reporting the issue HuWeihua. Choosing the right scheduling strategy when using Yarn with potentially infinite resources can be quite hard because you don't know over how many TaskExecutors one should distribute the tasks. It becomes easier if one can configure the minimum number of TaskExecutors a cluster should always have. This is currently being discussed and I hope that we can complete this feature for the next release. Cheers, Till On Wed, Jan 15, 2020 at 11:29 AM HuWeihua <[hidden email]> wrote:
|
Free forum by Nabble | Edit this page |