[DISCUSS] decentralized scheduling strategy is needed

classic Classic list List threaded Threaded
5 messages Options
Reply | Threaded
Open this post in threaded view
|

[DISCUSS] decentralized scheduling strategy is needed

胡伟华
Hi, All
We encountered some problems during the upgrade from Flink 1.5 to Flink 1.9. Flink's scheduling strategy has changed. Flink 1.9 prefers centralized scheduling, while Flink 1.5 prefers decentralized scheduling. This change has caused resources imbalance and blocked our upgrade plan. We have thousands of jobs that need to be upgraded.

For example,
There is a job with 10 sources and 100 sinks. Each source need 1 core and each sink need 0.1 core.
Try to run this job on Yarn, configure the numberOfTaskSlots is 10, yarn.containers.vcores is 2.

When using Flink-1.5:
Each TaskManager will run 1 source and 9 sinks, they need 1.9 cores totally. So the job with this configuration works very well. The schedule results is shown in Figure 1.
When using Flink-1.9:
The 10 sources will be scheduled to one TaskManager  and the 100 sinks will scheduled to other 10 TaskManagers.  The schedule results is shown in Figure 2.
In this scenario, the TaskManager which run sources need 10 cores, other TaskManagers need 1 cores. But TaskManager must be configured the same, So we need 11 TaskManager with 10 cores. 
This situation waste (10-2)*11 = 88 cores more than Flink 1.5.

In addition to the waste of resources, we also encountered other problems caused by centralized scheduling strategy.
  1. Network bandwidth. Tasks of the same type are scheduled to the one TaskManager, causing too much network traffic on the machine.
  1. Some jobs need to sink to the local agent. After centralized scheduling, the insufficient processing capacity of the single machine causes a backlog of consumption.

In summary, we think a decentralized scheduling strategy is necessary.


Figure 1. Flink 1.5 schedule results

Figure 2. Flink 1.9 schedule results



Best
Weihua Hu

Reply | Threaded
Open this post in threaded view
|

Re: [DISCUSS] decentralized scheduling strategy is needed

Andrey Zagrebin-4
HI HuWeihua,

I think your issue should resolve with 1.9.2 and 1.10 (not released but in progress).
You can check the related Jira ticket [1].

Best,
Andrey


On Wed, Jan 15, 2020 at 10:08 AM HuWeihua <[hidden email]> wrote:
Hi, All
We encountered some problems during the upgrade from Flink 1.5 to Flink 1.9. Flink's scheduling strategy has changed. Flink 1.9 prefers centralized scheduling, while Flink 1.5 prefers decentralized scheduling. This change has caused resources imbalance and blocked our upgrade plan. We have thousands of jobs that need to be upgraded.

For example,
There is a job with 10 sources and 100 sinks. Each source need 1 core and each sink need 0.1 core.
Try to run this job on Yarn, configure the numberOfTaskSlots is 10, yarn.containers.vcores is 2.

When using Flink-1.5:
Each TaskManager will run 1 source and 9 sinks, they need 1.9 cores totally. So the job with this configuration works very well. The schedule results is shown in Figure 1.
When using Flink-1.9:
The 10 sources will be scheduled to one TaskManager  and the 100 sinks will scheduled to other 10 TaskManagers.  The schedule results is shown in Figure 2.
In this scenario, the TaskManager which run sources need 10 cores, other TaskManagers need 1 cores. But TaskManager must be configured the same, So we need 11 TaskManager with 10 cores. 
This situation waste (10-2)*11 = 88 cores more than Flink 1.5.

In addition to the waste of resources, we also encountered other problems caused by centralized scheduling strategy.
  1. Network bandwidth. Tasks of the same type are scheduled to the one TaskManager, causing too much network traffic on the machine.
  1. Some jobs need to sink to the local agent. After centralized scheduling, the insufficient processing capacity of the single machine causes a backlog of consumption.

In summary, we think a decentralized scheduling strategy is necessary.


Figure 1. Flink 1.5 schedule results

Figure 2. Flink 1.9 schedule results



Best
Weihua Hu

Reply | Threaded
Open this post in threaded view
|

Re: [DISCUSS] decentralized scheduling strategy is needed

Chesnay Schepler-3
In reply to this post by 胡伟华
This is a known issue that's will be fixed in 1.9.2/1.10.0; see
https://issues.apache.org/jira/browse/FLINK-12122 .

On 15/01/2020 10:07, HuWeihua wrote:

> Hi, All
> We encountered some problems during the upgrade from Flink 1.5 to
> Flink 1.9. Flink's scheduling strategy has changed. Flink 1.9 prefers
> centralized scheduling, while Flink 1.5 prefers decentralized
> scheduling. This change has caused resources imbalance and blocked our
> upgrade plan. We have thousands of jobs that need to be upgraded.
>
> For example,
> There is a job with 10 sources and 100 sinks. Each source need 1 core
> and each sink need 0.1 core.
> Try to run this job on Yarn, configure the numberOfTaskSlots is 10,
> yarn.containers.vcores is 2.
>
> When using Flink-1.5:
> Each TaskManager will run 1 source and 9 sinks, they need 1.9 cores
> totally. So the job with this configuration works very well.
> The schedule results is shown in Figure 1.
> When using Flink-1.9:
> The 10 sources will be scheduled to one TaskManager  and the 100 sinks
> will scheduled to other 10 TaskManagers. The schedule results is shown
> in Figure 2.
> In this scenario, the TaskManager which run sources need 10 cores,
> other TaskManagers need 1 cores. But TaskManager must be configured
> the same, So we need 11 TaskManager with 10 cores.
> This situation waste (10-2)*11 = 88cores more than Flink 1.5.
>
> In addition to the waste of resources, we also encountered other
> problems caused by centralized scheduling strategy.
>
>  1. Network bandwidth. Tasks of the same type are scheduled to the one
>     TaskManager, causing too much network traffic on the machine.
>
>  2. Some jobs need to sink to the local agent. After centralized
>     scheduling, the insufficient processing capacity of the single
>     machine causes a backlog of consumption.
>
>
> In summary, we think a decentralized scheduling strategy is necessary.
>
>
> Figure 1. Flink 1.5 schedule results
>
> Figure 2. Flink 1.9 schedule results
>
>
>
> Best
> Weihua Hu
>

Reply | Threaded
Open this post in threaded view
|

Re: [DISCUSS] decentralized scheduling strategy is needed

胡伟华
In reply to this post by Andrey Zagrebin-4
Hi, Andrey

Thanks for your response.

I have checked this Jira ticket and I think it can work in standalone mode which TaskManager has been started before scheduling tasks.
But we are currently running flink on yarn in per-job cluster mode.

I noticed that this issue has already been raised. I will keep watching this ticket.

Thanks again.

Best
Weihua Hu

> 2020年1月15日 17:53,Andrey Zagrebin <[hidden email]> 写道:
>
> HI HuWeihua,
>
> I think your issue should resolve with 1.9.2 and 1.10 (not released but in progress).
> You can check the related Jira ticket [1].
>
> Best,
> Andrey
>
> [1] https://jira.apache.org/jira/browse/FLINK-12122 <https://jira.apache.org/jira/browse/FLINK-12122>
> On Wed, Jan 15, 2020 at 10:08 AM HuWeihua <[hidden email] <mailto:[hidden email]>> wrote:
> Hi, All
> We encountered some problems during the upgrade from Flink 1.5 to Flink 1.9. Flink's scheduling strategy has changed. Flink 1.9 prefers centralized scheduling, while Flink 1.5 prefers decentralized scheduling. This change has caused resources imbalance and blocked our upgrade plan. We have thousands of jobs that need to be upgraded.
>
> For example,
> There is a job with 10 sources and 100 sinks. Each source need 1 core and each sink need 0.1 core.
> Try to run this job on Yarn, configure the numberOfTaskSlots is 10, yarn.containers.vcores is 2.
>
> When using Flink-1.5:
> Each TaskManager will run 1 source and 9 sinks, they need 1.9 cores totally. So the job with this configuration works very well. The schedule results is shown in Figure 1.
> When using Flink-1.9:
> The 10 sources will be scheduled to one TaskManager  and the 100 sinks will scheduled to other 10 TaskManagers.  The schedule results is shown in Figure 2.
> In this scenario, the TaskManager which run sources need 10 cores, other TaskManagers need 1 cores. But TaskManager must be configured the same, So we need 11 TaskManager with 10 cores.
> This situation waste (10-2)*11 = 88 cores more than Flink 1.5.
>
> In addition to the waste of resources, we also encountered other problems caused by centralized scheduling strategy.
> Network bandwidth. Tasks of the same type are scheduled to the one TaskManager, causing too much network traffic on the machine.
> Some jobs need to sink to the local agent. After centralized scheduling, the insufficient processing capacity of the single machine causes a backlog of consumption.
>
> In summary, we think a decentralized scheduling strategy is necessary.
>
>
> Figure 1. Flink 1.5 schedule results
> <粘贴的图形-3.tiff>
>
> Figure 2. Flink 1.9 schedule results
> <粘贴的图形-4.tiff>
>
>
>
> Best
> Weihua Hu
>

Reply | Threaded
Open this post in threaded view
|

Re: [DISCUSS] decentralized scheduling strategy is needed

Till Rohrmann
Thanks for reporting the issue HuWeihua. Choosing the right scheduling
strategy when using Yarn with potentially infinite resources can be quite
hard because you don't know over how many TaskExecutors one should
distribute the tasks. It becomes easier if one can configure the minimum
number of TaskExecutors a cluster should always have. This is currently
being discussed and I hope that we can complete this feature for the next
release.

Cheers,
Till

On Wed, Jan 15, 2020 at 11:29 AM HuWeihua <[hidden email]> wrote:

> Hi, Andrey
>
> Thanks for your response.
>
> I have checked this Jira ticket and I think it can work in standalone mode
> which TaskManager has been started before scheduling tasks.
> But we are currently running flink on yarn in per-job cluster mode.
>
> I noticed that this issue has already been raised. I will keep watching
> this ticket.
>
> Thanks again.
>
> Best
> Weihua Hu
>
> 2020年1月15日 17:53,Andrey Zagrebin <[hidden email]> 写道:
>
> HI HuWeihua,
>
> I think your issue should resolve with 1.9.2 and 1.10 (not released but in
> progress).
> You can check the related Jira ticket [1].
>
> Best,
> Andrey
>
> [1] https://jira.apache.org/jira/browse/FLINK-12122
>
> On Wed, Jan 15, 2020 at 10:08 AM HuWeihua <[hidden email]> wrote:
>
>> Hi, All
>> We encountered some problems during the upgrade from Flink 1.5 to Flink
>> 1.9. Flink's scheduling strategy has changed. Flink 1.9 prefers centralized
>> scheduling, while Flink 1.5 prefers decentralized scheduling. This change
>> has caused resources imbalance and blocked our upgrade plan. We have
>> thousands of jobs that need to be upgraded.
>>
>> For example,
>> There is a job with 10 sources and 100 sinks. Each source need 1 core and
>> each sink need 0.1 core.
>> Try to run this job on Yarn, configure the numberOfTaskSlots is 10,
>> yarn.containers.vcores is 2.
>>
>> When using Flink-1.5:
>> Each TaskManager will run 1 source and 9 sinks, they need 1.9 cores
>> totally. So the job with this configuration works very well. The schedule
>> results is shown in Figure 1.
>> When using Flink-1.9:
>> The 10 sources will be scheduled to one TaskManager  and the 100 sinks
>> will scheduled to other 10 TaskManagers.  The schedule results is shown
>> in Figure 2.
>> In this scenario, the TaskManager which run sources need 10 cores, other
>> TaskManagers need 1 cores. But TaskManager must be configured the same, So
>> we need 11 TaskManager with 10 cores.
>> This situation waste (10-2)*11 = 88 cores more than Flink 1.5.
>>
>> In addition to the waste of resources, we also encountered other problems
>> caused by centralized scheduling strategy.
>>
>>    1. Network bandwidth. Tasks of the same type are scheduled to the one
>>    TaskManager, causing too much network traffic on the machine.
>>
>>
>>    1. Some jobs need to sink to the local agent. After centralized
>>    scheduling, the insufficient processing capacity of the single machine
>>    causes a backlog of consumption.
>>
>>
>> In summary, we think a decentralized scheduling strategy is necessary.
>>
>>
>> Figure 1. Flink 1.5 schedule results
>> <粘贴的图形-3.tiff>
>>
>> Figure 2. Flink 1.9 schedule results
>> <粘贴的图形-4.tiff>
>>
>>
>>
>> Best
>> Weihua Hu
>>
>>
>