Canbin Zheng created FLINK-15843:
------------------------------------
Summary: Do not violently kill TaskManagers
Key: FLINK-15843
URL:
https://issues.apache.org/jira/browse/FLINK-15843 Project: Flink
Issue Type: Sub-task
Components: Deployment / Kubernetes
Affects Versions: 1.10.0
Reporter: Canbin Zheng
Fix For: 1.11.0
The current solution of stopping a TaskManager instance when JobManager sends a deletion request is by directly calling ${\{KubernetesClient.pods().withName().delete}}, thus that instance would be violently killed with a _KILL_ signal and having no chance to clean up, which could cause problems because we expect the process to gracefully terminate when it is no longer needed.
Refer to the guide of [Termination of Pods|[
https://kubernetes.io/docs/concepts/workloads/pods/pod/#termination-of-pods]|
https://kubernetes.io/docs/concepts/workloads/pods/pod/#termination-of-pods],], we know that on Kubernetes a _TERM_ signal would be first sent to the main process in each container, and may be followed up with a force _KILL_ signal if the grace period has expired; the Unix signal will be sent to the process which has PID 1 ([Docker Kill|
https://docs.docker.com/engine/reference/commandline/kill/]), however, the TaskManagerRunner Process is spawned by {color:#172b4d}/opt/flink/bin/kubernetes-entry.sh {color}and could never have PID 1, so it would never receive the Unix signal_._
One walk around could be that JobManager firstly sends a *KILL_WORKER* message to the TaskManager, then the TaskManager gracefully terminates itself to ensure that the clean-up is completely finished, lastly, the JobManager deletes the Pod after a configurable graceful shut-down period.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)