Bhagavan created FLINK-17969:
--------------------------------
Summary: Enhance Flink (Task) logging to include job name as context diagnostic information
Key: FLINK-17969
URL:
https://issues.apache.org/jira/browse/FLINK-17969 Project: Flink
Issue Type: Improvement
Components: Runtime / Task
Affects Versions: 1.10.0
Reporter: Bhagavan
Problem statement:
We use a shared session cluster (Standalone/Yarn) to execute jobs. All logs from the cluster are shipped using log aggregation framework (Logstash/Splunk) so that application diagnostic is easier.
However, we are missing one vital information in the logline. i.e. Job name so that we can filter the logs for a single job.
Background
Currently, Flink logging uses SLF4J as API to abstract away from concrete logging implementation (log4j 1.x, Logback or log4j2) and configuration of logging pattern and implementation can be configured at deployment, However, there is no MDC info from framework indicating job context.
Proposed improvement.
Add jobName field to Task class so that we can add it as MDC when task thread starts executing.
Change is trivial and uses SLF4J MDC API.
With this change, user can customise logging pattern to include MDC (e.g. in Logback [%X{jobName}])
Change required.
{code:java}
@@ -319,6 +323,7 @@ public class Task implements Runnable, TaskSlotPayload, TaskActions, PartitionPr
this.jobId = jobInformation.getJobId();
+ this.jobName = jobInformation.getJobName();
this.vertexId = taskInformation.getJobVertexId();
@@ -530,8 +535,10 @@ public class Task implements Runnable, TaskSlotPayload, TaskActions, PartitionPr
@Override
public void run() {
try {
+ MDC.put("jobName", this.jobName);
doRun();
} finally {
+ MDC.remove("jobName");
terminationFuture.complete(executionState);
}
}
{code}
if we are in agreement for this small change. Will raise PR.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)