Troubleshooting with the Job Comparison Feature
Steps for comparing two different runs of the same job, which is especially useful when you notice unexpected changes. For example, when you have a job that consistently completes within a specific amount of time and then it starts taking longer, comparing two runs of the same job enables you to analyze the differences so that you can troubleshoot the cause.
- In a supported browser, log in to Workload XM.
- In the Search field of the Clusters page, enter the name of the cluster whose workloads you want to analyze.
From the navigation panel under Data Engineering, select
Jobs and then from the time-range list in the Cluster
Summary page, select a time period that meets your requirements.
View the list of jobs that have executed during the selected time period:
The following example reveals that though the
spark-etljob runs often, the last three runs have taken significantly longer. Where, on August 2, the duration was 27 minutes, but on August 3, the duration almost doubled to 51 minutes. The Job Comparison tool will enable you to examine both runs to determine why the duration changed:
Select one of the runs of the
spark-etljob, and then in the Jobs detail page, click the Trends tab:Up to 30 runs prior to the selected job are displayed. Notice that in the Input and Output columns there are different amounts of data processed by the job. For example, on August 2, the job processed 2.4 GB of data and outputted 1.8 GB. However, on August 3, the job processed 4.2 GB, almost twice as much data, and it outputted 4.6 GB. The Job Comparison tool will enable you to examine both runs to determine why the amount of data changed:
To compare two job runs, select the check boxes adjacent to the job runs you
require, in this case the runs for August 2 and August 3 are selected, and then
The Job Comparison page opens displaying more details about each job. For this example's comparison, the tabs that will contain more information are the Structure, Configurations, and the SQL Executions tabs:
The Structure tab page, displays the sub-jobs executed
for both runs of the
spark-etljob:In this example, the job that took 27 minutes only executed 9 sub-jobs and the job that took 51 minutes, almost twice as much time, executed 16 sub-jobs, almost twice as many. Selecting any of the listed sub-jobs displays more details.When the Configurations tab was examined, it revealed that the configurations between the two runs of this job were identical, so a configuration change probably did not cause this anomaly.
When the SQL Executions tab was selected, it revealed
that twice as many Spark queries executed for the job that took the longest