College of Computer, National University of Defense Technology, Changsha410073, China
Show Author Information
Hide Author Information
Abstract
Data-parallel computing platforms, such as Hadoop and Spark, are deployed in computing clusters for big data analytics. There is a general tendency that multiple users share the same computing cluster. The schedule of multiple jobs becomes a serious challenge. Over a long period in the past, the Shortest-Job-First (SJF) method has been considered as the optimal solution to minimize the average job completion time. However, the SJF method leads to a low system throughput in the case where a small number of short jobs consume a large amount of resources. This factor prolongs the average job completion time. We propose an improved heuristic job scheduling method, called the Densest-Job-Set-First (DJSF) method. The DJSF method schedules jobs by maximizing the number of completed jobs per unit time, aiming to decrease the average Job Completion Time (JCT) and improve the system throughput. We perform extensive simulations based on Google cluster data. Compared with the SJF method, the DJSF method decreases the average JCT by 23.19% and enhances the system throughput by 42.19%. Compared with Tetris, the job packing method improves the job completion efficiency by 55.4%, so that the computing platforms complete more jobs in a short time span.
J.Dean and S.Ghemawat, MapReduce: Simplified data processing on large clusters, in Proc. 6th Conf. Operating System Design & Implementation, San Francisco, CA, USA, 2004, pp. 137-150.
[3]
M.Zaharia, M.Chowdhury, M. J.Franklin, S.Shenker, and I.Stoica, Spark: Cluster computing with working sets, in Proc. 2nd USENIX Workshop on Hot Topics in Cloud Computing, Boston, MA, USA, 2020, pp. 1-10.
[4]
J. C.Tang, M.Xu, S. J.Fu, and K.Huang, A scheduling optimization technique based on reuse in spark to defend against apt attack, Tsinghua Science and Technology, vol. 23, no. 5, pp. 550-560, 2018.
R.Grandl, M.Chowdhury, A.Akella, and G.Ananthanarayanan, Altruistic scheduling in multi-resource clusters, in Proc. 12th USENIX Conf. Operating Systems Design and Implementation, Savannah, GA, USA, 2016, pp. 65-80.
[6]
R.Grandl, G.Ananthanarayanan, S.Kandula, S.Rao, and A.Akella, Multi-resource packing for cluster schedulers, in Proc. 2014 ACM Special Interest Group on Data Communication, Chicago, IL, USA, 2014, pp. 455-466.
B.Hindman, A.Konwinski, M.Zaharia, A.Ghodsi, A. D.Joseph, R.Katz, S.Shenker, and I.Stoica, Mesos: A platform for fine-grained resource sharing in the data center, in Proc. 8th USENIX Conf. on Networked Systems Design and Implementation, Boston, MA, USA, 2011, pp. 429-483.
[9]
V. K.Vavilapalli, A. C.Murthy, C.Douglas, S.Agarwal, M.Konar, R.Evans, T.Graves, J.Lowe, H.Shah, S.Seth, et al., Apache hadoop YARN: Yet another resource negotiator, in Proc. 4th Annual Symp. Cloud Computing, Santa Clara, CA, USA, 2013, pp. 1-16.
[10]
T.Bonald, L.Massoulié, A.Proutière, and J.Virtamo, A queueing analysis of max-min fairness, proportional fairness and balanced fairness, Queueing Syst., vol. 53, nos. 1&2, pp. 65-84, 2006.
A.Ghodsi, M.Zaharia, B.Hindman, A.Konwinski, S.Shenker, and I.Stoica, Dominant resource fairness: Fair allocation of multiple resource types, in Proc. 8thUSENIX Conf. Networked Systems Design and Implementation, Boston, MA, USA, 2013, pp. 323-336.
[12]
J. A.Hartigan and M. A.Wong, Algorithm AS 136: A K-means clustering algorithm, J. Roy. Stat. Soc. Ser. C (Appl. Stat.), vol. 28, no. 1, pp. 100-108, 1979.
M. K.Pakhira, A linear time-complexity k-means algorithm using cluster shifting, in Proc. 2014 IEEE Int. Conf. Computational Intelligence and Communication Networks, Bhopal, India, 2014, pp. 1047-1051.
[14]
J. O.Iglesias, L.Murphy, M. DeCauwer, D.Mehta, and B.O’Sullivan, A methodology for online consolidation of tasks through more accurate resource estimations, in Proc. 2014 IEEE ACM 7thInt. Conf. Utility and Cloud Computing, London, UK, 2014, pp. 89-98.
[15]
P.Janus and K.Rzadca, SLO-aware colocation of data center tasks based on instantaneous processor requirements, in Proc. 2017 Symp. Cloud Computing, Santa Clara, CA, USA, 2017, pp. 256-268.
[16]
M.Carvalho, D. A.Menascé, and F.Brasileiro, Capacity planning for IaaS cloud providers offering multiple service classes, Future Gener. Comput. Syst., vol. 77, no. 4, pp. 97-111, 2017.
Z. Y.Hu, D. S.Li, and D. K.Guo, Balance resource allocation for spark jobs based on prediction of the optimal resources, Tsinghua Science and Technology, vol. 25, no. 4, pp. 487-497, 2020.
S.Venkataraman, Z. H.Yang, M.Franklin, B.Recht, and I.Stoica, Ernest: Efficient performance prediction for large-scale advanced analytics, in Proc. 13th USENIX Conf. Networked Systems Design and Implementation, Santa Clara, CA, USA, 2016, pp. 363-378.
[19]
Z. D.Bei, Z. B.Yu, H. L.Zhang, W.Xiong, C. Z.Xu, L.Eeckhout, and S. Z.Feng, RFHOC: A random-forest approach to auto-tuning Hadoop’s configuration, IEEE Transactions on Parallel and Distributed Systems, vol. 27, no. 5, pp. 1470-1483, 2016.
Z. B.Yu, Z. D.Bei, and X. H.Qian, Datasize-aware high dimensional configurations auto-tuning of in-memory cluster computing, in Proc. 23rd ACM Int. Conf. Architectural Support for Programming Languages and Operating Systems, Williamsburg, VA, USA, 2018, pp. 564-577.
[21]
Z. Y.Hu, D. S.Li, D. X.Zhang, and Y. X.Chen, ReLoca: Optimize resource allocation for data-parallel jobs using deep learning, in Proc. IEEE Conf. Computer Communications, Toronto, Canada, 2020, pp. 1163-1171.
[22]
Y.Dong, J.Chen, Y.Tang, J. J.Wu, H. Q.Wang, and E. Q.Zhou, Lazy scheduling based disk energy optimization method, Tsinghua Science and Technology, vol. 25, no. 2, pp. 203-216, 2020.
W.Zhang, X.Chen, and J. H.Jiang, A multi-objective optimization method of initial virtual machine fault-tolerant placement for star topological data centers of cloud systems, Tsinghua Science and Technology, vol. 26, no. 1, pp. 95-111, 2021.
D.Shen, J. Z.Luo, F.Dong, and J. X.Zhang, VirtCo: Joint coflow scheduling and virtual machine placement in cloud data centers, Tsinghua Science and Technology, vol. 24, no. 5, pp. 630-644, 2019.
C.Xue, C.Lin, and J.Hu, Scalability analysis of request scheduling in cloud computing, Tsinghua Science and Technology, vol. 24, no. 3, pp. 249-261, 2019.
Hu Z, Li D. Improved Heuristic Job Scheduling Method to Enhance Throughput for Big Data Analytics. Tsinghua Science and Technology, 2022, 27(2): 344-357. https://doi.org/10.26599/TST.2020.9010047
The articles published in this open access journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/).
10.26599/TST.2020.9010047.F001
Comparison of the SJF method and a heuristic job scheduling method. The heuristic job scheduling method decreases the average JCT by 32.07.
10.26599/TST.2020.9010047.F002
Workflow of the improved heuristic job scheduling method.
10.26599/TST.2020.9010047.F003
Performance of Tetris, CS, FS, and DRF methods in YARN[
6
].
3.1 Concept of job completion efficiency
To decrease the average JCT, we further explore the cause why the SJF method performs worse than the heuristic method, as shown in
Fig. 1
. If jobs with short JCT may consume a large amount of resources, the SJF method prolongs the completion of the majority of jobs, which have a request for a small amount of resources. To sum up, the SJF method does not consider the resources capacity of the cluster. As a result, the number of parallel jobs may decrease. To address this problem, we propose an improved heuristic job scheduling method to pack as many jobs as possible into the job set and then determine the scheduling priority of the job set according to the job completion efficiency. The job completion efficiency is defined as follows.
Definition 1 (job completion efficiency). Given that a batch of jobs , , , are executed in parallel and packed into the set of jobs . Let denote the time span that all jobs in are completed. The job completion efficiency is as follows:
where is the number of jobs.
In
Fig. 1
, the heuristic method packs Jobs 2-5 together. We regard such packed jobs as the set of jobs. Although the heuristic method takes 4.2 s to complete jobs in the job set, the number of completed jobs is four. This achieves a high job completion efficiency (up to approximately 0.95). By contrast, the SJF method prefers to execute Job 1. This condition leads to a low job completion efficiency (i.e., 0.25). To sum up, we leverage the job completion efficiency as a priority to schedule the set of jobs. Definition 1 indicates that the value of the job completion efficiency is equivalent to the system throughput, i.e., how many jobs are completed in a given amount of time. Next, we prove that such a scheduling order, which is determined according to the job completion efficiency, achieves the smallest average JCT.
Theorem 1 Given a batch of jobs, we pack them into two sets of jobs and . Note that and are executed in a successive order. If the job set with a higher job completion efficiency for scheduling, then the average JCT is smaller. Let and denote the job completion efficiency of and , respectively. If , then the scheduling order achieves a smaller average JCT than another scheduling order .
Proof We list the information about the set of jobs in
Table 1
. and denote the number of jobs in and , respectively. and denote the time span of and , respectively.
10.26599/TST.2020.9010047.T001
Example for the scheduling of two job sets.
Job set
Number of jobs
Time span
Case 1 If a scheduler prefers to schedule the job set with a higher job completion efficiency, then the scheduling preference is . The completion time of is . The sum of JCT of all jobs in is approximately . Similarly, the completion time of is . The sum of JCT in is approximately . In this case, the sum of JCT in and is
Case 2 If a scheduler prefers to schedule the job set with a smaller job completion efficiency, then the scheduling preference is . The completion time of is . The sum of JCT of all jobs in is approximately . Similarly, the completion time of is . The sum of JCT in is approximately . In this case, the sum of JCT in and is
Because has a larger job completion efficiency than , we have . Hence, we have . We compare Formula (
2
) with Formula (
3
). The sum of JCT in Case 1 is smaller than that in Case 2. We make a conclusion that for any two job sets and , the job set with a larger job completion efficiency should be preferred for scheduling to achieve a smaller average JCT.
We extend the above conclusion to a new case where three job sets are scheduled. To extend the conclusion, we introduce the following lemma.
Lemma 1 Given that any two job sets and can be executed serially, the two job sets are scheduled in a successive order. We regard and as a new job set, denoted by . Let and denote the job completion efficiency of and , respectively. We have
Proof Let and denote the number of jobs in and , respectively. Let and denote the time span of and , respectively. We have and . For the job set , the number of jobs in is . The time span of is . Thus, the job completion efficiency of is . Assume that the job completion efficiency of is larger than that of . Owing to , i.e., , we have . Hence, we have . The following formula holds:
Hence, is proven.
Similarly, because , we have ,
Hence, is proven.
Next, we expand Theorem 1 to the case where over three job sets, i.e., , , and , are scheduled. We sort them by the value of the job completion efficiency. Assume that . According to Lemma 1, we regard and as a new set of jobs. Let denote the new job set. denotes the job completion efficiency of . According to Lemma 1, we have . Then the scheduling problem becomes the case in Theorem 1, where two job sets, and , are scheduled. According to Theorem 1, we prefer to schedule because the job completion efficiency of is larger than those of the other job sets. Similarly, after scheduling , we prefer to schedule rather than . The final scheduling preference is , which achieves the smallest average JCT.
To sum up, Theorem 1 can be recursively applied in multiple-job-set cases. We make the following general theorem.
Theorem 2 Given multiple job sets, the job set with the largest job completion efficiency should be scheduled with the highest priority to decrease the average JCT.
10.26599/TST.2020.9010047.F004
Illustrative example of low job completion efficiency that is incurred by Tetris. In the example, the heuristic job packing method can achieve a high job completion efficiency.
10.26599/TST.2020.9010047.F005
Illustrative example: Cumulative distribution curve of JCT.
10.26599/TST.2020.9010047.F006
Number of jobs changes with real JCT.
The job packing method adopts the -means clustering algorithm to group a batch of JCT according to the length of JCT. The -means clustering algorithm is a well-known data mining and machine learning tool, which is used to assign data points (i.e., the JCT of multiple jobs) into groups without any prior knowledge of those relationships. The -means clustering algorithm attempts to learn the difference between data points and finds which group, a data point belongs to. The -means clustering algorithm is one of the simplest clustering techniques, which is widely used in biometrics, medical imaging, and related fields. The -means clustering algorithm determines the group where data points should be assigned into rather than the user instructing to build the group of data points based on heuristic rules.
We are motivated to leverage the -means clustering algorithm[
12
] to group jobs with different JCT. Algorithm 2 shows the main steps of the JCT clustering algorithm. The -means clustering based job packing method places a job to the group of jobs if the sum of the difference between the centroid of the group and the JCT of the jobs is at a minimum. A less variation in the group results in the jobs within the group to have similar JCT. The -means based JCT clustering algorithm is known to have a time complexity of [
13
].
In addition to the JCT alignment, the job packing method should increase the number of jobs packed in the job set. Given a group of jobs, the job packing method prefers to pack jobs that request a small amount of resources. This method aims to enhance the number of parallel jobs in the job set. In
Fig. 5
, is the number of jobs, of which the JCT is smaller than . Similarly, is the number of jobs, of which the JCT is smaller than . Assume that we pack the jobs, of which the JCT is larger than but smaller than , into a new job set. The number of jobs in the job set is . The time span of the job set is . Thus, the job completion efficiency of the job set is . To enhance the value of the job completion efficiency, the job packing method maximizes the value of .
Algorithm 3 shows the main steps of the job packing method. After the JCT clustering method outputs the group of jobs, the job packing method packs jobs into the job set. The job packing method calculates the average JCT of each group. Then the job group that has smaller average JCT is preferred to be used in Algorithm 3. Within each group, the job packing method sorts jobs by the amount of resources. Jobs that request a small amount of resources are packed into the job set with high priorities. If the remaining resources is not enough for packing a job into the job set, then the job is packed into the next job set. When the last job is packed, the job packing method stops. Algorithm 3 has two For loops. The complexity of the job packing method is , where is the number of groups.
The job packing method adopts the -means clustering algorithm to group a batch of JCT according to the length of JCT. The -means clustering algorithm is a well-known data mining and machine learning tool, which is used to assign data points (i.e., the JCT of multiple jobs) into groups without any prior knowledge of those relationships. The -means clustering algorithm attempts to learn the difference between data points and finds which group, a data point belongs to. The -means clustering algorithm is one of the simplest clustering techniques, which is widely used in biometrics, medical imaging, and related fields. The -means clustering algorithm determines the group where data points should be assigned into rather than the user instructing to build the group of data points based on heuristic rules.
We are motivated to leverage the -means clustering algorithm[
12
] to group jobs with different JCT. Algorithm 2 shows the main steps of the JCT clustering algorithm. The -means clustering based job packing method places a job to the group of jobs if the sum of the difference between the centroid of the group and the JCT of the jobs is at a minimum. A less variation in the group results in the jobs within the group to have similar JCT. The -means based JCT clustering algorithm is known to have a time complexity of [
13
].
In addition to the JCT alignment, the job packing method should increase the number of jobs packed in the job set. Given a group of jobs, the job packing method prefers to pack jobs that request a small amount of resources. This method aims to enhance the number of parallel jobs in the job set. In
Fig. 5
, is the number of jobs, of which the JCT is smaller than . Similarly, is the number of jobs, of which the JCT is smaller than . Assume that we pack the jobs, of which the JCT is larger than but smaller than , into a new job set. The number of jobs in the job set is . The time span of the job set is . Thus, the job completion efficiency of the job set is . To enhance the value of the job completion efficiency, the job packing method maximizes the value of .
Algorithm 3 shows the main steps of the job packing method. After the JCT clustering method outputs the group of jobs, the job packing method packs jobs into the job set. The job packing method calculates the average JCT of each group. Then the job group that has smaller average JCT is preferred to be used in Algorithm 3. Within each group, the job packing method sorts jobs by the amount of resources. Jobs that request a small amount of resources are packed into the job set with high priorities. If the remaining resources is not enough for packing a job into the job set, then the job is packed into the next job set. When the last job is packed, the job packing method stops. Algorithm 3 has two For loops. The complexity of the job packing method is , where is the number of groups.
10.26599/TST.2020.9010047.F007
JCT distribution of 1000 jobs that are used in simulation.
10.26599/TST.2020.9010047.F008
JCT distribution of 1000 jobs that are completed by Tetris, SJF method, and DJSF method. The majority of jobs that are completed by the DJSF method have shorter JCT.
10.26599/TST.2020.9010047.F009
JCT reduction of DJSF method against SJF method and Tetris.
10.26599/TST.2020.9010047.F010
System throughput when the scheduling time increases.
10.26599/TST.2020.9010047.F011
Job completion efficiency of different job sets. The average job completion efficiency of Tetris and k-means job packing method is 0.0426 and 0.0662, respectively.
10.26599/TST.2020.9010047.F012
Normalized resources usage of the CPU, memory, and disk space when 1000 jobs are scheduled by Tetris, SJF method, and DJSF method, respectively.