Characterizing a High Throughput Computing Workload: The Compact Muon Solenoid (CMS) Experiment at LHC
Author(s) -
Rafael Ferreira da Silva,
Mats Rynge,
Gideon Juve,
I. Sfiligoi,
Ewa Deelman,
James Letts,
F. Würthwein,
Miron Livny
Publication year - 2015
Publication title -
procedia computer science
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.334
H-Index - 76
ISSN - 1877-0509
DOI - 10.1016/j.procs.2015.05.190
Subject(s) - compact muon solenoid , computer science , large hadron collider , workload , throughput , solenoid , operating system , particle physics , embedded system , physics , wireless , quantum mechanics
High throughput computing (HTC) has aided the scientific community in the analysis of vast amounts of data and computational jobs in distributed environments. To manage these large workloads, several systems have been developed to efficiently allocate and provide access to distributed resources. Many of these systems rely on job characteristics estimates (e.g., job runtime) to characterize the workload behavior, which in practice is hard to obtain. In this work, we perform an exploratory analysis of the CMS experiment workload using the statistical recursive partitioning method and conditional inference trees to identify patterns that characterize particular behaviors of the workload. We then propose an estimation process to predict job characteristics based on the collected data. Experimental results show that our process es-timates job runtime with 75% of accuracy on average, and produces nearly optimal predictions for disk and memory consumption
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom