Extending the distributed computing infrastructure of the CMS experiment with HPC resources

J. Adelman-McCarthy,T. Boccali, R. Caspart,A. Delgado Peris,M. Fischer, J. Flix Molina,M. Giffels,J. M. Hernandez, D. Hufnagel, E. Kuhn, T. Madlener, A. K. Mohapatra, H. Ozturk,A. Perez-Calero Yzquierdo, D. Spiga, C. Wissing

Journal of Physics: Conference Series(2023)

引用 1|浏览21
暂无评分
摘要
Particle accelerators are an important tool to study the fundamental properties of elementary particles. Currently the highest energy accelerator is the LHC at CERN, in Geneva, Switzerland. Each of its four major detectors, such as the CMS detector, produces dozens of Petabytes of data per year to be analyzed by a large international collaboration. The processing is carried out on the Worldwide LHC Computing Grid, that spans over more than 170 compute centers around the world and is used by a number of particle physics experiments. Recently the LHC experiments were encouraged to make increasing use of HPC resources. While Grid resources are homogeneous with respect to the used Grid middleware, HPC installations can be very different in their setup. In order to integrate HPC resources into the highly automatized processing setups of the CMS experiment a number of challenges need to be addressed. For processing, access to primary data and metadata as well as access to the software is required. At Grid sites all this is achieved via a number of services that are provided by each center. However at HPC sites many of these capabilities cannot be easily provided and have to be enabled in the user space or enabled by other means. At HPC centers there are often restrictions regarding network access to remote services, which is again a severe limitation. The paper discusses a number of solutions and recent experiences by the CMS experiment to include HPC resources in processing campaigns.
更多
查看译文
关键词
distributed computing,hpc resources,cms experiment,infrastructure
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要