Online Data Handling And Storage At The Cms Experiment

J-M Andre,A. Andronidis,U. Behrens,J. Branson, O. Chaze, S. Cittolin,G-L Darlea, C. Deldicque,Z. Demiragli,M. Dobson,A. Dupont,S. Erhan, D. Gigi, F. Glege, G. Gomez-Ceballos,J. Hegeman,A. Holzner, R. Jimenez-Estupinan, L. Masetti,F. Meijers,E. Meschi,Rk Mommsen,S. Morovic, C. Nunez-Barranco-Fernandez, V. O'Dell,L. Orsini, C. Paus,A. Petrucci, M. Pieri,A. Racz,P. Roberts, H. Sakulin, C. Schwick,B. Stieger, K. Sumorok,J. Veverka,S. Zaza,P. Zejdl

21ST INTERNATIONAL CONFERENCE ON COMPUTING IN HIGH ENERGY AND NUCLEAR PHYSICS (CHEP2015), PARTS 1-9(2015)

引用 24|浏览15
暂无评分
摘要
During the LHC Long Shutdown 1, the CMS Data Acquisition (DAQ) system underwent a partial redesign to replace obsolete network equipment, use more homogeneous switching technologies, and support new detector back-end electronics. The software and hardware infrastructure to provide input, execute the High Level Trigger (HLT) algorithms and deal with output data transport and storage has also been redesigned to be completely file based. All the metadata needed for bookkeeping are stored in files as well, in the form of small documents using the JSON encoding. The Storage and Transfer System (STS) is responsible for aggregating these files produced by the HLT, storing them temporarily and transferring them to the TO facility at CERN for subsequent offline processing. The STS merger service aggregates the output files from the HLT from X62 sources produced with an aggregate rate of r`-' 2 G B / s An estimated bandwidth of 7GB/s in concurrent read/write mode is needed. Furthermore, the STS has to be able to store several days of continuous running, so an estimated of 250TB of total usable disk space is required. In this article we present the various technological and implementation choices of the three components of the STS: the distributed file system, the merger service and the transfer system.
更多
查看译文
关键词
general
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要