Abstract

The ATLAS Trigger and Data Acquisition system is responsible for the online processing of live data, streaming from the ATLAS experiment at the Large Hadron Collider at CERN. The online farm is composed of ∼3000 servers, processing the data read out from ∼100 million detector channels through multiple trigger levels. During the two years of the first Long Shutdown there has been a tremendous amount of work done by the ATLAS Trigger and Data Acquisition System Administrators, implementing numerous new software applications, upgrading the OS and the hardware, changing some design philosophies and exploiting the High- Level Trigger farm with different purposes. The OS version has been upgraded to SLC6; for the largest part of the farm, which is composed of net booted nodes, this required a completely new design of the net booting system. In parallel, the migration to Puppet of the Configuration Management systems has been completed for both net booted and local booted hosts; the Post-Boot Scripts system and Quattor have been consequently dismissed. Virtual Machine usage has been investigated and tested and many of the core servers are now running on Virtual Machines. Virtualisation has also been used to adapt the High-Level Trigger farm as a batch system, which has been used for running Monte Carlo production jobs that are mostly CPU and not I/O bound. Finally, monitoring the health and the status of ∼3000 machines in the experimental area is obviously of the utmost importance, so the obsolete Nagios v2 has been replaced with Icinga, complemented by Ganglia as a performance data provider. This paper serves for reporting of the actions taken by the Systems Administrators in order to improve and produce a system capable of performing for the next three years of ATLAS data taking.

Highlights

  • After three years of Large Hadron Collider (LHC) beam (Run 1) and two years of upgrades performed during the Long Shutdown 1 (LS1), the LHC has officially restarted.During the time of the LS1, a tremendous amount of work has been done by the ATLAS [1] Trigger and Data Acquisition (TDAQ) Systems Administrators (SysAdmins)

  • Software applications have been implemented, operating systems (OS) and most of the hardware has been upgraded, some design philosophies have been changed and the SysAdmins have exploited the High-Level Trigger (HLT) [2] computing farm to be used as a batch system to run specialised offline jobs

  • There is one Local File Servers (LFS) installed per rack for the HLT farm, with a couple being shared between various sub-detector systems

Read more

Summary

Introduction

After three years of Large Hadron Collider (LHC) beam (Run 1) and two years of upgrades performed during the Long Shutdown 1 (LS1), the LHC has officially restarted. During the time of the LS1, a tremendous amount of work has been done by the ATLAS [1] Trigger and Data Acquisition (TDAQ) Systems Administrators (SysAdmins).

Primary Author
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.