Abstract

The computing infrastructures serving the LHC experiments have been designed to cope at most with the average amount of data recorded. The usage peaks, as already observed in Run-I, may however originate large backlogs, thus delaying the completion of the data reconstruction and ultimately the data availability for physics analysis. In order to cope with the production peaks, the LHC experiments are exploring the opportunity to access Cloud resources provided by external partners or commercial providers. In this work we present the proof of concept of the elastic extension of a local analysis facility, specifically the Bologna Tier-3 Grid site, for the LHC experiments hosted at the site, on an external OpenStack infrastructure. We focus on the Cloud Bursting of the Grid site using DynFarm, a newly designed tool that allows the dynamic registration of new worker nodes to LSF. In this approach, the dynamically added worker nodes instantiated on an OpenStack infrastructure are transparently accessed by the LHC Grid tools and at the same time they serve as an extension of the farm for the local usage.

Highlights

  • The physics experiments operating at the Large Hadron Collider at CERN are collecting large amount of collision data and in the future the overall data volume is foreseen to steadily increase

  • Data processing requires large computing resources, the data processing campaigns are often not constant over the time, having periods of peak usage when resources usage greatly increases with respect to periods of standard usage

  • Instead of buying new resources to be fully used only for short time periods, a possible solution could be to access external Cloud resources provided both by external partners and commercial providers, operating in the so called Cloud Bursting mode. In this contribution we describe a mechanism based on the DynFarm [4] software to dynamically extend the Bologna local farm of the INFN-BOLOGNA-T3 data center to the external cloud facility of the INFN-CNAF both for the ATLAS and CMS experiments

Read more

Summary

Introduction

The physics experiments (such ATLAS [1] and CMS [2]) operating at the Large Hadron Collider at CERN are collecting large amount of collision data and in the future the overall data volume is foreseen to steadily increase. Data processing requires large computing resources, the data processing campaigns are often not constant over the time, having periods of peak usage when resources usage greatly increases with respect to periods of standard usage During those peak periods, the world-wide distributed WLCG [3] computing centers may run over the pledged resources and generate long backlog queues to completely absorb the computing workload. Instead of buying new resources to be fully used only for short time periods, a possible solution could be to access external Cloud resources provided both by external partners and commercial providers, operating in the so called Cloud Bursting mode In this contribution we describe a mechanism based on the DynFarm [4] software to dynamically extend the Bologna local farm of the INFN-BOLOGNA-T3 data center to the external cloud facility of the INFN-CNAF both for the ATLAS and CMS experiments.

Methods
Findings
Conclusion

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.