Abstract

The LHCb experiment records millions of proton collisions every second, but only a fraction of them are useful for LHCb physics.In order to filter out the "bad events" a large farm of x86-servers (~2000 nodes) has been put in place. These servers boot from and run from NFS, however they use their local disk to temporarily store data, which cannot be processed in real-time ("data-deferring"). These events are subsequently processed, when there are no live-data coming in. The effective CPU power is thus greatly increased. This gain in CPU power depends critically on the availability of the local disks. For cost and power-reasons, mirroring (RAID-1) is not used, leading to a lot of operational headache with failing disks and disk-errors or server failures induced by faulty disks.To mitigate these problems and increase the reliability of the LHCb farm, while at same time keeping cost and power-consumption low, an extensive research and study of existing highly available and distributed file systems has been done. While many distributed file systems are providing reliability by "file replication", none of the evaluated ones supports erasure algorithms.A decentralised, distributed and fault-tolerant "write once read many" file system has been designed and implemented as a proof of concept providing fault tolerance without using expensive – in terms of disk space – file replication techniques and providing a unique namespace as a main goals.This paper describes the design and the implementation of the Erasure Codes File System (ECFS) and presents the specialised FUSE interface for Linux.Depending on the encoding algorithm ECFS will use a certain number of target directories as a backend to store the segments that compose the encoded data. When target directories are mounted via nfs/autofs – ECFS will act as a file-system over network/block-level raid over multiple servers.

Highlights

  • Event Filter Farm ~2000 nodesTomasz Rybczynski, Enrico Bonaccorsi, Niko Neufeld CHEP 2013High Level TriggerEvent Filter Farm Requirements Fault tolerance – data redundancy Single namespace POSIX semantics Write once read many Sufficient performanceSearching for a solution Linux RAID over HA cluster GlusterFS Tahoe-LAFS tahoe-lafs.org ...Erasure Codes File SystemECFS - distributed, fault-tolerant file systemProject based on: FUSE NFS Erasure Codes

  • Successful decoding possible if the number of missing symbols is less or equal n Code word with some symbols missing

  • ECFS meets technical requirements of the project

Read more

Summary

Introduction

A decentralized, distributed and fault-tolerant FUSE filesystem for the LHCb online farm Session: Data Stores, Data Bases, and Storage Systems

Results
Conclusion
Full Text
Paper version not known

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call

Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.