Orateur
M.
Dario Barberis
(Genoa University/INFN)
Description
The ATLAS experiment is taking data steadily since Autumn 2009, collecting
close to 1 fb-1 of data (several petabytes of raw and reconstructed data
per year of data-taking). Data are calibrated, reconstructed, distributed
and analysed at over 100 different sites using the World-wide LHC
Computing Grid and the tools produced by the ATLAS Distributed Computing
project.
In addition to event data, ATLAS produces a wealth of information on
detector status,luminosity, calibrations, alignments, and data processing
conditions. This information is stored in relational databases, online and
offline, and made transparently available to analysers of ATLAS data
world-wide through an infrastructure consisting of distributed database
replicas and web servers that exploit caching technologies. This paper
reports on the experience of using this distributed computing
infrastructure with real data and in real time, on the evolution of the
computing model driven by this experience, and on the system performance
during the first two years of operation
Auteur principal
Naoko Kanaya
(ICEPP,University of Tokio)