Plans of the WLCG for Run3 and HL-LHC era
Jose F. Salt Cairols Instituto de Física Corpuscular
XI CPAN DAYS 21-23 October 2019
23/10/2019 XI CPAN Days 1
Plans of the WLCG for Run3 and HL-LHC era Jose F. Salt Cairols - - PowerPoint PPT Presentation
XI CPAN DAYS 21-23 October 2019 Plans of the WLCG for Run3 and HL-LHC era Jose F. Salt Cairols Instituto de Fsica Corpuscular 23/10/2019 XI CPAN Days 1 Overview 1.-The WLCG Global Collaboration 2.-Run 3 and HL/LHC Plan 3.- The Spanish
XI CPAN DAYS 21-23 October 2019
23/10/2019 XI CPAN Days 1
23/10/2019 XI CPAN Days 2
Distributed High- throughput computing infrastructure to store, process and analyze data produced by the LHC experiments. In numbers:
and overlay over NREN s (LHCONE) with 10/100 Gbps links
CERN Computing Center
The equipment purchased by the centers (T0&T1 &T2) give service to the whole collaboration (as a detector) WLCG is a worldwide and non-stop infrastructure
Contributes to the scientific and technological progress
participates in WLCG: scientific infrastructure, expert perssonel, etc
23/10/2019 XI CPAN Days 3
BEST GUESS Run 3:
23/10/2019 XI CPAN Days 4
From I. Bird’s talk at 7th Scientific Computing Forum, 4/10/210 SCF, 4th Oct 2019, CERN
23/10/2019 XI CPAN Days 5
requirements for HL-LHC
23/10/2019 XI CPAN Days 6
23/10/2019 XI CPAN Days 7
Clouds:
The PIC Cloud (ES)
ATLAS, CMS and LHCb
:
○
CMS Spanish Tier2
○
CIEMAT Madrid
○
IFCA Santander
○
ATLAS Spanish Tier2 IFIC Valencia IFAE Barcelona UAM Madrid)
○
LHCb Spanish Tier2
○
USC Santiago de Compostela
○
UB (Universitat de Barcelona=
○
LIP Lisbon, Portugal
○
UTFSM Santiago, Chile
○
UNLP La Paz, Argentina (inactive)
following the ATLAS/CMS/LHCb computing models
Tier-1s ones
23/10/2019 XI CPAN Days 8
Total accounting of Resources: CPU (HS06) =182K Disk (PB) = 14.5 Tape (PB) = 19.6
LCG-ES
More than 22 million finished jobs On average, 5000 slots occupied by running jobs daily More than 196 million events proccessed More than 46 million files produced
9 23/10/2019 XI CPAN Days
– Euro HPC Beur funding 2 aprox 200 PFlps machines by 2021, 2 EXaFlops by 2024
23/10/2019 XI CPAN Days 10
– Recommendation of using the computing resources of BSC coming from Funding Agency – ATLAS: : effort devoted to addapt the queues at BSC to run simulation production jobs . In 2018, start to call for computing time (IFIC, IFAE) and several requests have been granted
Supercomputing Network (RES) and Europe (PRACE), being granted for the IFAE 2.8 M hours and IFIC 1.2M hours in the Mare Nostrum (BSC) and 2M hours in Lusitania (Cenit)
execution of simulation work of the ATLAS detector in these HPCs, so in this way we have used resources outside the Spanish Tiers centers. We have simulated more than 60 million event
ATLAS simulation when profiting of
resources
millions of events simulated
jobs ended successfully
– CMS:
CMS, we still cannot use them due to the lack of network connectivity from the nodes, which is necessary in CMS to integrate them into the WMS. There is a project with the HTCondor team to address that limitation.
Adaptation of ALTAMIRA (node of RES in Cantabria) within the GRID Infrastructure (input de Ibán) – The grid infrastructure of the T2 has been redesigned so that when the T2 is saturated, check the availability of free HPC resources and forward them there. At the moment pilot examples are operating using altamira in "parasitic" mode, but it can be easily changed.
at the spanish level the LHCb groups have not started with these activities yet
23/10/2019 XI CPAN Days 11
Take the example of another special ‘project ‘agreement with BSC
users (hopefully)
23/10/2019 XI CPAN Days 12
Meeting at BSC in December 2018
View of Mare Nostrum
23/10/2019 XI CPAN Days 13
Experiments have run large scale tests using Cloud compute nodes Google Cloud, Amazon AWS, Microsoft Azure
=>Commercial cloud is
computing,
without investing ⇒ Currentely essentially no commercial cloud use for LHC computing ⇒ Potential future opportunties:
European Open Science Cloud (EOSC) A EU model for use of cloud computing in the private and public sector
23/10/2019 XI CPAN Days 14
European Science cluster of Astronomy & Particle Physics
ESFRI Research Infrastructure
23/10/2019 XI CPAN Days 15
23/10/2019 XI CPAN Days 16
23/10/2019 XI CPAN Days 17
– The idea is localize bulk data in a cloud service (data lake): minimize replication, assure availability – Serve data to remote ( or local) compute grid, cloud, HPC, ??? – Simple caching is all that is needed at compute site (or none, if fast network) – Federated data at national, regional, global scales
23/10/2019 XI CPAN Days 18
PIC is contributing actively in the first group with studies in Data
Access and Popularity for a CMS at PIC and CIEMA measuring the effect on the applications to real data in a remote way
– 50 millions of lines of code mainly C++ – “a project / experiment cannot afford to have bad software” (Graeme’s talk in Granada)
– HEP Software Foundation – IRIS-HEP: Institute for Research & Innovation in Software for HEP, 25M$, 5 years – Proposal a EU Scientific Software Institute – In Spain: COMCHA forum
– High level parallelism , new instructions sets,… – Support in software frameworks for heterogenous hardware
– Machine Learning/Deep Learning – Rewrite physics algorithms for new hardware
23/10/2019 XI CPAN Days 19
Improvement in CPU consumption by using faster phyisics algortithms in FASTSIM/FASTRECO
23/10/2019 XI CPAN Days 20
DUNE and CTA will leverage the WLCG for its Computing Infrastructure Nuclear Physics Coll: ESCAPE address FAIR data management The LHC Computing Model has been adapated to the needs and the size of AGATA collaboration Computing @ Future Accelerators
Meeting May 2019: Addressing the outstanding questions CLIC and Future Circular Cilliders
23/10/2019 XI CPAN Days 21
23/10/2019 XI CPAN Days 22
23/10/2019 XI CPAN Days 23