Elisabetta Pennacchio, IPNL WA105 Collaboration Meeting, March 23 rd - - PowerPoint PPT Presentation

elisabetta pennacchio ipnl wa105 collaboration meeting
SMART_READER_LITE
LIVE PREVIEW

Elisabetta Pennacchio, IPNL WA105 Collaboration Meeting, March 23 rd - - PowerPoint PPT Presentation

Development and implementation of the WA105 6x6x6 online storage/processing on the 3x1x1 online storage and processing small scale test farm Elisabetta Pennacchio, IPNL WA105 Collaboration Meeting, March 23 rd , 2017 1 1 Outline 1. Online


slide-1
SLIDE 1

1 1

Elisabetta Pennacchio, IPNL

WA105 Collaboration Meeting, March 23rd , 2017

Development and implementation of the WA105 6x6x6

  • nline storage/processing on the 3x1x1 online storage

and processing small scale test farm

slide-2
SLIDE 2

2

Outline

  • 1. Online storage/processing motivation
  • 2. Implementation on the 3x1x1
  • 3. Data Availability at CERN
  • 4. On going activities for the 6x6x6 online processing and storage farm
  • 5. General information concerning 6x6x6 offline processing

Points 1. 2. 3. have already been discussed at different SB meetings: https://indico.fnal.gov/conferenceDisplay.py?confId=13286 https://indico.fnal.gov/conferenceDisplay.py?confId=13769

slide-3
SLIDE 3

3

  • 1. Online storage/processing motivation
slide-4
SLIDE 4

4 4

DUNE meeting September 2016: Online processing and storage facility of 6x6x6

  • ffline
  • nline
slide-5
SLIDE 5

5 5

SPSC report, April 2016 Online storage/processing farm motivation : 6x6x6

slide-6
SLIDE 6

6

  • 2. Implementation on the 3x1x1
slide-7
SLIDE 7

7 7

3x1x1

A description of the hardware configuration of the farm has been provided during the 3x1x1 biweekly meeting of September 22nd https://indico.fnal.gov/conferenceDisplay.py?confId=12944 The EOS system was selected months ago for the data storage system thanks to the extensive performance tests made by Denis Pugnere in order to grant high bandwidth concurrent write/read access (see TB slides) and it has been implemented also in the 3x1x1 smaller farm. EOS is a network distributed file system using a meta-data server (https://indico.fnal.gov/conferenceDisplay.py?confId=12347)

slide-8
SLIDE 8

8

farm rack proximity rack (event builder)

https://indico.fnal.gov/conferenceDisplay.py?confId=12944

slide-9
SLIDE 9

 Some main architectural characteristics of the farm :

  • Event builder machine storage space: 48 TB
  • Online storage/processing farm storage size: 192 TB
  • Filesystem of the online storage system: EOS (high performance/bandwidth distributed

filesystem requiring a metadata server machine)

  • Protocol to copy data from Event Builder to the online storage system : XRootD
  • Resources manager for the batch system: TORQUE (installed on a dedicated machine)
  • Batch workers: 7 CPU units  112 processors, possibility of having up to 112 jobs running

simultaneously (jobs sequential monocore) The farm has been setup by Denis Pugnere (IPNL) and Thierry Viant (ETHZ).  Reconstruction software installed:

  • the latest version of WA105Soft (see Slavic presentation) and related libraries (root 5.34.23

XRootD 4.0.4, same versions installed at CCIN2P3 and on lxplus)

  • The farm is foreseen for fast reconstruction of the raw data and purity and gain online

measurement  only the code related to the fast reconstruction has been installed (the code needed for generation of Monte Carlo events is not available)

  • This code is available on the svn server (https://svn.in2p3.fr/wa105/311onlinefarm/)

9

slide-10
SLIDE 10

10

 Accounts on the online farm:

  • shift  used by people on shift, to run the DAQ, the event display and monitor results

see for instance the shifter DAQ doc.: http://lbnodemo.ethz.ch:8080/Plone/wa105/daq/daq-shifters-instructions-for-3x1x1-running/view

  • prod  to maintain the automatic data processing machinery: scripts for file transfers,

batch processing, copy to EOS and CASTOR, system monitoring

  • evtbd  DAQ account for the event builder software maintenance

The working environment for the 3 accounts is automatically setup at login

slide-11
SLIDE 11

11

Data flow

 Binary files are written by the DAQ in the storage server of the proximity rack: each file is composed by 335 events  1GB/file (optimal file size for storage systems) not compressed  each run can be composed by several files (this number is not fixed but depends on the duration of the run). The filename is runid-seqid:

1-0.dat

1-1.dat 2-0.dat 2-1.dat 2-2.dat

3 possible filetypes: .dat for rawdata, .ped.cal for pedestal data , .pul.cal for pulser data

The automatic online data processing includes these 3 steps (not in strict time order):

1) As soon as a data file is produced, it is copied to the EOS storage area of the farm, Depending on the filetype, a different processing chain is followed. In case of rawdata a script to run reconstruction is automatically generated and submitted to the batch system 2) Results from reconstruction (root files) are also stored in the storage area and analyzed to evaluate purity and gain, to monitor the behavior of the detector in time (online analysis) 3) The binary data files are also copied to the CERN EOS and CASTOR, where they are available to the users for offline analysis. Analysis results are stored on central EOS as well

slide-12
SLIDE 12

12

It is important to stress that this small scale farm is a test bench.

The operating experiences gained

  • 1. during the data taking (also if the data flow is very small)
  • 2. by performing mock data challenge with simulated and real data

are fundamental for the validation of the design of the final high rate system for the 6x6x6.

slide-13
SLIDE 13

13

automatic online data processing scheme

2b) analysis

Benchmark Purity analysis Gain analysis

Files are written by the DAQ

3 filetypes: dat  raw data

ped.cal  pedestal pul.cal  pulser They are immediately copied on local EOS the transfer to central EOS and CASTOR is scheduled Pedestal files and raw data files are processed Pedestal : an ascii file with pedestal value is produced (required by event display and reconstruction) raw data : a script to run reconstruction is automatically generated and submitted to the batch system The output root file (reconstructed data) is scheduled for transfer to central (EOS only) are run on reconstructed data and results are used to monitor the behavior of the detector in time (online analysis) . 1) copy to local EOS

2a) data processing 3) copy to CERN

shifter

slide-14
SLIDE 14

14

1) copy to local EOS

2a) data processing 3) copy to CERN 2b) analysis

  • Each of these steps is handled by processes from different directories of the production

account

  • To keep all the steps synchronized among them, a set of “bridge” directories has been put in
  • place. These directories are filled with the information on files to be treated by different

processing steps

  • Every processing step reads the bridge directory written by the previous step, and writes in its
  • wn one.
  • This mechanism allows to propagate the information on the files to be treated with a minimal

impact on the system

  • Every operation is recorded in a dedicated log file: this allows to monitor the processing
slide-15
SLIDE 15

15

1) copy to local EOS

  • As soon as a new data file is written by the DAQ , it is copied to the local EOS storage area of

the farm. To verify data integrity and validate the transfer the checksum value is verified

  • The detection of the completion of this new file is based on inotify, a Linux kernel feature

that monitors file systems and immediately alerts an attentive application to relevant events. It is used within a bash script, running in background. This mechanism avoids to scan the storage area every n seconds to look for new file The files are scheduled for transfer to EOS and CASTOR raw data file and pedestal files are scheduled for processing

slide-16
SLIDE 16

16

2a) data processing

In order to handle the processing the manager script processing.sh is periodically executed from the crontab:

  • It looks for entries to be processed in the bridge directory filled by previous step

2 possibilities: 1) If a pedestal file is detected, it launches its processing in interactive mode using caliana.exe. An ascii and a root file are produced, and their copy to EOS is scheduled 2) If a raw data file is detected, it creates a processing script and submits it to the batch system where the load is automatically balanced among workers The output root files are stored in local EOS, scheduled for transfer to central CERN EOS to cern

slide-17
SLIDE 17

17

In order to handle the analysis, the manager script checkforanalysis.sh is periodically executed from the crontab:

  • It looks for entries to be processed in the bridge directory filled by previous step

(reconstruction)

  • It creates a processing script and submits it to the batch system

2b) analysis

to cern

slide-18
SLIDE 18

18

 The analysis is run in 3 steps: 1) Production of benchmarking histograms 2) Purity evaluation 3) Gain evaluation  Since a run can be composed by several sequences: it is checked if results from previous sequences are available, if it is the case the analysis is also run on the full statistics for that run.  Analysis results are then scheduled for transfer to central EOS  Results from benchmark are the input of the monitoring task (the code for the monitoring task has been developed by Slavic, and it has been modified to be integrated into the farm working environment and processing scheme)  An example is shown in the next slides: the monitoring task is run on the farm, from the shift account, to monitor results of run 5001.

slide-19
SLIDE 19

19

Example (from the shift account)

number of hits reconstructed for each strip reconstructed charge per strip (from hit reconstructions)

View 0 View 1

run 5001

slide-20
SLIDE 20

20

3) copy to CERN

  • Files are copy by the service account wa105daq by using xrdcp
  • authentification: kerberos, with automatic renewal of tokens
  • To verify data integrity and validate the transfer the checksum value is tested at the final

destination of the file

  • In case of failure, the transfer is rescheduled, with a maximum of 3 attempts.
  • All files written by the DAQ are copied to EOS and CASTOR, output from reconstruction and

analysis are copied automatically only to EOS: files on CASTOR should have a minimal size of 1 GB for decent performances

  • The CERN EOS space allocated to WA105 has been “organized” in order to setup a

directory where data pushed from the online farm are stored . This space is accessible (read mode) by WA105 members (see later)

  • 10TB are available at the moment, more space can be added when needed
  • The creation of a “directory” for WA105 on CASTOR has also been requested and

implemented: (/castor/cern.ch/wa105)

  • Following a suggestion of Denis, the possibility to exploit the Third Party Copy protocol (data

flows from server to server  direct dataflow without intermediate machines) is under investigation

slide-21
SLIDE 21

21

  • The online processing has been tested during the different campaigns of noise measurement in

December and January: all the files written by the DAQ have been transferred to local EOS and then moved to the CERN computing center

  • In particular, to test the stability of the system, the scripts to handle the processing and the data

pushing to CERN EOS are continuously running from January 12th. Some numbers (from December 2nd) :

  • Up to ~ 400GB have been transferred from event builder to local EOS : transfer time ~4sec for

1GB file

  • These files have also been pushed to central EOS:

transfer time ~11 sec for 1GB file ~93MB/sec this value refers to the “basic” transfer, with only one stream

(December 2nd : DAQ commissioning, see Dario presentation here:

https://indico.fnal.gov/conferenceDisplay.py?confId=13680 )

Conclusions on the implementation on the 3x1x1

slide-22
SLIDE 22

22

  • Every operation (file copy, script creation, batch submission……) is recorded in ad-hoc log

files, needed to make performances studies.

  • An incremental copy of the logs is also automatically created
  • For redundancy these logs are copied also to CERN (from crontab).
  • Some background process have been setup ad hoc to monitor the smoothness of the data

flow

  • Two mock data challenges have already been performed using simulated compressed and

uncompressed data. In both cases the simulated trigger rate was of ~ 30Hz: the batch processing and files transfer ended smoothly. New mock data challenges will be repeated using real data

slide-23
SLIDE 23

23

  • 3. Data Availability at CERN
slide-24
SLIDE 24

24

  • A complete working environment has been set up CERN in July 2016

https://indico.fnal.gov/conferenceDisplay.py?confId=12617

  • The code installed and running on the farm is available here:

/afs/cern.ch/exp/wa105/Public/WA105Soft

  • All the data (raw data, reconstructed data, logs….) from the farm are accessible

as well here: /eos/experiment/wa105/data/311

slide-25
SLIDE 25

25

results from benchmarking, purity and gain analysis (slide 17) root file from raw data reconstruction (slide 16) dat pul.cal ped.cal (slide 11)

Summary information for every run: they are automatically generated on the farm, and then copied here with rsync. At the moment this is done once a day, but of course this frequency can be optimized

pedestal files from caliana.exe (slide 15)

slide-26
SLIDE 26

26

Summary information: some examples

pedestal

slide-27
SLIDE 27

27

Summary information: some examples

(fake) raw data 

slide-28
SLIDE 28

28

slide-29
SLIDE 29

29

These information are also stored in a database (see SB meeting 01/02/2107)

Thierry has setup this database, and WEB interface to access it:

https://wa105runs.web.cern.ch/wa105runs/ All runs are listed:

slide-30
SLIDE 30

30

Clicking here: Run details are shown These fields are filled during the data processing  it is possible to follow the data flow Not yet filled

slide-31
SLIDE 31

31

The shifter can add some comments on the run An click here to validate  The design of this database has just started , we are still working on it, so some changes can be foreseen in the following days To access the WEB interface: 1) you are asked to enter your CERN password 2) you have to be member of wa105-comp group

slide-32
SLIDE 32

32

  • All files on /eos/experiment/wa105/data are accessible to every member of

wa105-comp group

  • The run database is also accessible by people belonging to wa105-comp group
  • The next 3 slides will provide:
  • 1. some instructions on how to be part of wa105-comp group at CERN
  • 2. the path to the personal working area on /eos and some links to related

documentation 3.links to past presentations (>1 YEAR!!) explaining how to use WA105Soft to run reconstruction, to access reconstructed data….

slide-33
SLIDE 33

33 33

HOW to add your account to wa105-comp group: 1) https://e-groups.cern.ch/e-groups/EgroupsSearchForm.do 2) Select wa105-comp 3) subscribe

slide-34
SLIDE 34

34

Then you ask that wa105-comp becomes your main primary computing group: https://resources.web.cern.ch/resources/Help/?kbid=067030

It is also possible to have some working space on /eos 

for the login id alogin The working directory is here: /eos/experiment/wa105/user/a/alogin

No backup is performed on eos: a removed file is a lost file!

castor tutorial: https://cern.service-now.com/service-portal/article.do?n=KB0001103 How to increase your AFS quota: https://resources.web.cern.ch/resources/Help/?kbid=067040

slide-35
SLIDE 35

35

Presentations given at CM march2016: they are a summary of activities up to march2016. @CCIN2P3 /sps/hep/lbno/indico/EP_GM_08032016_SW.pdf /sps/hep/lbno/indico/qsim_wa105gm_08032016.pdf Cosmic ray reconstruction in 3x1x1, benchmarking, purity and gain analysis: https://indico.fnal.gov/conferenceDisplay.py?confId=12481 Software organization at CERN https://indico.fnal.gov/conferenceDisplay.py?confId=12617 Qscan interface for the 3x1x1 raw data https://indico.fnal.gov/conferenceDisplay.py?confId=13081 Study of cosmic ray selection for 6x6x6 online monitoring https://indico.fnal.gov/conferenceDisplay.py?confId=13872 Muon backgrounds in 6x6x6 https://indico.fnal.gov/conferenceDisplay.py?confId=13983

slide-36
SLIDE 36

36

  • 4. On going activities for the 6x6x6 online processing and

storage farm

slide-37
SLIDE 37

37 37 Marzio Nessi, DUNE GM, CERN, 01/23/2017

slide-38
SLIDE 38

38 38

slide-39
SLIDE 39

39

  • Starting for February 2017, the IT division is scheduling biweekly meeting with

representatives of protoDUNE DP and SP to help reviewing and following up the requests from the 2 protoDUNEs on online/offline computing and infrastructure

  • We are now discussing in details the network architecture in EHN1: for instance we

discussed the switches configuration, how to connect to the link to the CERN IT division common router, how many and which kind of switches are needed, specifications of the storage servers procurement of computing material

  • The discussion on database space needs also started
  • These meeting are very helpful, since IT experts for different items are available for
  • discussions. It is important to stress that the online computing part is only under the

responsibility of the experiment, nevertheless IT staff is available for suggestions/help

  • People from DP attending: Denis Pugnere, Thierry Viant , Dario and myself.
  • For what concerns the protoDUNEs data offline storage and processing, starting from

August 2017 onwards the protoDUNE community (SP+DP) will have access to 3 PB of EOS disks space, 6PB of Castor tape space and 1500 cores inside Condor. The

  • rganization of this space is under definition. The funding has been provided byt the

Neutrino Platform

  • All information related to the final setup of the farm for the 6x6x6 will be circulated in the

Science Board and Technical Board meetings.

slide-40
SLIDE 40

40

  • 5. General information concerning 6x6x6 offline processing
slide-41
SLIDE 41

41

A new group has also recently setup: the Fermilab/CERN/DUNE Interface Group (this group was already existing in Spring 2016, then its activity « stopped » for sometimes and now it has been resumed) The reflection on offline processing has also started in this interface committee (Andrew Norman-chair, Stuard Fuess, Berndt Panzer, Elisabetta Pennacchio, Ruth Pordes) Once protoDUNE data will be on CERN EOS they will copied at FERMILAB by exploiting the already existing link. Fermilab will also provide data analysis resources for both ProtoDUNE detectors and these data will be made available to the entire DUNE collaboration. The bulk of the data processing will be done at CERN and FERMILAB, but of course there is room for other centers. 2 questions from the interface committee to the WA105/protoDUNE DP collaboration: 1) Is your institution interested/available in providing computing resources to the data processing? As, for example, the CCIN2P3 is doing already for WA105 If yes, please contact Elisabetta, to start the discussion in the interface committee. 2) You may be interested in asking an account to FERMILAB: please let me (Elisabetta) know,

  • r, in case already asked and you had troubles, please let me know as well.
slide-42
SLIDE 42

42

Conclusions

  • All the steps for online data processing have been set up and commissioned
  • The automatic file transfer to local EOS, and the data pushing to CERN (EOS and

CASTOR) are running smoothly from January 12th.

  • A complete working environment has been setup at CERN (lxplus/EOS) as well)
  • The operating experiences gained during the data taking and by performing new mock data

challenge with real data are fundamental for the validation of the design of the final high rate system for the 6x6x6.

  • The collaboration with IT people is well in place

Thank you to Denis and Thierry for all the discussions and the collaboration

Thank you for your attention