ideas for real time analysis for hl lhc using the cms daq
play

Ideas for Real-Time Analysis for HL-LHC using the CMS DAQ System - PowerPoint PPT Presentation

Ideas for Real-Time Analysis for HL-LHC using the CMS DAQ System Remigius K Mommsen Fermilab Disclaimer The idea of the L1 scou0ng originates from Emilio Meschi (CERN) This talk is based to a large extend on material presented by Hannes


  1. Ideas for Real-Time Analysis for HL-LHC 
 using the CMS DAQ System Remigius K Mommsen Fermilab

  2. Disclaimer The idea of the L1 scou0ng originates from Emilio Meschi (CERN) This talk is based to a large extend on material presented by Hannes Sakulin (CERN) 
 at CHEP 2019, Adelaide, Australia Any mistakes or misinterpreta0ons are mine 2

  3. All-new CMS for HL-LHC (2027 onwards) Barrel Calorimeters • new BE/FE electronics Muon Systems • ECAL: lower temperature • new DT/CSC BE/FE electronics • HCAL: partially new scintillator • GEM/RPC coverage in 1.5<| ƞ |<2.4 • Muon-tagging in 2.4<| ƞ |<3.0 Endcap Calorimeters • high granularity calorimeter • radiation tolerant scinitllator • 3D capability and timing MIP Timing Detector • 30-60 ps resolution L1 rate: 
 • coverage up to | ƞ | = 3.0 750 kHz HLT rate: 
 ~7.5 kHz Tracker Event size: 
 • radiation tolerant, high granularity, 7.5 MB low material budget • coverage up to | ƞ | = 3.8 300 TB/s 
 • track trigger at L1 @ 40 MHz 3

  4. CMS Trigger & DAQ — 2 Trigger Levels Only Phase 0 & 1 — 2008-24 Phase 2 — 2027- 40 MHz 40 MHz Digitizers LV1 Front end pipelines 100 kHz 750 kHz μ s 1.5 MB 
 7.5 MB 
 Event-builder nodes event size event size 0.15 TB/s 5.5 TB/s HLT sec 2 kHz 7.5 kHz Storage (pt5/tier 0) 4

  5. L1 Trigger for HL-LHC High resolu0on objects • Tracker track reconstruc0on in firmware • Vertex finding • Kalman filter muon 12 µs latency reconstruc0on • Displaced muons • High precision calorimetry • Par0cle flow reconstruc0on Topological algorithms including invariant/transverse mass cuts Machine learning algorithms Inter-bx algorithms 
 (limited to +/- 3 bx) 5

  6. What is Real-Time Analysis? Analyze events while the data is being taken • Par0al events with limited resolu0on • Full events with sub-op0mal calibra0ons • Much higher rate than possible with offline analysis • Stringent 0me constrains Store summary results for certain topologies at higher rate • E.g. low-mass di-jets, three-jet resonances, di-muons LHCb will does most analysis in “real-0me” • 2-step HLT selec0on • 2nd step is run aier calibra0ons have been done • Same physics quality as offline for most objects 6

  7. HLT Real-Lme Analysis Data Scou0ng at HLT used successfully in CMS since 2011 • Save HLT physics objects to disk Detectors • Perform offline analysis on these objects rather than on Digitizers offline reconstructed en00es LV1 Front end pipelines • No raw data is saved and no further reconstruc0on is performed for these events μ s • Typically 1-5 kHz of scou0ng data O(100 MB/s) Readout buffers Switching networks HLT Processor farms sec Tiny event 
 at higher rate 7

  8. L1 Trigger ScouLng Acquire L1 trigger data at full bunch crossing rate Detectors • No back pressure 40 MHz Level-1 
 Digitizers Trigger Scou0ng System • Drop data if system cannot 
 keep up with rate LV1 Front end pipelines μ s Analyze certain topologies at full rate • Real-0me analysis Readout buffers • Store 0ny event record Planned for HL-LHC Switching networks • Prototyping now • Tes0ng during run 3 HLT Processor farms sec 8

  9. Physics to Look at with L1 scouLng (non-exhausLve) Physics use case • Rare process • Difficult to select at Level-1 trigger - despite upgraded L1 trigger 
 (Available cuts give low efficiency at amributed rate budget) • Analysis is possible with resolu0on available at Level-1 • Scou0ng for new signal -> then point L1 trigger to it Several Physics channels iden0fied where L1 scou0ng could poten0ally make a difference 9

  10. Other uses for Level-1 Trigger scouLng Scou0ng provides invaluable diagnos0c and monitoring opportuni0es as well • BX-to-BX correla0ons available at all 0mes (cosmics, pre/post firing, etc.) • Real-0me heat maps to immediately spot problema0c channels • High-stat cross-check of algorithms (e.g. GT inputs/outputs) Per-bunch luminosity measurement using physics channels with high sta0s0cs Anomaly detec0on with deep-learning algorithms 10

  11. HL-LHC 40 MHz L1 ScouLng 
 Stageable Architecture

  12. 
 
 
 
 ScouLng system components Amached storage 
 long term Expect Xilinx Kintex Ultrascale+ based HW board to be commercially available Query- Infiniband HDR, 200 GbE op7cal based HPC Interconnect(s) Other 
 analysis 25 Gb/s GPU CPU Accel. from Input board • Zero suppression trigger distributed processing (MPI ?) • Pre-processing Input HW Feature Key-value 
 • Re-calibra0on 8x Features or 
 DB 
 store ? SoAware ZS 200 using ML DMA full events Short term medium Kintex Ultrascale+ Gbps storage 
 term PCIe 
 Kintex Ultrascale+ 2 min 
 NIC (mul7-bx possible) Same 
 Gen4 1-3 TB protocol 1 or 2 boards RAM ? 
 Distributed … NVRAM ? as in 
 (global) trigger I/O node … stream processing no back- 
 pressure 12

  13. Ingredients Trigger data captured directly from the Level-1 using spare outputs of the processing boards • Assuming same 16/25 Gbps serial op0cal links used for the Level-1 interconnects and using the same protocol Input hardware: PCIe boards with (modest) FPGA in 1U PC (I/O node) – (uGMT scou0ng uses KCU1500 [limited to 16 Gbps]) • Zero-suppression, local pre-processing (e.g. re-calibra0on using ML) in FPGA • DMA to host memory for short-term buffering (~2 min) • Baseline: eight op0cal inputs per board (PCIe Gen4 ~ 200Gbps over 16 lanes), one or two input boards per PC I/O nodes (CPU, GPU, other accelerators) use distributed algorithms to extract features while data are buffered in memory • 1-3 TB short-term buffer (e.g. NVRAM, could be cheaper with acceptable latency) • 200 Gbps low-latency interconnect (e.g. InfiniBand HDR or 200 GbE) • Interes0ng features and/or full “events” (mul0-bx possible) streamed over interconnect to global processing “farm” Distributed global stream processing and storage into “feature DB” • Organizes features in “searchable” data structures • Search-engine-like system op0mized for numerical data, medium term storage (e.g. key-value store) Analysis by query, analysis results to permanent storage 13

  14. L1 Trigger System 14

  15. 
 
 L1 Trigger System Scou0ng System I/O nodes Amached • Local processing storage 
 • Transient storage long term Trigger Query- Primi0ves Infiniband HDR, 200 GbE based HPC Interconnect(s) analysis Tracker Feature Calo DB 
 medium Muon term Distributed Global (global) stream Decision processing 14

  16. 
 
 L1 Trigger System Scou0ng System I/O nodes Amached • Local processing storage 
 • Transient storage long term Trigger Query- Primi0ves Infiniband HDR, 200 GbE based HPC Interconnect(s) analysis Tracker Feature Calo DB 
 medium Muon term Distributed Global Stage1: (global) 9 nodes 
 stream @ 200 Gbps Decision processing 14

  17. 
 
 L1 Trigger System Scou0ng System I/O nodes Amached • Local processing storage 
 • Transient storage long term Trigger Query- Primi0ves Infiniband HDR, 200 GbE based HPC Interconnect(s) analysis Tracker Feature Calo Stage 2: DB 
 add 28 nodes 
 medium Muon @ 200 Gbps term Distributed Global Stage1: (global) 9 nodes 
 stream @ 200 Gbps Decision processing 14

  18. 
 
 L1 Trigger System Scou0ng System I/O nodes Amached • Local processing storage 
 • Transient storage long term Trigger Query- Primi0ves Infiniband HDR, 200 GbE based HPC Interconnect(s) Stage 3: analysis Tracker add 98 nodes 
 @ 200 Gbps Feature Calo Stage 2: DB 
 add 28 nodes 
 medium Muon @ 200 Gbps term Distributed Global Stage1: (global) 9 nodes 
 stream @ 200 Gbps Decision processing 14

  19. 
 
 L1 Trigger System Scou0ng System I/O nodes Amached • Local processing storage 
 • Transient storage long term Stage 4: Trigger add 100's nodes 
 Query- Primi0ves @ 200 Gbps Infiniband HDR, 200 GbE based HPC Interconnect(s) Stage 3: analysis Tracker add 98 nodes 
 @ 200 Gbps Feature Calo Stage 2: DB 
 add 28 nodes 
 medium Muon @ 200 Gbps term Distributed Global Stage1: (global) 9 nodes 
 stream @ 200 Gbps Decision processing 14

  20. GMT scouLng prototype in Run 2

  21. Global Muon Trigger ScouLng in Run 2 When: Oct / Nov 2018 Types of runs: • 1 week of pp run • Large part of HI run Capture @ 40 MHz • Up to 8 final muon candidates • Up to 8 intermediate muon 
 candidates from barrel region • GMT adds bunch and orbit counters 
 40 MHz Scouting 
 Prototype System 16

  22. Global Muon Trigger (GMT) ScouLng Prototype 2x QSFP = 
 8x 10 Gbps op7cal PCIe Gen3x8 (2x) KCU 1500 
 Xilinx Kintex Ultrascale 115 10 Gb/s from GMT SoAware ZS 
 Infini 10 40 BZIP (1/2) DMA KCU1500 10/40 8x RAMdisk 
 (1/8) band Gbps Gbps RAM 
 RAID 
 Lustre Gbps mount firmware ZS 
 PCIe 
 disk 8 TB switch NIC NIC NIC (1/20) Gen3 8 GB/s max 800 MB/s max 100 MB/s max 50 MB/s Dell R720 Dell R720 1.1 TB/24 hour beam day 
 Controller PC 
 aier compression in pp @2E34 Firmware update & monitoring 17

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend