the Belle II Experiment David Asner Pacific Northwest National - - PowerPoint PPT Presentation

the belle ii experiment
SMART_READER_LITE
LIVE PREVIEW

the Belle II Experiment David Asner Pacific Northwest National - - PowerPoint PPT Presentation

Computing and Networking for the Belle II Experiment David Asner Pacific Northwest National Laboratory July 31, 2013 Belle Computing deployed at PNNL in response to earthquark & tsunami Following the earthquake & tsunami of March


slide-1
SLIDE 1

Computing and Networking for the Belle II Experiment

David Asner

Pacific Northwest National Laboratory July 31, 2013

slide-2
SLIDE 2

Belle Computing deployed at PNNL in response to earthquark & tsunami

Following the earthquake & tsunami

  • f March 12, 2011, electrical power

available to KEK was dramatically reduced, and most of Belle computing – centralized at KEK – was offline. DOE-supported Belle computing at PNNL came online July 2011. Belle Analysis Software Framework installed on 1500 core cluster with 1 pb disk/1 pb type ! BCOMP KEK

Data + 1x "(4S) +1x "(5S) MC sample populated via scp from KEK 3x MC "(4S) + 2x MC "(5S) + 150x Rare MC generated at PNNL 90+ Belle users with PNNL accounts

2 2 July 31, 2013 Belle II Computing

slide-3
SLIDE 3

Significant strength and depth in High Performance Computing and Computational Science

3

  • Deep in data sciences; especially strong in

visual analytics and real-time HPC

  • Development of massively parallel

simulation codes

  • Critical mass in computer science and

applied math – especially strong in performance, power and reliability

  • Institutional hardware includes 160 TF

Olympus HPC cluster and archive storage – available to ALL staff and their collaborators

  • Programs may buy nodes or storage, “own”

those resources = priority usage

  • Redundant access to ESnet
  • LEED Gold-certified facility using

groundwater heating and cooling 624 ! 800 compute nodes 32 cores per node (20k cores) 2 GB of memory per core 4/3 PB of storage on disk/tape

July 31, 2013 Belle II Computing

slide-4
SLIDE 4

Expected data rates from the Belle II experiment are high. Expected to exceed the sum of the CERN Large Hadron Collider experiments

July 31, 2013 4 Belle II Computing

!"#$%&'$()* !+$()*,&-$*./01)$23* 45)$*.6-3* 45)$*.701)$2823* !"#$%&'()%*+),'&"-%.-&%/)00)%11%2345! /)00)%11* 677* 89777* :9;77* <=>%?2@%AB77CD5! 3<1=E%A!1D% :B9C77% :77% :9BC7% 3<1=E%AFFD% :9777% :77% :77% 3?<3G% :9877% B77% 6B7% =HG% :9C77% :C7% BBC% <!=I% BC% B9777% C7%

slide-5
SLIDE 5

Projection of Data Volume/Rate at SuperKEKB

July 31, 2013 Belle II Computing 5

Goal of Belle II/SuperKEKB Will accumulate 50 ab-1 corresponding to ~130 PB raw data volume in 2022!

9 months/year! 20 days/month!

Commissioning starts in 2015.! Shutdown! for upgrade!

Integrated luminosity ! (ab-1)! Peak luminosity ! (cm-2s-1)!

Year!

Will reach ~30 PB/year ! data rate in 2020!

slide-6
SLIDE 6

Evolution from Belle to Belle II Computing

Belle Experiment ran from 1999-2010 with computing centralized at KEK Belle II: 50 times more data requiring ~50 times more computing resources, distributed collaboration Belle II Collaboration adopted a distributed computing model based on the grid More complicated, requires more effort to set up and maintain Allows Belle II members to contribute via computing facilities in their own country It provides redundancy The distributed infrastructure already exists Use existing grid sites and services (gLite, EMI, OSG, DIRAC)

July 31, 2013 6 Belle II Computing

J&-K(%.&-L%)MF)&"),+)%-.%<!=% )MF)&"L),(*%',N%($)"&%O)00%% )*('I0"*$)N%',N%L'(P&)%*-0PQ-,*%

n e na ch)

slide-7
SLIDE 7

Grid Sites in Belle II Collaborating Countries

9:;()%1* ,&)$2* 0$<<$*=>* 9:''$()* 3P*(&'0"'% ?")&%BR6% GPFF-&()N% =0-PN%*S*()L%F0',,)N% 3P*(&"'% ?")&%B% =$",'% ?")&%B% 21@3=%G)&T)&% =U)+$% ?")&%B% GPFF-&()N% >)&L',S% ?")&%:RB% GPFF-&()N% 1,N"'% ?")&%B% V)O%N'('%+),()&%F0',,)N% W'F',% XEX% GPFF-&()N% X-&)'% ?")&%B% GPFF-&()N% J-0',N% ?")&%BR6% GPFF-&()N% =0-PN%*S*()L%N)T)0-F)N% @P**"'% ?")&%B% G0-T),"'% ?")&%B% GPFF-&()N% ?'"O',% ?")&%:RB% YG3% ZG>% GPFF-&()N% JVV<%*"()%O"($%21@3=%*)&T)&%

July 31, 2013 7 Belle II Computing

slide-8
SLIDE 8

Computing Tasks

July 31, 2013 8 Belle II Computing

Raw data processing

Store (tape) and process at KEK Replication to just one remote site (PNNL)

Monte Carlo Production

6 times the real data size Produced in managed way, (almost) no input data needed Well suited for a distributed environment, including cloud

Physics Analysis

Random, uncoordinated access

Store input data on disk

  • Ntuple analysis on local resources for fast turn-around
slide-9
SLIDE 9

The Belle II experiment will use a distributed computing model

July 31, 2013 9 Belle II Computing

V(PF0)% 3,'0S*"*% H-,()%='&0-% ',N%V(PF0)%% F&-NP+Q-,% H-,()%='&0-%% F&-NP+Q-,% A-FQ-,'0D% @'O%N'('%*(-&'#)% ',N%F&-+)**",#% @'O%N'('%*(-&'#)% ',N%F&-+)**",#%

slide-10
SLIDE 10

PNNL-KEK-KIT-OSG January 2012 Workshop

Joint computing workshop held at PNNL January 25-26, 2012 Objectives:

Establish and reinforce relationships between PNNL, KEK Show PNNL computational capabilities Share KEK and PNNL plans for computational aspects of Belle II collaboration Discuss formal agreements for collaboration

Outcomes:

PNNL and KEK working jointly on formal Memorandum of Understanding PNNL will establish pilot Grid site supporting Belle II PNNL investigating possibility of acting as backup site for all Belle experimental data

200 additional petabytes by 2020 Japan – US Bandwidth needs (~2 GB/sec) may tax existing infrastructure DOE HEP will be target funding agency

Attendees: KEK: Mitsuaki Nozaki, Koichi Murakami, Go Iwai, Hiroyuki Matsunaga, Takanori Hara, Takashi Sasaki. KIT: Thomas Kuhr. PNNL: Jerry Johnson, Mike Thompson, David Asner, Kevin Regimbal, Dave Cowley, Shaun O’Leary, Jim Schroeder, Tim Carlson, Jason Jensen OSG: Chander Sehgal, Gabriele, Garzoglio. US Belle: Romulus Godang, Leo Piilonen

July 31, 2013 Belle II Computing 10

slide-11
SLIDE 11

“Pacific Network & Computing Requirements” Workshop hosted by PNNL - Oct 17-18, 2012

Organized by PNNL-KEK-ESNET

The purpose of this workshop was to begin preparation for addressing the wide-area networking requirements for science in general and of the Belle II experiment in

  • particular. Objectives include: developing

a common understanding of Belle II science objectives, discovery workflows, cyber-infrastructure requirements, and data models; discussing challenges posed by Belle II data rates; and developing a concrete plan for establishing a Belle II grid site at PNNL and assuring that adequate capabilities for data transport (including monitoring and measurement infrastructure) are in place and thoroughly tested before they are needed by the experiment.

July 31, 2013

11

!"##"$%%&'()"*+,"-.&/".01*2&3"45+*","-.6

!"##"$%%&'()"*+,"-.&/0##120*1.+0- '-"*34&56+"-6"7&8".90*:

!"#$%&'()*+),-(./).

$[F5RROOO\)*\,)(R'**)(*RFPI*]F&)*-*R/)00)^11^EMF)&"L),(^ V)(O-&_^@)`P"&)L),(*^a-&_*$-F^T:;^K,'0\FN.%

%

Belle II Computing

slide-12
SLIDE 12

Belle II Computing Requirements – CY15-20

Note: Integrated data volume doubles CY20 to CY22 KEK Site

July 31, 2013 12 Belle II Computing

?&2@5<*A$5%* BCDE* BCDF* BCDG* BCDH* BCDI* BCBC* ?'F)R2"*_%bJ/c% B\;R6\7% B\;R6\7% d\8Re\8% BfR:6% CBRBB% ffReB% =JY%b_!EJGF)+c% e7% e7% de% :8;% :;C% :d8% a3V%b>I"(R*c% B\C% e\7% d\f% :;\;% Be\6% BC%

PNNL Site

?'F)R2"*_%bJ/c% 7\7RB\7% 7\7RC\7% d\8R:B% BfR:f% CBRBB% ffRBf% =JY%b_!EJGF)+c% :7% :C% Cd\:% dC\;% f8\8% ;B\f% a3V%b>I"(R*c% B\C% e\7% ;\f% :C\;% :d\6% :d\e%

Regional Site

2"*_%bJ/c% 7% 7\6% 6\B% d\:% :C\B% B:\e% =JY%b_!EJGF)+c% 7\f% e\6% e:\e% fB% fB% f;% a3V%b>I"(R*c% 7\7e% 7\6% B\f% 6\8% 6\d% e\7%

slide-13
SLIDE 13

GV<<% JVV<% <3V<% GV<3% W<3/% JJJ<% /V<% 3HEG% <<V<% W>1% </V<% 3V<% G'0(%<'_)% >g2<% JY%J$S*"+*% %GYVV% %GE3?% %G?3@% %3</Y% <3Gh% %G2G=%

%<ZG3% =<Eh% :77%

:7% :7% :7% :7% :7%

:77%

:7% :7% :7%

:%

:77% :77% :77% :77% :77%

:7% :7%

:77% :77% :7%

:7% :7% :7% :7%

:77%

:77%

:7% :7% :7%

:77%

:7%

:77% :7% :77% :77% :77% :77% :77% :77% %E4=!%

"#$%&'()*!+,! $-.*!&#(&#,#-/'0$-'.!

:%

Esnet Network Infrastructure

July 31, 2013 Belle II Computing 13

slide-14
SLIDE 14

Summary

Belle II experiment at SuperKEKB will search for New Physics with O(50) times more data than Belle Huge data volume is a challenge for network & computing

Sustaining very high (1.8 GByte / second) data rates In Japan, Across the Pacific, In the US Handling (re)processing tasks in US (as well as Japan) Distributed computing system based on existing technologies and infrastructures Formation of grid sites federation in progress

3 logical layers (main centers, grid sites, local resources)

Distributed computing user interface: gbasf2

Workflow abstraction with projects and datasets Easy transition from offline software framework: basf2

Scientific program requires upgrade of software & computing system to complement the accelerator and detector upgrade projects

July 31, 2013 14 Belle II Computing

http://belle2.kek.jp