First National Workshop of the National Knowledge Network Indian - - PowerPoint PPT Presentation

first national workshop of the national knowledge network
SMART_READER_LITE
LIVE PREVIEW

First National Workshop of the National Knowledge Network Indian - - PowerPoint PPT Presentation

B.S. JAGADEESH, COMPUTER DIVISION, BARC, TROMBAY, MUMBAI 400 085 First National Workshop of the National Knowledge Network Indian Institute of Technology, Powai 31/October/2012 Bunches, each containing 100 billion protons, cross 40


slide-1
SLIDE 1

B.S. JAGADEESH, COMPUTER DIVISION, BARC, TROMBAY, MUMBAI – 400 085

First National Workshop of the National Knowledge Network Indian Institute of Technology, Powai 31/October/2012

slide-2
SLIDE 2

Bunches, each containing 100 billion protons, cross 40 million times a second in the centre of each experiment 1 billion proton-proton interactions per second in ATLAS & CMS ! Large Numbers of collisions per event ~ 1000 tracks stream into the detector every 25 ns a large number of channels (~ 100 M ch)  ~ 1 MB/25ns i.e. 40 TB/s !

slide-3
SLIDE 3

LHC is a very large scientific instrument…

Lake Geneva

Large Hadron Collider

27 km circumference

CMS ATLAS LHCb ALICE

slide-4
SLIDE 4

James Casey, CERN, IT Department

slide-5
SLIDE 5

    

slide-6
SLIDE 6

 

most rudimentary

slide-7
SLIDE 7

Exper erime iment CPU (MSi2 i2k * year ar) Disk (TB) B) MSS (TB) Alic ice 59.2 .2 23903 17880 ATLAS AS 150 72,453 ,453 4839 398 CMS CMS 108.2 .2 34,40 ,403 46,80 ,800 LHCb Cb 17.88 47,49 11632

2,00,000 Computers !!!!

slide-8
SLIDE 8

Result

slide-9
SLIDE 9

Same (as) Yet different (from) Web (information) Allows collaboration too. (many resources of which information is one of them) Cluster/distributed computing (unifies resources) Unifies resources belonging to different administrative domains Virtualization (single resource) Allows virtualization of large

  • no. of resources like,

computation, Data, storage, information etc

slide-10
SLIDE 10

Ian.Bird@cern.ch 10

slide-11
SLIDE 11

11

 More than 140

computing centres

 12 large centres for

primary data management: CERN (Tier-0) and eleven Tier- 1s

 138 federations of smaller

Tier-2 centres

 India – BARC, TIFR,

VECC

 Relies on EGEE and

OSG Grids

slide-12
SLIDE 12

Tier2 Centre ~1 TIPS Online System Offline Processor Farm ~20 TIPS CERN Computer Centre FermiLab ~4 TIPS France Regional Centre Italy Regional Centre Germany Regional Centre Institute Institute Institute Institute ~0.25TIPS Physicist workstations ~622 Mbits/sec ~1 MBytes/sec

There is a “bunch crossing” every 25 nsecs. There are 100 “triggers” per second Each triggered event is ~1 MByte in size Physicists work on analysis “channels”. Each institute will have ~10 physicists working on one or more channels; data for these channels should be cached by the institute server

Physics data cache

~PBytes/sec

~622 Mbits/sec

  • r Air Freight (deprecated)

Tier2 Centre ~1 TIPS Tier2 Centre ~1 TIPS Tier2 Centre ~1 TIPS Caltech ~1 TIPS ~622 Mbits/sec

1 TIPS is approximately 25,000 SpecInt95 equivalents

slide-13
SLIDE 13
slide-14
SLIDE 14

Specific to WLCG– Gftp, LFC, GUID

MIDDLEWARE FUNCTIONALITIES ?

slide-15
SLIDE 15

DAE-CERN Joint Co-ordination Meeting, Nov 2, 2011, CERN

slide-16
SLIDE 16

   

slide-17
SLIDE 17

Node Configuration Manager NCM CompA CompB CompC ServiceA ServiceB RPMs / PKGs SW Package Manager SPMA Managed Nodes SW server(s)

HTTP

SW Repository

RPMs Install server

HTTP / PXE

System installer Install Manager

base OS XML configuration profiles

HTTP

CDB

SQL backend

SQL

CLI XML backend

SOAP

GUI SCRIPTS

ServiceC

slide-18
SLIDE 18
slide-19
SLIDE 19

 The key monitoring areas in GridView

include

 Service Availability Monitoring  File Transfer Monitoring  Job Monitoring

slide-20
SLIDE 20

SAM DB GRIDVIEW DB Service Nodes

SAM tests SAM Test Results SAM Framework Publishing Web Service R-GMA Archiver Module Web Service Archiver Module SAM XSQL Export Module

RBs SEs (gridftp)

WS Client RB Job Logs Gridftp Logs Gridftp Logs Fabric Monitoring System at Site (LEMON / Nagios) WS Client HTTP/XML Availability Metrics

GOCDB

GOCDB Sync Module Data Analysis & Summarization Module

Visualization Module

Graphs & Reports

slide-21
SLIDE 21

 Displays periodic Graphs and Reports for

 Detailed SAM test results for tests run for services at

a particular site

 Hourly, Daily, Weekly and Monthly basis  Full traceability from aggregate Availability to detailed

SAM test results

 Provision for saving user preferences based on

certificates

 Refer http://gridview.cern.ch/GRIDVIEW/

slide-22
SLIDE 22
slide-23
SLIDE 23

 Gridview computes job statistics based on RB

job logs

 Displays periodic Graphs and Reports for

 Job Status (Total Number of Jobs in various States)  Job Success Rate  Job Resource Utilization (Elapsed time,CPU, Memory)  Average Job Turnaround time (RB Waiting, Site

Waiting, Execution Time)

 Site, VO and RB-wise distribution  Hourly, Daily, Weekly and Monthly reports

slide-24
SLIDE 24

 Displays periodic Graphs and Reports for

 Overall Summary

○ sites with high/low job execution rate ○ sites with high/low job success rate ○ VOs running more/less jobs etc

 Possible to view job statistics for any user selected

combination of VO, Site and RB

 Refer http://gridview.cern.ch/GRIDVIEW/

slide-25
SLIDE 25

Please visit: http://gridview.cern.ch/GRIDVIEW/

Most recent snap shot

slide-26
SLIDE 26

 Fully web based system providing

 Tracking : Tracking reported bugs, defects,

feature requests, etc.

 Assignment : Automatic routing and

notification to support staff to get issues resolved

 Communication : Capturing discussion and

sharing knowledge

 Enforcement : Automatic reminders

according to severity of the issues

 Accountability : History and Logs

slide-27
SLIDE 27
slide-28
SLIDE 28

NKN is a state-of-the-art multi-gigabit pan-India network: www.nkn.in

►9th April 2009:

President of India Inaugurated the NKN Project.

► 16 PoP ► 26 Backbone Links ► 57 Edge Links ► 100 Crores allocated

in 2008 budget

► Completed in ¾ year

Connect R & D, educational, health, agri, labs institutes etc.. NIC working as the Project Execution Agency Idea of setting NKN was finalized at the Office of PSA & NKC. 1500+ institutes in final phase GoI approved a budget of INR 5990 crores for NKN in March, 2010

►5th March,

2011: Launched the Logo & Website of NKN

► 27 PoP ► 76 Backbone Links ► 216 Edge Links

HAS ENABLED EVERYONE TO COME ON BOARD !!

slide-29
SLIDE 29

National Grid Computing CDAC, Pune WLCG Collaboration Common Users Group (CUG) Anunet (DAE units) BARC – IGCAR

NKN Router

NKN-Internet (Grenoble-France) NKN-General (National Collaborations)

Intranet segment of BARC Internet segment of BARC

Logical Communication Domains Through NKN

slide-30
SLIDE 30

Category Examples Characteristics Distributed Supercomputing Ab-initio Molecular Dyn Large Cpu/ memory reqd High Throughput Cryptography Harness Idle cycles On Demand Medical instruments Cost effectiveness Data Intensive CERN LHC Info from Large Data sets Collaborative Data Exploration Support communication

slide-31
SLIDE 31

KEEPING PROVENANCE INFORMATION TO MAKE DATA DISCERNABLE TO NEXT GENERATION MEETING PROVISIONING CHALLENGES ( CLOUDMAN PROJECT) COMPLETE SWICTH OVER TO “CLOUDS” ? (SECURITY OF DATA?)

slide-32
SLIDE 32