cesnet storage activities overview
play

CESNET storage activities overview Ji Hork (jiri.horky@cesnet.cz) - PowerPoint PPT Presentation

12 th TF-Storage Meeting - Berlin CESNET storage activities overview Ji Hork (jiri.horky@cesnet.cz) 6th of March, 2013 History Relatively new activity to CESNET started in autumn 2010 with 2FTEs preparation phase continued until


  1. 12 th TF-Storage Meeting - Berlin CESNET storage activities overview Jiří Horký (jiri.horky@cesnet.cz) 6th of March, 2013

  2. History • Relatively new activity to CESNET • started in autumn 2010 with 2FTEs • preparation phase continued until the end of 2011 with no HW  no services • Some childhood problems appeared afterwards • the department restructured in April 2012 • 6.5 FTEs in total right now 6/3/2013 CESNET storage activities overview 2

  3. Financing and scale EU structural funds + projects of Ministry of Education, Youth and Sports • Research and Development for Innovations OP (“R&DI OP”) - project “ eIGeR ” • 100 mil. CZK ~ 4 mil. EUR (infrastructure only) • May 2011 – October 2013 (end of pilot phase) • Project “Large infrastructure” • 2011 – 2015, operational costs • commitment to sustain operation at least to 2018 6/3/2013 CESNET storage activities overview 3

  4. Data storage facility • Three geographically separated storage locations (Pilsen, PardubiceJihlava, Brno) to support research and science community • Large research projects • Public universities • Academy of Sciences • Public Libraries • Digitalization of books • ... • Total capacity 16.6 PB + another tender in preparation 6/3/2013 CESNET storage activities overview 4

  5. Distributed storage locations 6/3/2013 CESNET storage activities overview 5

  6. Technical assumptions • Assumptions: • Emphasis on economical aspects + transparent behavior to users -> HSM system • Backup and archival demands foreseen -> requirement for a tape library • Difference in usage patterns will be covered by migration policies, e.g.: • archival – move everything from disk to tape almost immediately • input files for batch jobs – migrate only files touched > one month ago 6/3/2013 CESNET storage activities overview 6

  7. The Pilsen storage site • Hosted by University of Western Bohemia, delivered in the end of 2011, in pilot operation since May 2012 • Multi-tier, managed by DMF (Data Migration Facility), SLES OS, CXFS file system (SGI) • Disk systems in two IS4600 arrays: • Tier1: 50 TB FC drives • Tier2: 450 TB SATA drives • Tier3: Spectra T-Finity tape library with dual robotics • 2200 slots with LTO5 (3.3PB in total), 2300 free slots • 8Gbit redundant FC SAN, 10 Gbit Ethernet • 2x10 Gbit connection to CESNET2 network • 6 frontends managed by Pacemaker in HA cluster • 2 HSM servers in HA mode, one system for administration 6/3/2013 CESNET storage activities overview 7

  8. The Pilsen storage site 6/3/2013 CESNET storage activities overview 8

  9. The Jihlava storage site • Tender finished in the beginning of 2013 • Hosted by Vyso čina Region Authority • Multi-tier, GPFS filesystem, Tivoli Storage Manager (IBM) • Disk systems in two DSC3700 arrays (+JBODs): • Tier1: 800 TB SATA drives • Tier2: • 2.5 PB MAID (Promise) • 3.8 PB Tape – IBM TS3500 with TS1140 drives, dual robotics • 16Gbit redundant FC SAN, 10 Gbit Ethernet • 2x10 Gbit connection to CESNET2 network • 5 frontends, 2 HSM servers in HA mode 6/3/2013 CESNET storage activities overview 9

  10. The Jihlava storage site Only being delivered during this week! 6/3/2013 CESNET storage activities overview 10

  11. The Brno storage site • Tender finished in the very beggining of 2013 • Hosted by Brno University of Technology • Multi-tier, GPFS filesystem, Tivoli Storage Manager (IBM) • Disk systems in two DSC3700 arrays (+JBODs): • Tier1: 400 TB SATA drives • Tier2: • 1.8 PB MAID (Proware) • 3.5 PB Tape – IBM TS3500 with TS1140 tape drives, dual robotics • 16Gbit redundant FC SAN, 10 Gbit Ethernet • 2x10 Gbit connection to CESNET2 network • 5 frontends, 2 HSM servers in HA mode 6/3/2013 CESNET storage activities overview 11

  12. The Brno storage site Server room still under construction! 6/3/2013 CESNET storage activities overview 12

  13. Use cases • Long term storage with high redundancy (geographical replicas) – but no data curing • Institutional backups and/or archive – individual backups supported as well • Input/output data for batch jobs • Exchange point of scientific data among collaborations using standard file protocols • not suitable for – on-line data processing – critical services 6/3/2013 CESNET storage activities overview 13

  14. Access to storage services • File access – access to the same namespace • our “core bussiness” – taking advantage of file- level migration to lower tiers – NFSv4 with strong Kerberos authentication – SCP, SFTP (sshfs) – FTPS – rsync 6/3/2013 CESNET storage activities overview 14

  15. Access to storage services • Filesender • our first storage service – but not heavily advertised yet • up to 400 GB files (images of systems) • still stable, can resume!  • upload speed limit of ~100Mbit  • command line API? 6/3/2013 CESNET storage activities overview 15

  16. Access to storage services • Grid Storage Element – dCache for ATLAS (LHC at CERN) and AUGER experiments – coupled with HSM to enable efficient storage of not heavily used data – 30TB on disks + 2x100TB on tapes – data are actually read here, not only written  6/3/2013 CESNET storage activities overview 16

  17. Access to storage services • Block access – iSCSI directly from arrays – no migration possible using HSM – speed (latency) issues – only in special and well justified cases (7TB so far) – but very stable service  6/3/2013 CESNET storage activities overview 17

  18. User management • One user identity in the whole CESNET e-Infrastructure (data storage group, NGI_CZ, video conferencing services...) • Divide and conquer in practice: – users arranged in groups with an admin from the group – resource owners negotiate on service quantity and quality (quotas, migration policies...) with the group admin – the group admin decides who can join the group and what quality of service each particular user receives • Managed by Perun v3 system developed in collaboration with other projects 6/3/2013 CESNET storage activities overview 18

  19. User management • We require users to be well verified – “somebody we trust has seen user’s ID” – managed by eduID.cz federation – then we create a Kerberos principal & password which forms his e-Infrastructure identity – decoupled from federation after registration • Grid Storage Element and iSCSI are special 6/3/2013 CESNET storage activities overview 19

  20. Open questions • Inter site replication – FTP backend for each HSM in the beginning • each site will migrate to another site as to lower tier • Long term goal: One namespace across all three sites accessible by standard file protocols – suggestions? 6/3/2013 CESNET storage activities overview 20

  21. Conclusion • 3.8 PB in pilot operation • mainly backups and archive data • 16.6 PB available soon • Broad range of user demands • broad range of technical challenges (aka problems :-) • surprisingly many bugs & patches of commonly used tools/apps! 6/3/2013 CESNET storage activities overview 21

  22. Thank you for your attention. Questions? Jiří Horký (jiri.horky@cesnet.cz) 6/3/2013 CESNET storage activities overview 22

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend