SOFTDRIVE.NL, SOFTDRIVE.NL, CVMFS FOR THE CVMFS FOR THE MASSES - - PowerPoint PPT Presentation

softdrive nl softdrive nl cvmfs for the cvmfs for the
SMART_READER_LITE
LIVE PREVIEW

SOFTDRIVE.NL, SOFTDRIVE.NL, CVMFS FOR THE CVMFS FOR THE MASSES - - PowerPoint PPT Presentation

SOFTDRIVE.NL, SOFTDRIVE.NL, CVMFS FOR THE CVMFS FOR THE MASSES MASSES DENNIS VAN DOK DENNIS VAN DOK Generic Components of the eScience Infrastructure Ecosystem 14th IEEE eScience Conference Amsterdam, Monday 2018-10-29 1 GRID


slide-1
SLIDE 1

SOFTDRIVE.NL, SOFTDRIVE.NL, CVMFS FOR THE CVMFS FOR THE MASSES MASSES

DENNIS VAN DOK DENNIS VAN DOK

Generic Components of the eScience Infrastructure Ecosystem — 14th IEEE eScience Conference Amsterdam, Monday 2018-10-29

1

slide-2
SLIDE 2

GRID COMPUTING A.K.A. PAAS GRID COMPUTING A.K.A. PAAS

Large scale common science infrastructure for high throughput batch computing. the only guaranteed environment is the base OS and some middleware no persistent local storage between jobs bring your own soware

2 . 1

slide-3
SLIDE 3

CHALLENGES OF SOFTWARE CHALLENGES OF SOFTWARE DISTRIBUTION DISTRIBUTION

Bringing soware with every job incurs much

  • verhead

Projects to develop common soware distributions have a slow upgrade cycle Negotiating a locally writable soware area for each site takes time, effort and coordination

2 . 2

slide-4
SLIDE 4

SOFTWARE DISTRIBUTION WITH SOFTWARE DISTRIBUTION WITH CVMFS CVMFS

CVMFS spun off the the CERN Virtual Machine content delivery based on http data is distributed as objects referenced by hashes read-only, so trivial to replicate massively transactionally consistent indices garbage collected

3 . 1

slide-5
SLIDE 5

ARCHITECTURE ARCHITECTURE

/cvmfs/atlas.cern.ch CVMFS fuse driver automount local cache disk read squid cache http Atlas stratum 1@CERN http Atlas stratum 1@RAL http Atlas stratumn0@CERN periodic sync over http Atlas stratum 0@CERN periodic sync over http

3 . 2

slide-6
SLIDE 6

DRAWBACKS DRAWBACKS

CVMFS is great for large organisations. But for small teams it can be a real challenge: set up and maintain a repository take care of a Stratum-0 server negotiate the replication at Stratum-1 sites negotiate with sites to include the repository in their CVMFS configuration I imagined dozens of small e-science groups knocking

  • n my door to get their repositories mounted.

3 . 3

slide-7
SLIDE 7

OUR SOLUTION: SOFTDRIVE.NL OUR SOLUTION: SOFTDRIVE.NL

Nikhef and SURFSara have jointly set up /cvmfs/sodrive.nl to offer a single CVMFS repository for all e-science users in the Netherlands.

4 . 1

slide-8
SLIDE 8

The system consists of a user interface system, where users can log on (with ssh) and upload their soware a Stratum-0 server which copies the user's files at regular intervals Stratum-1 at Nikhef and RAL mounted by default on all grid resources in the Netherlands

user softdrive.grid.sara.nl ssh stratum-0 rsync

ARCHITECTURE ARCHITECTURE

/cvmfs/softdrive.nl/z /cvmfs/softdrive.nl/c /cvmfs/softdrive.nl/p /cvmfs/softdrive.nl/s /cvmfs/softdrive.nl/t /cvmfs/softdrive.nl/a /cvmfs/softdrive.nl/r /cvmfs/softdrive.nl/f /cvmfs/softdrive.nl/k

4 . 2

slide-9
SLIDE 9

RULES RULES

User requests account at SURFSara Standard quota of 2GB (could be extended) Manage soware on sodrive.grid.sara.nl Copy soware to /cvmfs/softdrive.nl/$USER Run the publish command which touches the sodrive.modified file

4 . 3

slide-10
SLIDE 10

MECHANISM MECHANISM

Automated rsync from Stratum-0 server at Nikhef Two stage process:

  • 1. rsync the sodrive.modified files
  • 2. rsync those directories with updated

sodrive.modified files

4 . 4

slide-11
SLIDE 11

QUIRKS QUIRKS

Catalog size exploded when monitoring was put in

  • place. The monitoring triggered an update every five

minutes and thereby a completely new, full catalog of all files. This was ultimately understood and remedied by making subcatalogs per user.

4 . 5

slide-12
SLIDE 12

USER EXPERIENCE USER EXPERIENCE

To complement the technical implementation, the total user experience was taken care of by having proper documentation, monitoring and guidance.

5 . 1

slide-13
SLIDE 13

DOCUMENTATION DOCUMENTATION

The user documentation is right there when logging on to the system. The message of the day, printed for login shells, gives a summary of the workings of the system and how to publish data. More extensive documentation was . written and placed

  • n-line

5 . 2

slide-14
SLIDE 14

MONITORING MONITORING

End to end monitoring of the system is done by automatically triggering a change to the system every hour and measuring the time it takes for the data to reach a client machine. Alerts are raised if the delay reaches a certain threshold, prompting the technicians to inspect what went wrong.

5 . 3

slide-15
SLIDE 15

SUMMARY SUMMARY

The sodrive model has proven to be succesful; it is easy for users to maintain their own soware; the soware is lightweight and the maintainance burden

  • n the administrators is very light.

There is no plan at this point to add more bells and whistles to the system. Even as the PaaS infrastructure dwindles in favour of IaaS (infrastructure as a service), the CVMFS system could still be a viable component for delivering soware.

6 . 1

slide-16
SLIDE 16

SOME NUMBERS SOME NUMBERS

25 active users last 6 months 393k files, 178 GB

6 . 2

slide-17
SLIDE 17

INTERESTED? INTERESTED?

Some other national grid infrastructures offer something similar to sodrive, but I've not heard of anyone interested in cloning our setup. If you have plans to provide CVMFS to your users, and would perhaps like to use (parts of) the sodrive system, don't hesitate to contact me.

6 . 3

slide-18
SLIDE 18

ACKNOWLEDGEMENTS ACKNOWLEDGEMENTS

Coen Schrijvers and colleagues at SURFSara for user documentation and monitoring. Catalin Condurache (RAL) for the fail-over Stratum-1. Ronald Starink for the initial setup of the CVMFS system at Nikhef. http://doc.grid.surfsara.nl/en/latest/Pages/Advanced/

7 . 1