orchestrator on ra fu internals benefits and
play

Orchestrator on Ra fu : internals, benefits and considerations - PowerPoint PPT Presentation

Orchestrator on Ra fu : internals, benefits and considerations Shlomi Noach GitHub PerconaLive 2018 About me @github/database-infrastructure Author of orchestrator , gh-ost , freno , ccql and others. Blog at http://openark.org @ShlomiNoach


  1. Orchestrator on Ra fu : internals, benefits and considerations Shlomi Noach GitHub PerconaLive 2018

  2. About me @github/database-infrastructure Author of orchestrator , gh-ost , freno , ccql and others. Blog at http://openark.org @ShlomiNoach

  3. Agenda Raft overview Why orchestrator/raft orchestrator/raft implementation and nuances HA, fencing Service discovery Considerations

  4. Ra fu Consensus algorithm � Quorum based In-order replication log � � Delivery, lag Snapshots � �

  5. HashiCorp ra fu golang raft implementation Used by Consul Recently hit 1.0.0 github.com/hashicorp/raft

  6. orchestrator � � � MySQL high availability solution and � replication topology manager � Developed at GitHub � � � � Apache 2 license � � github.com/github/orchestrator � � � � � �

  7. Why orchestrator/ra fu Remove MySQL backend dependency DC fencing And then good things happened that were not planned: Better cross-DC deployments DC-local KV control Kubernetes friendly

  8. orchestrator/ra fu � � � n orchestrator nodes form a raft cluster � Each node has its own,dedicated backend � database (MySQL or SQLite) � � � � All nodes probe the topologies � � All nodes run failure detection � � Only the leader runs failure recoveries � � � �

  9. Implementation & deployment @ GitHub 5 Nodes (2xDC1, 2xDC2, 1xDc3) � 2xDC1 � 1 second raft polling interval step-down � 2xDC2 � raft-yield SQLite-backed log store � DC3 � MySQL backend (SQLite backend use case in the works)

  10. A high availability scenario o2 is leader of a 3-node orchestrator/raft setup � o1 � � � � o2 � � � � � � o3 �

  11. Injecting failure master: killall -9 mysqld o2 detects failure. About to recover, but… � o1 � � � � o2 � � � � � � o3 �

  12. Injecting 2nd failure o2: DROP DATABASE orchestrator; o2 freaks out. 5 seconds later it steps � down o1 � � � � o2 � � � � � � o3 �

  13. orchestrator recovery o1 grabs leadership � o1 � � � � o2 � � � � � � o3 �

  14. MySQL recovery o1 detected failure even before stepping up as leader. � � o1 , now leader, kicks recovery, fails over o1 � MySQL master � � � o2 � � � � o3 �

  15. orchestrator self health tests Meanwhile, o2 panics and bails out. � � o1 � � � � o2 � � � � o3 �

  16. puppet Some time later, puppet kicks orchestrator service back on o2 . � � o1 � � � � o2 � � � � o3 �

  17. orchestrator startup orchestrator service on o2 bootstraps, creates orchestrator schema and � tables. � o1 � � � � o2 � � � � o3 �

  18. Joining ra fu cluster o2 recovers from raft snapshot, acquires raft log from an active node, rejoins the � group � o1 � � � � o2 � � � � o3 �

  19. Grabbing leadership Some time later, o2 grabs leadership � � o1 � � � � o2 � � � � o3 �

  20. DC fencing Assume this 3 DC setup One orchestrator node in each DC � DC1 � Master and a few replicas in DC2 � � � DC2 � � � What happens if DC2 gets network partitioned? � � � DC3 � i.e. no network in or out DC2

  21. DC fencing From the point of view of DC2 servers, and in particular in the point of view of DC2 ’s orchestrator node: � DC1 � Master and replicas are fine. � � � DC2 � � � DC1 and DC3 servers are all dead. � � � DC3 No need for fail over. � However, DC2 ’s orchestrator is not part of a quorum, hence not the leader. It doesn’t call the shots.

  22. DC fencing In the eyes of either DC1 ’s or DC3 ’s orchestrator: � All DC2 servers, including the master, are DC1 � dead. � � � DC2 � � � There is need for failover. � � � DC3 DC1 ’s and DC3 ’s orchestrator nodes form a � quorum. One of them will become the leader. The leader will initiate failover.

  23. DC fencing Depicted potential failover result. New � � � master is from DC3 . � � DC1 � DC2 � � � DC3 � � �

  24. orchestrator/ra fu & consul orchestrator is Consul-aware Upon failover orchestrator updates Consul KV with identity of promoted master Consul @ GitHub is DC-local, no replication between Consul setups orchestrator nodes, update Consul locally on each DC

  25. Considerations, watch out for Eventual consistency is not always your best friend What happens if, upon replay of raft log, you hit two failovers for the same cluster? NOW() and otherwise time-based assumptions Reapplying snapshot/log upon startup

  26. orchestrator/ra fu roadmap Kubernetes ClusterIP-based configuration in progress Already container-friendly via auto-reprovisioning of nodes via Raft

  27. Thank you! Questions? github.com/shlomi-noach @ShlomiNoach

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend