Support #699
Updated by Daniel Curtis about 9 years ago
{{>toc}} Here’s step-by-step guide to install and configure a MariaDB Galera Clustered Database server software on Debian 8. h2. Prepare the Environments *NOTE*: All three nodes should have their environments setup in the same way. * Make sure the system is up to date: <pre> sudo apt-get update && sudo apt-get upgrade </pre> * Install prerequisite Debian Packages: <pre> sudo apt-get install software-properties-common python-software-properties rsync </pre> * Setup the MariaDB 10.0 Repository: <pre> sudo apt-key adv --recv-keys --keyserver keyserver.ubuntu.com 0xcbcb082a1bb943db sudo add-apt-repository 'deb http://ftp.osuosl.org/pub/mariadb/repo/10.0/debian jessie main' sudo apt-get update </pre> * Install MariaDB 10 with Galera: <pre> sudo apt-get install mariadb-galera-server galera </pre> # Ensure the /etc/hosts files are the same on each Node and that it contains entries for each DB Node’s IP addresses and hostname. <pre> vi /etc/hosts </pre> #* And add/modify it similar to the following <pre> 192.168.0.1 db1 db1.example.com 192.168.0.2 db2 db2.example.com 192.168.0.3 db3 db3.example.com </pre> h2. Configure Node 1 * Create a mysql cluster config: <pre> vi /etc/mysql/conf.d/cluster.cnf </pre> #* And add the following: <pre> [mysqld] binlog_format=ROW default-storage-engine=innodb innodb_autoinc_lock_mode=2 query_cache_size=0 query_cache_type=0 bind-address=0.0.0.0 # Galera Provider Configuration wsrep_provider=/usr/lib/galera/libgalera_smm.so #wsrep_provider_options="gcache.size=32G" # Galera Cluster Configuration wsrep_cluster_name="example_cluster" # Initialize the cluster; switch after --wsrep-new-cluster has been run and other nodes have joined. wsrep_cluster_address="gcomm://192.168.0.1,192.168.0.2,192.168.0.3" wsrep_cluster_address="gcomm:// # Set cluster node addresses (hostnames work too) #wsrep_cluster_address="gcomm://192.168.0.1,192.168.0.2,192.168.0.3" # Galera Synchronization Congifuration wsrep_sst_method=rsync #wsrep_sst_auth=user:pass # Galera Node Configuration wsrep_node_address="192.168.0.1" wsrep_node_name="db1.example.com" </pre> * Restart the MySQL Server on Node 1: <pre> service mysql stop service mysql bootstrap start --wsrep-new-cluster </pre> * Check how MySQL Galera Cluster servers is going after restart with: <pre> tail -n 20 /var/log/mysql/mysqld_safe.log </pre> h2. Configure Node 2 * Load MariaDB Cluster config file, listing Node 1′s IP address as the initial ‘gcomm’ Cluster server, on Node 2 do: <pre> vi /etc/mysql/conf.d/cluster.cnf </pre> #* And add the following: <pre> [mysqld] binlog_format=ROW default-storage-engine=innodb innodb_autoinc_lock_mode=2 query_cache_size=0 query_cache_type=0 bind-address=0.0.0.0 # Galera Provider Configuration wsrep_provider=/usr/lib/galera/libgalera_smm.so #wsrep_provider_options="gcache.size=32G" # Galera Cluster Configuration wsrep_cluster_name="example_cluster" wsrep_cluster_address="gcomm://192.168.0.1,192.168.0.2,192.168.0.3" # Galera Synchronization Congifuration wsrep_sst_method=rsync #wsrep_sst_auth=user:pass # Galera Node Configuration wsrep_node_address="192.168.0.1" wsrep_node_name="db2.example.com" </pre> * Restart the MySQL Server on Node 2 with the above @/etc/mysql/conf.d/cluster.cnf@ file installed: <pre> service mysql restart </pre> * Check how MySQL Galera Cluster servers is going after restart with: Note: Any UUID numbers are replaced with ‘…’ in the following output <pre> cat /var/log/mysql/mysqld_safe.log </pre> #* You should see log file entries about each node joining and being syncronised into the Galera MySQL Cluster on Node 1, entries like: <pre> [Note] WSREP: declaring ... stable [Note] WSREP: Node ... state prim [Note] WSREP: (..., 'tcp://0.0.0.0:4567') turning message relay requesting on, nonlive peers: tcp://192.168.0.3:4567 [Note] WSREP: view(view_id(PRIM,...,11) memb { ..., ..., } joined { } left { } partitioned { ..., }) [Note] WSREP: New COMPONENT: primary = yes, bootstrap = no, my_idx = 1, memb_num = 2 [Note] WSREP: STATE EXCHANGE: Waiting for state UUID. [Note] WSREP: forgetting ... (tcp://192.168.0.3:4567) [Note] WSREP: (..., 'tcp://0.0.0.0:4567') turning message relay requesting off [Note] WSREP: STATE EXCHANGE: sent state msg: ... [Note] WSREP: STATE EXCHANGE: got state msg: ... from 0 (node1.domain.name) [Note] WSREP: STATE EXCHANGE: got state msg: ... from 1 (node2.domain.name) [Note] WSREP: Quorum results: version = 2, component = PRIMARY, conf_id = 8, members = 2/2 (joined/total), act_id = 162, last_appl. = 0, protocols = 0/4/2 (gcs/repl/appl), group UUID = ... [Note] WSREP: Flow-control interval: [23, 23] [Note] WSREP: New cluster view: global state: ...:162, view# 9: Primary, number of nodes: 2, my index: 1, protocol version 2 [Note] WSREP: wsrep_notify_cmd is not defined, skipping notification. [Note] WSREP: Assign initial position for certification: 162, protocol version: 2 14:24:37 [Note] WSREP: cleaning up ... (tcp://192.168.0.3:4567) [Note] WSREP: declaring ... stable [Note] WSREP: declaring ... stable [Note] WSREP: Node ... state prim [Note] WSREP: view(view_id(PRIM,...,12) memb { ..., ..., ..., } joined { } left { } partitioned { }) [Note] WSREP: New COMPONENT: primary = yes, bootstrap = no, my_idx = 1, memb_num = 3 [Note] WSREP: STATE EXCHANGE: Waiting for state UUID. [Note] WSREP: STATE EXCHANGE: sent state msg: ... [Note] WSREP: STATE EXCHANGE: got state msg: ... from 0 (node1.domain.name) [Note] WSREP: STATE EXCHANGE: got state msg: ... from 1 (node2.domain.name) [Note] WSREP: STATE EXCHANGE: got state msg: ... from 2 (node3.domain.name) [Note] WSREP: Quorum results: version = 2, component = PRIMARY, conf_id = 9, members = 3/3 (joined/total), act_id = 162, last_appl. = 0, protocols = 0/4/2 (gcs/repl/appl), group UUID = ... [Note] WSREP: Flow-control interval: [28, 28] [Note] WSREP: New cluster view: global state: ...:162, view# 10: Primary, number of nodes: 3, my index: 1, protocol version 2 [Note] WSREP: wsrep_notify_cmd is not defined, skipping notification. [Note] WSREP: Assign initial position for certification: 162, protocol version: 2 [Note] WSREP: Member 2 (node3.domain.name) synced with group. </pre> * Again, use the mysql client on Node 2 to query Node 2′s Galera Server. <pre> mysql -uroot -p -e "SHOW STATUS LIKE 'wsrep_%';" </pre> h2. Configure Node 3 * Load MariaDB Cluster config file, listing Node 1 and 2′s IP address as the initial ‘gcomm’ Cluster server, on Node 3 do: <pre> vi /etc/mysql/conf.d/cluster.cnf </pre> #* And add the following: <pre> [mysqld] binlog_format=ROW default-storage-engine=innodb innodb_autoinc_lock_mode=2 query_cache_size=0 query_cache_type=0 bind-address=0.0.0.0 # Galera Provider Configuration wsrep_provider=/usr/lib/galera/libgalera_smm.so #wsrep_provider_options="gcache.size=32G" # Galera Cluster Configuration wsrep_cluster_name="example_cluster" wsrep_cluster_address="gcomm://192.168.0.1,192.168.0.2,192.168.0.3" # Galera Synchronization Congifuration wsrep_sst_method=rsync #wsrep_sst_auth=user:pass # Galera Node Configuration wsrep_node_address="192.168.0.1" wsrep_node_name="db3.example.com" </pre> * Restart the MySQL Server on Node 3 with the above @/etc/mysql/conf.d/cluster.cnf@ file installed: <pre> service mysql restart </pre> Check the @mysqld_safe.log@ logfiles on all nodes to see that Node 3 also joins and syncronises with the Cluster. At this point you have all 3 x Nodes in the Galera MySQL Cluster and you can now update Node 1 and Node 2′s cluster.cnf file to set the IP Addresses of all Nodes in the Cluster. * On Node 1, change the line in @/etc/mysql/conf.d/cluster.cnf@ that says: <pre> wsrep_cluster_address = 'gcomm://192.168.0.1,192.168.0.2,192.168.0.3' </pre> And then restart MySQL on Node 1, checking the relevant log files and mysql client command to query the Cluster membership once Node 1′s mysql server software has restarted, with: <pre> service mysql restart </pre> * Check the log for any problems: <pre> tail -f /var/log/mysql/mysqld_safe.log </pre> * Check the status of the Galera cluster <pre> mysql -uroot -p -e "SHOW STATUS LIKE 'wsrep_%';" </pre> h2. Finalize the Galera Cluster configuration Once Node 1 has been reconfigured to know of all Cluster Members, do the same step above on Node 2 and Node 3, setting Node 2 and Node 3′s *wsrep_cluster_address* in @/etc/mysql/conf.d/cluster.cnf@ to: <pre> wsrep_cluster_address = 'gcomm://192.168.0.1,192.168.0.2,192.168.0.3' </pre> And restarting their MySQL Servers and checking that they then each rejoin the live Cluster. Finally by now you should have a running live 3 Node MariaDB Galera MySQL Cluster, the status of which on each Node should list that there’s 3 members of the Cluster, via: <pre> mysql -uroot -pMYSQLROOTPASSWD -e "SHOW STATUS LIKE 'wsrep_%';" | egrep "wsrep_incoming_addresses|wsrep_cluster_size" </pre> h3. Transferring the Debian Maintenance Configuration Currently, Ubuntu and Debian's MariaDB servers use a special maintenance user to do routine maintenance. Some tasks that fall outside of the maintenance category also are run as this user, including important functions like stopping MySQL. With our cluster environment being shared between the individual nodes, the maintenance user, who has randomly generated login credentials on each node, will be unable to execute commands correctly. Only the initial server will have the correct maintenance credentials, since the others will attempt to use their local settings to access the shared cluster environment.This can fix this by simply copying the contents of the maintenance file to each individual node: * On server that bootstrapped the cluster, open the Debian maintenance configuration file: <pre> sudo nano /etc/mysql/debian.cnf </pre> #* The file will look similar to the following: <pre> [client] host = localhost user = debian-sys-maint password = VRP84dIknkX31uOf socket = /var/run/mysqld/mysqld.sock [mysql_upgrade] host = localhost user = debian-sys-maint password = VRP84dIknkX31uOf socket = /var/run/mysqld/mysqld.sock basedir = /usr </pre> * We simply need to copy this information and paste it into the same file on each node. h2. h3. Automatic Cluster Initialization at Boot Newer versions of Galera have options that can enable the cluster to start at boot. If you shut down all nodes, you effectively terminated the cluster (not the data of course, but the running cluster), hence the right way is to start the all the nodes with @gcomm://<node1 address>,<node2 address>,...?pc.wait_prim=no@ again. * On ever node in the Galera cluster edit the mysql cluster configuration file <pre> vi /etc/mysql/conf.d/cluster.cnf </pre> #* And modify the @wsrep_cluster_address@ paramter: <pre> wsrep_cluster_address="gcomm://192.168.0.1,192.168.0.2,192.168.0.3?pc.wait_prim=no" </pre> * On one of the nodes set global @wsrep_provider_options="pc.bootstrap=true"@. <pre> vi /etc/mysql/conf.d/cluster.cnf </pre> #* And add the @wsrep_provider_options@ parameter: <pre> wsrep_provider_options="pc.bootstrap=true" </pre> h2. (Optional) BONUS - Setup Xtrabackup Installing xtrabackup is a decent way to increase the performance and add a minor level of security to the Galera cluster. Xtrabackup is maintained by Percona and as such we need to add their repo and install it the xtrabackup program, then finally modify the cluster.cnf file to use xtrabackup. * Add the percona repo keys <pre> apt-key adv --keyserver keys.gnupg.net --recv-keys 1C4CBDCDCD2EFD2A </pre> * Create the percona repo list file: <pre> vi /etc/apt/sources.list.d/percona.list </pre> #* And add the following: <pre> deb http://repo.percona.com/apt jessie main deb-src http://repo.percona.com/apt jessie main </pre> * Refresh the local cache: <pre> apt-get update </pre> * Install xtrabackup <pre> apt-get install xtrabackup </pre> * On ever node in the Galera cluster edit the mysql cluster configuration file <pre> vi /etc/mysql/conf.d/cluster.cnf </pre> #* And modify the @wsrep_sst_method@ parameter: <pre> wsrep_sst_method=xtrabackup </pre> #* NOTE: If using authentication make sure add the following parameter to the cluster.cnf file: <pre> wsrep_sst_auth=username:SuperSecretPassword </pre> h2. Resources * https://www.digitalocean.com/community/tutorials/how-to-configure-a-galera-cluster-with-mariadb-on-ubuntu-12-04-servers * https://mariadb.com/kb/en/mariadb/documentation/replication/galera/getting-started-with-mariadb-galera-cluster/ * https://mariadb.com/kb/en/mariadb/installing-mariadb-deb-files/ * https://www.linode.com/docs/databases/mariadb/clustering-with-mariadb-and-galera