4.2 KiB
Assuming the following details, and we are trying to recover mysql-innodb-cluster/0
mysql-innodb-cluster/0 active idle 0/lxd/6 10.0.1.137 Unit is ready: Mode: R/O, Cluster is ONLINE and can tolerate up to ONE failure.
mysql-innodb-cluster/1 active idle 1/lxd/6 10.0.1.114 Unit is ready: Mode: R/W, Cluster is ONLINE and can tolerate up to ONE failure.
mysql-innodb-cluster/2* active idle 2/lxd/6 10.0.1.156 Unit is ready: Mode: R/O, Cluster is ONLINE and can tolerate up to ONE failure.
-
Stop the mysql on the unit
mysql-innodb-cluster/0sudo systemctl stop mysql sudo mv /var/lib/mysql /var/lib/mysql.old -
Remove the member from the cluster
juju run-action --wait mysql-innodb-cluster/leader remove-instance address=10.0.1.137If the above command doesn't work run with the parameter
force=truejuju run-action --wait mysql-innodb-cluster/leader remove-instance address=10.0.1.137 force=trueConfirm it worked by checking the IP is removed from:
juju run-action --wait mysql-innodb-cluster/leader cluster-status -
Re-initialise the DB on the machine locally on problematic node i.e.
mysql-innodb-cluster/0juju ssh mysql-innodb-cluster/0 cd /var/lib mv mysql mysql.old.$(date +%s) mkdir mysql chown mysql:mysql mysql chmod 700 mysql mysqld --initialize systemctl start mysqlCheck the mysql status:
sudo systemctl status mysql -
A temporary password would have been created via the
--initializeabove, so this needs to be updated. You will find the temporary password in/var/log/mysql/error.log, and the line should look similar to the one below[Note] [MY-010454] [Server] A temporary password is generated for root@localhost: wyPd_?kEd03pFirst we need to get the root password that is stored
juju run --unit mysql-innodb-cluster/leader 'leader-get mysql.passwd'Now, we can update the password, login using the password that was suggested in the error.log i.e.
wyPd_?kEd03pmysql -p -u rootOnce logged in, we can update the password to the value that mysql-innodb-cluster charm knows about
ALTER USER 'root'@'localhost' IDENTIFIED BY 'zn8K73dmqnkZd99JZxXwcFmxWqTxPYgw3Hjx5sk'; -
Remove the flags using Juju:
Clear flags to force charm to re-create cluster users
juju run --unit mysql-innodb-cluster/0 -- charms.reactive clear_flag local.cluster.user-created juju run --unit mysql-innodb-cluster/0 -- charms.reactive clear_flag local.cluster.all-users-created juju run --unit mysql-innodb-cluster/0 -- ./hooks/update-statusAfter that, you can confirm it worked by getting the password:
juju run --unit mysql-innodb-cluster/leader leader-get cluster-passwordConnect to the unit
mysql-innodb-cluster/0and use the password above:mysql -u clusteruser -p -e 'SELECT user,host FROM mysql.user' -
Re-add instance to cluster:
juju run-action --wait mysql-innodb-cluster/leader add-instance address=10.0.1.137 juju run-action --wait mysql-innodb-cluster/leader cluster-status
Note: If the instance is not added to the cluster use mysqlsh to do this with the step below:
juju ssh mysql-innodb-cluster/2
mysqlsh clusteruser@10.0.1.156 --password=<clusterpassword> --cluster
cluster.add_instance("10.0.1.137:3306")
Choose the option => "[C]lone YES"
You might need to run the command below to configure your instance if you have the error-output below:
"NOTE: Please use the dba.configure_instance() command to repair these issues."
If you have the output above run the command below using the mysqlsh CLI:
dba.configure_instance("clusteruser@10.0.1.137:3306")
Note: You will be asked for the password of the user clusteruser and after this step you can add the instance back to the cluster:
cluster.add_instance("10.0.1.137:3306")
choose the option => "[C]lone YES"
After that check the cluster status:
juju run-action --wait mysql-innodb-cluster/leader cluster-status