mirror of
https://github.com/ceph/ceph
synced 2024-12-24 20:33:27 +00:00
d1d848276f
wait_for_replay_complete() doesn't wait for image status to get updated. This didn't matter previously because these tests are run on two different pools and nothing else was following. Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
99 lines
4.8 KiB
Bash
Executable File
99 lines
4.8 KiB
Bash
Executable File
#!/usr/bin/env bash
|
|
#
|
|
# rbd_mirror_bootstrap.sh - test peer bootstrap create/import
|
|
#
|
|
|
|
set -ex
|
|
|
|
RBD_MIRROR_MANUAL_PEERS=1
|
|
RBD_MIRROR_INSTANCES=${RBD_MIRROR_INSTANCES:-1}
|
|
. $(dirname $0)/rbd_mirror_helpers.sh
|
|
|
|
setup
|
|
|
|
testlog "TEST: bootstrap cluster2 from cluster1"
|
|
# create token on cluster1 and import to cluster2
|
|
TOKEN=${TEMPDIR}/peer-token
|
|
TOKEN_2=${TEMPDIR}/peer-token-2
|
|
CEPH_ARGS='' rbd --cluster ${CLUSTER1} mirror pool peer bootstrap create ${POOL} > ${TOKEN}
|
|
CEPH_ARGS='' rbd --cluster ${CLUSTER1} mirror pool peer bootstrap create ${PARENT_POOL} > ${TOKEN_2}
|
|
cmp ${TOKEN} ${TOKEN_2}
|
|
|
|
CEPH_ARGS='' rbd --cluster ${CLUSTER2} --pool ${POOL} mirror pool peer bootstrap import ${TOKEN} --direction rx-only
|
|
CEPH_ARGS='' rbd --cluster ${CLUSTER2} --pool ${PARENT_POOL} mirror pool peer bootstrap import ${TOKEN} --direction rx-tx
|
|
|
|
start_mirrors ${CLUSTER1}
|
|
start_mirrors ${CLUSTER2}
|
|
|
|
testlog "TEST: verify rx-only direction"
|
|
# rx-only peer is added immediately by "rbd mirror pool peer bootstrap import"
|
|
rbd --cluster ${CLUSTER2} --pool ${POOL} mirror pool info --format json | jq -e '.peers[0].direction == "rx-only"'
|
|
# tx-only peer is added asynchronously by mirror_peer_ping class method
|
|
while ! rbd --cluster ${CLUSTER1} --pool ${POOL} mirror pool info --format json | jq -e '.peers | length > 0'; do
|
|
sleep 1
|
|
done
|
|
rbd --cluster ${CLUSTER1} --pool ${POOL} mirror pool info --format json | jq -e '.peers[0].direction == "tx-only"'
|
|
|
|
create_image_and_enable_mirror ${CLUSTER1} ${POOL} image1
|
|
|
|
wait_for_image_replay_started ${CLUSTER2} ${POOL} image1
|
|
write_image ${CLUSTER1} ${POOL} image1 100
|
|
wait_for_replay_complete ${CLUSTER2} ${CLUSTER1} ${POOL} image1
|
|
wait_for_replaying_status_in_pool_dir ${CLUSTER2} ${POOL} image1
|
|
|
|
testlog "TEST: verify rx-tx direction"
|
|
# both rx-tx peers are added immediately by "rbd mirror pool peer bootstrap import"
|
|
rbd --cluster ${CLUSTER1} --pool ${PARENT_POOL} mirror pool info --format json | jq -e '.peers[0].direction == "rx-tx"'
|
|
rbd --cluster ${CLUSTER2} --pool ${PARENT_POOL} mirror pool info --format json | jq -e '.peers[0].direction == "rx-tx"'
|
|
|
|
create_image ${CLUSTER1} ${PARENT_POOL} image1
|
|
create_image ${CLUSTER2} ${PARENT_POOL} image2
|
|
|
|
enable_mirror ${CLUSTER1} ${PARENT_POOL} image1
|
|
enable_mirror ${CLUSTER2} ${PARENT_POOL} image2
|
|
|
|
wait_for_image_replay_started ${CLUSTER2} ${PARENT_POOL} image1
|
|
write_image ${CLUSTER1} ${PARENT_POOL} image1 100
|
|
wait_for_replay_complete ${CLUSTER2} ${CLUSTER1} ${PARENT_POOL} image1
|
|
wait_for_replaying_status_in_pool_dir ${CLUSTER2} ${PARENT_POOL} image1
|
|
|
|
wait_for_image_replay_started ${CLUSTER1} ${PARENT_POOL} image2
|
|
write_image ${CLUSTER2} ${PARENT_POOL} image2 100
|
|
wait_for_replay_complete ${CLUSTER1} ${CLUSTER2} ${PARENT_POOL} image2
|
|
wait_for_replaying_status_in_pool_dir ${CLUSTER1} ${PARENT_POOL} image2
|
|
|
|
testlog "TEST: pool replayer and callout cleanup when peer is updated"
|
|
test_health_state ${CLUSTER1} ${PARENT_POOL} 'OK'
|
|
test_health_state ${CLUSTER2} ${PARENT_POOL} 'OK'
|
|
POOL_STATUS=$(get_pool_status_json ${CLUSTER2} ${PARENT_POOL})
|
|
jq -e '.summary.health == "OK"' <<< ${POOL_STATUS}
|
|
jq -e '.summary.daemon_health == "OK"' <<< ${POOL_STATUS}
|
|
jq -e '.daemons[0].health == "OK"' <<< ${POOL_STATUS}
|
|
jq -e '.daemons[0] | has("callouts") | not' <<< ${POOL_STATUS}
|
|
OLD_SERVICE_ID=$(jq -r '.daemons[0].service_id' <<< ${POOL_STATUS})
|
|
OLD_INSTANCE_ID=$(jq -r '.daemons[0].instance_id' <<< ${POOL_STATUS})
|
|
# mess up the peer on one of the clusters by setting a bogus user name
|
|
PEER_UUID=$(rbd --cluster ${CLUSTER2} --pool ${PARENT_POOL} mirror pool info --format json | jq -r '.peers[0].uuid')
|
|
rbd --cluster ${CLUSTER2} --pool ${PARENT_POOL} mirror pool peer set ${PEER_UUID} client client.invalid
|
|
wait_for_health_state ${CLUSTER2} ${PARENT_POOL} 'ERROR'
|
|
test_health_state ${CLUSTER1} ${PARENT_POOL} 'WARNING'
|
|
POOL_STATUS=$(get_pool_status_json ${CLUSTER2} ${PARENT_POOL})
|
|
jq -e '.summary.health == "ERROR"' <<< ${POOL_STATUS}
|
|
jq -e '.summary.daemon_health == "ERROR"' <<< ${POOL_STATUS}
|
|
jq -e '.daemons[0].health == "ERROR"' <<< ${POOL_STATUS}
|
|
jq -e '.daemons[0].callouts == ["unable to connect to remote cluster"]' <<< ${POOL_STATUS}
|
|
# restore the correct user name
|
|
rbd --cluster ${CLUSTER2} --pool ${PARENT_POOL} mirror pool peer set ${PEER_UUID} client client.rbd-mirror-peer
|
|
wait_for_health_state ${CLUSTER2} ${PARENT_POOL} 'OK'
|
|
test_health_state ${CLUSTER1} ${PARENT_POOL} 'OK'
|
|
POOL_STATUS=$(get_pool_status_json ${CLUSTER2} ${PARENT_POOL})
|
|
jq -e '.summary.health == "OK"' <<< ${POOL_STATUS}
|
|
jq -e '.summary.daemon_health == "OK"' <<< ${POOL_STATUS}
|
|
jq -e '.daemons[0].health == "OK"' <<< ${POOL_STATUS}
|
|
jq -e '.daemons[0] | has("callouts") | not' <<< ${POOL_STATUS}
|
|
NEW_SERVICE_ID=$(jq -r '.daemons[0].service_id' <<< ${POOL_STATUS})
|
|
NEW_INSTANCE_ID=$(jq -r '.daemons[0].instance_id' <<< ${POOL_STATUS})
|
|
# check that we are running the same service (daemon) but a newer pool replayer
|
|
((OLD_SERVICE_ID == NEW_SERVICE_ID))
|
|
((OLD_INSTANCE_ID < NEW_INSTANCE_ID))
|