mirror of
https://github.com/ceph/go-ceph
synced 2024-12-22 22:24:03 +00:00
6f2a33a012
Add a function to bring up a 2nd cephfs file system. Execute it on ceph versions pacific and later. While the API exists on nautilus and octopus those versions are already deprecated in go-ceph. Additionally, the ceph docs for "multifs" state, "Beginning with the Pacific release, multiple file system support is stable and ready to use." [1] Therefore we won't bother running our test automation on anything earlier. 1 - https://docs.ceph.com/en/latest/cephfs/multifs/#multiple-ceph-file-systems Signed-off-by: John Mulligan <jmulligan@redhat.com>
213 lines
5.9 KiB
Bash
Executable File
213 lines
5.9 KiB
Bash
Executable File
#!/bin/bash
|
|
#
|
|
# Copyright (C) 2013,2014 Loic Dachary <loic@dachary.org>
|
|
#
|
|
# This program is free software: you can redistribute it and/or modify
|
|
# it under the terms of the GNU Affero General Public License as published by
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# This program is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU Affero General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU Affero General Public License
|
|
# along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
#
|
|
set -e
|
|
set -x
|
|
set -u
|
|
|
|
DIR=${1}
|
|
|
|
# reset
|
|
pkill ceph || true
|
|
rm -rf "${DIR:?}"/*
|
|
LOG_DIR="${DIR}/log"
|
|
MON_DATA="${DIR}/mon"
|
|
MDS_DATA="${DIR}/mds"
|
|
MOUNTPT="${MDS_DATA}/mnt"
|
|
OSD_DATA="${DIR}/osd"
|
|
RGW_DATA="${DIR}/radosgw"
|
|
mkdir "${LOG_DIR}" "${MON_DATA}" "${OSD_DATA}" "${MDS_DATA}" "${MOUNTPT}" "${RGW_DATA}"
|
|
MDS_NAME="Z"
|
|
FS_NAME="cephfs"
|
|
ALT_MDS_NAME="Y"
|
|
ALT_FS_NAME="altfs"
|
|
MON_NAME="a"
|
|
MGR_NAME="x"
|
|
MIRROR_ID="m"
|
|
RGW_ID="r"
|
|
S3_ACCESS_KEY=2262XNX11FZRR44XWIRD
|
|
S3_SECRET_KEY=rmtuS1Uj1bIC08QFYGW18GfSHAbkPqdsuYynNudw
|
|
|
|
FSID="$(uuidgen)"
|
|
export CEPH_CONF=${DIR}/ceph.conf
|
|
|
|
generate_ceph_conf() {
|
|
# cluster wide parameters
|
|
cat >> "${CEPH_CONF}" <<EOF
|
|
[global]
|
|
fsid = ${FSID}
|
|
osd crush chooseleaf type = 0
|
|
run dir = ${DIR}/run
|
|
auth cluster required = none
|
|
auth service required = none
|
|
auth client required = none
|
|
osd pool default size = 1
|
|
mon host = ${HOSTNAME}
|
|
|
|
[mds.${MDS_NAME}]
|
|
host = ${HOSTNAME}
|
|
|
|
[mon.${MON_NAME}]
|
|
log file = ${LOG_DIR}/mon.log
|
|
chdir = ""
|
|
mon cluster log file = ${LOG_DIR}/mon-cluster.log
|
|
mon data = ${MON_DATA}
|
|
mon data avail crit = 0
|
|
mon addr = ${HOSTNAME}
|
|
mon allow pool delete = true
|
|
|
|
[osd.0]
|
|
log file = ${LOG_DIR}/osd.log
|
|
chdir = ""
|
|
osd data = ${OSD_DATA}
|
|
osd journal = ${OSD_DATA}.journal
|
|
osd journal size = 100
|
|
osd objectstore = memstore
|
|
osd class load list = *
|
|
osd class default list = *
|
|
|
|
[mgr.${MGR_NAME}]
|
|
log_file = ${LOG_DIR}/mgr.log
|
|
|
|
[client.rgw.${RGW_ID}]
|
|
rgw dns name = ${HOSTNAME}
|
|
rgw enable usage log = true
|
|
rgw usage log tick interval = 1
|
|
rgw usage log flush threshold = 1
|
|
rgw usage max shards = 32
|
|
rgw usage max user shards = 1
|
|
log file = ${LOG_DIR}/client.rgw.${RGW_ID}.log
|
|
rgw frontends = beast port=80
|
|
ms mon client mode = crc
|
|
EOF
|
|
}
|
|
|
|
launch_mon() {
|
|
ceph-mon --id ${MON_NAME} --mkfs --keyring /dev/null
|
|
touch "${MON_DATA}/keyring"
|
|
ceph-mon --id ${MON_NAME}
|
|
}
|
|
|
|
launch_osd() {
|
|
OSD_ID=$(ceph osd create)
|
|
ceph osd crush add "osd.${OSD_ID}" 1 root=default
|
|
ceph-osd --id "${OSD_ID}" --mkjournal --mkfs
|
|
ceph-osd --id "${OSD_ID}" || ceph-osd --id "${OSD_ID}" || ceph-osd --id "${OSD_ID}"
|
|
}
|
|
|
|
launch_mds_server() {
|
|
local mds="$1"
|
|
local fs="$2"
|
|
|
|
ceph auth get-or-create "mds.${mds}" mon 'profile mds' mgr 'profile mds' mds 'allow *' osd 'allow *' >> "${MDS_DATA}/keyring"
|
|
ceph osd pool create "${fs}_data" 8
|
|
ceph osd pool create "${fs}_metadata" 8
|
|
ceph fs new "${fs}" "${fs}_metadata" "${fs}_data"
|
|
ceph fs ls
|
|
ceph-mds -i "${mds}"
|
|
ceph status
|
|
while ! ceph mds stat | grep -q "up:active"; do sleep 1; done
|
|
|
|
}
|
|
|
|
launch_mds() {
|
|
launch_mds_server "${MDS_NAME}" "${FS_NAME}"
|
|
}
|
|
|
|
launch_mds2() {
|
|
launch_mds_server "${ALT_MDS_NAME}" "${ALT_FS_NAME}"
|
|
echo "${ALT_FS_NAME}" > "${DIR}/altfs.txt"
|
|
}
|
|
|
|
launch_mgr() {
|
|
ceph-mgr --id ${MGR_NAME}
|
|
}
|
|
|
|
launch_rbd_mirror() {
|
|
ceph auth get-or-create client.rbd-mirror.${MIRROR_ID} mon 'profile rbd-mirror' osd 'profile rbd'
|
|
rbd-mirror --id ${MIRROR_ID} --log-file "${LOG_DIR}/rbd-mirror.log"
|
|
}
|
|
|
|
launch_cephfs_mirror() {
|
|
ceph auth get-or-create "client.cephfs-mirror.${MIRROR_ID}" \
|
|
mon 'profile cephfs-mirror' \
|
|
mds 'allow r' \
|
|
osd 'allow rw tag cephfs metadata=*, allow r tag cephfs data=*' \
|
|
mgr 'allow r'
|
|
cephfs-mirror --id "cephfs-mirror.${MIRROR_ID}" \
|
|
--log-file "${LOG_DIR}/cephfs-mirror.log"
|
|
ceph fs authorize cephfs client.cephfs-mirror-remote / rwps > "${DIR}/cephfs-mirror-remote.out"
|
|
# the .out file above is not used by the scripts but can be used for debugging
|
|
}
|
|
|
|
launch_radosgw() {
|
|
ceph auth get-or-create client.rgw."${RGW_ID}" osd 'allow rwx' mon 'allow rw' -o "${RGW_DATA}/keyring"
|
|
radosgw -n client.rgw."${RGW_ID}" -k "${RGW_DATA}/keyring"
|
|
# not going to try to make shellcheck happy with this line at this time
|
|
# shellcheck disable=SC2016
|
|
timeout 60 sh -c 'until [ $(ceph -s | grep -c "rgw:") -eq 1 ]; do echo "waiting for rgw to show up" && sleep 1; done'
|
|
radosgw-admin user create --uid admin --display-name "Admin User" --caps "buckets=*;users=*;usage=read;metadata=read" --access-key="$S3_ACCESS_KEY" --secret-key="$S3_SECRET_KEY"
|
|
}
|
|
|
|
selftest() {
|
|
ceph --version
|
|
ceph status
|
|
test_pool=$(uuidgen)
|
|
temp_file=$(mktemp)
|
|
ceph osd pool create "${test_pool}" 0
|
|
rados --pool "${test_pool}" put group /etc/group
|
|
rados --pool "${test_pool}" get group "${temp_file}"
|
|
diff /etc/group "${temp_file}"
|
|
ceph osd pool delete "${test_pool}" "${test_pool}" --yes-i-really-really-mean-it
|
|
rm "${temp_file}"
|
|
}
|
|
|
|
FEATURESET="${CEPH_FEATURESET-}"
|
|
if [ -z "$FEATURESET" ] ; then
|
|
case "${CEPH_VERSION-}" in
|
|
nautilus|octopus)
|
|
FEATURESET="mon osd mgr mds rbd-mirror rgw selftest"
|
|
;;
|
|
*)
|
|
FEATURESET="mon osd mgr mds mds2 rbd-mirror cephfs-mirror rgw selftest"
|
|
;;
|
|
esac
|
|
fi
|
|
|
|
generate_ceph_conf
|
|
for fname in ${FEATURESET} ; do
|
|
case "${fname}" in
|
|
mon) launch_mon ;;
|
|
osd) launch_osd ;;
|
|
mds) launch_mds ;;
|
|
mds2) launch_mds2 ;;
|
|
mgr) launch_mgr ;;
|
|
rbd-mirror) launch_rbd_mirror ;;
|
|
cephfs-mirror) launch_cephfs_mirror ;;
|
|
rgw|radosgw) launch_radosgw ;;
|
|
selftest) selftest ;;
|
|
*)
|
|
echo "Invalid feature: ${fname}"
|
|
exit 2
|
|
;;
|
|
esac
|
|
done
|
|
|
|
touch "${DIR}/.ready"
|
|
|
|
# vim: set ts=4 sw=4 sts=4 et:
|