mirror of
https://github.com/ceph/ceph
synced 2024-12-18 17:37:38 +00:00
e061d8d148
Add ability to set monitor_interface and public_network values (as opposed to using hard coded values) This change should be squashed with the previous change before merging this pull request Signed-off-by: Warren Usui <wusui@magna002.ceph.redhat.com> |
||
---|---|---|
.. | ||
ceph_install_w_ansible | ||
execs | ||
files | ||
ceph_install.sh | ||
connectceph.sh | ||
copy_func.sh | ||
fix_conf_file.sh | ||
image_create.sh | ||
openstack.sh | ||
packstack.sh | ||
README |
This directory contains scripts to quickly bring up an OpenStack instance, attach a ceph cluster, create a nova compute node, and store the associated glance images, cinder volumes, nova vm, and cinder backup on ceph via rbd. execs is a directory that contains executables that are copied and remotely run on the OpenStack instance files is a directory that contains templates used to initialize OpenStack conf files. These templates reflect the state of these conf files on 5/17/2016. If further development is necessary in the future, these templates should probably be removed and direct editing of the OpenStack conf files should probably be performed. These scripts also assume that either there is a rhel iso file named rhel-server-7.2-x86_64-boot.iso in the user's home directory, or the exported variable RHEL_ISO is set to point at an existing rhel iso file. If one is also running the ceph-deploy based ceph_install.sh, this script also assumes that there is a file named rhceph-1.3.1-rhel-7-x86_64-dvd.iso in the files directory. These iso files can be obtained from the rhel site and are not stored with these scripts. To install openstack: ./openstack.sh <openstack-admin-node> <ceph-monitor-node> This assumes that the ceph cluster is already set up. To setup a ceph-cluster using an iso and ceph-deploy: ./ceph_install.sh <admin-node> <mon-node> <osd-node> <osd-node> <osd-node> To setup a ceph-cluster using the cdn and ceph-ansible: cd ceph_install_w_ansible ./ceph_install.sh <admin-node> <mon-node> <osd-node> <osd-node> <osd-node>