Add a manual RADOSGW installation procedure to
doc/install/manual-deployment.rst. This procedure was developed by Janne
Johansson and reported to the ceph-users mailing list on 29 Jan 2024
here: https://lists.ceph.io/hyperkitty/list/ceph-users@ceph.io/message/LB3YRIKAPOHXYCW7MKLVUJPYWYRQVARU/
Co-authored-by: Janne Johansson <icepic.dz@gmail.com>
Co-authored-by: Anthony D'Atri <anthony.datri@gmail.com>
Signed-off-by: Zac Dover <zac.dover@proton.me>
Remove misleading material that would give readers the wrong idea about
when stale submodules are present. This commit is made in response to
information given to me by Ilya Dryomov here: https://github.com/ceph/ceph/pull/54929#issuecomment-1859237986.
Signed-off-by: Zac Dover <zac.dover@proton.me>
add --progress flag to git submodule update commands
Reviewed-by: Xiubo Li <xiubli@redhat.com>
Reviewed-by: Dhairya Parmar <dparmar@redhat.com>
Reviewed-by: Patrick Donnelly <pdonnell@redhat.com>
Ceph has lots of submodules that needs to be cloned before building
binaries from the repository. Seeing the progress when these submodules
are being cloned is useful, especially when developers/users have a
network issue or a slow network.
Signed-off-by: Rishabh Dave <ridave@redhat.com>
* Don't call cephadm a script, it could be misleading
* Use the current release in the example for copy-pasters
* Improve grammar (I think)
Signed-off-by: John Mulligan <jmulligan@redhat.com>
First off, these instructions are wrong because the file has been
renamed to cephadm.py, and will not succeed as written. Amusingly, due
to curl's default behavior (and my lack of access to a time machine) the
initial command will seem to have succeeded until the user discovers
that HTML delivered as part of the body of a 404 response is not an
executable script.
Furthermore, in preparation for Reef we updated the documentation in the
cephadm subsection to use the results of a build, rather than source
files, because reef now contains a "compiled" cephadm. Currently, the
contents of the compilation is essentially the same-old cephadm script
but that will no longer be the case in the very near future (see
7c6a2cad96405616719b85c740aabd98558c93ac). We missed this general
section of the Ceph docs that repeats the instructions to copy a raw
python file out of the Ceph tree on GitHub. Update the section to
contain a simplified version of the instruction found in
doc/cephadm/install.rst.
Signed-off-by: John Mulligan <jmulligan@redhat.com>
The ceph main branch daily build on openEuler has been built.
https://github.com/openeuler-mirror/ceph-daily-build
The verification includes:
- build ceph from source
- run unit test
- build RPM package
Signed-off-by: liuqinfei <lucas.liuqinfei@huawei.com>
The current documentation tries really hard to convince people to set
both `osd_pool_default_pg_num` and `osd_pool_default_pgp_num` in their
configs, but at least the latter has undesirable side effects on any
Ceph version that has PG autoscaling enabled by default (at least quincy
and beyond).
Assume a cluster with defaults of `64` for `pg_num` and `pgp_num`.
Starting `radosgw` will fail as it tries to create various pools without
providing values for `pg_num` or `pgp_num`. This triggers the following
in `OSDMonitor::prepare_new_pool()`:
- `pg_num` is set to `1`, because autoscaling is enabled
- `pgp_num` is set to `osd pool default pgp_num`, which we set to `64`
- This is an invalid setup, so the pool creation fails
Likewise, `ceph osd pool create mypool` (without providing values for
`pg_num` or `pgp_num`) does not work.
Following this rationale:
- Not providing a default value for `pgp_num` will always do the right
thing, unless you use advanced features, in which case you can be
expected to set both values on pool creation
- Setting `osd_pool_default_pgp_num` in your config breaks pool creation
for various cases
This commit:
- Removes `osd_pool_default_pgp_num` from all example configs
- Adds mentions of the autoscaling and how it interacts with the default
values in various places
For each file that was touched, the following maintenance was also
performed:
- Change interternal spaces to underscores for config values
- Remove mentions of filestore or any of its settings
- Fix minor inconsistencies, like indentation etc.
There is also a ticket which I think is very relevant and fixed by this,
though it only captures part of the broader issue addressed here:
Fixes: https://tracker.ceph.com/issues/47176
Signed-off-by: Conrad Hoffmann <ch@bitfehler.net>
Update index.rst by making minor grammar improvements. This file was
long overdue for a backport to Reef, Quincy, and Pacific, so this commit
was a good way to pass a human eyeball over the text before making those
backports.
Signed-off-by: Zac Dover <zac.dover@proton.me>
Refine English sentences in doc/install/index.rst. Remove adverbial
phrases of time that refer to Nautilus-era features as "new", since that
was four years ago.
Signed-off-by: Zac Dover <zac.dover@gmail.com>
Link to "Installing Ceph" in the cephadm documentation instead of (as
was the case before this commit) to the cephadm overview page. Anyone
who clicks on the "cephadm" link in the context of the
doc/install/index.rst page is more likely to expect installation
instructions than to expect an explanation of what cephadm is.
Signed-off-by: Zac Dover <zac.dover@gmail.com>
This commit does not exist in main. It will be forward ported once the
release branches appropriately redirect /en/$release/releases to
/en/latest/releases.
Signed-off-by: Patrick Donnelly <pdonnell@redhat.com>
(cherry picked from commit 01986dffbd8caf366e328f7ab4650cbb76663d3e)
Remove "Docker Hub" from the list of locations from which Ceph can be
acquired.
This commit represents a second attempt to get
https://github.com/ceph/ceph/pull/49034 past the Jenkins CI tests, and
that PR was itself an attempt to incorporate the change in
https://github.com/ceph/ceph/pull/44310, which could not be merged
because the commit message associated with that PR did not begin with
the string "doc", which meant that the "Signed-off-by" check failed.
Signed-off-by: Zac Dover <zac.dover@gmail.com>
hk.ceph.com - notified mirror maintainer list of mirror decommission on 2020-08-31
us-east.ceph.com - mirror has been offline for years.
Signed-off-by: Matthew Taylor <matthew.taylor@digitalpacific.com.au>
add and change the configuration steps to enable latest version of QAT
driver on QAT based encryption and compression
Signed-off-by: Miaomiao Liu <miaomiao.liu@intel.com>
since we are able to build and run on ubuntu focal without extra
dependencies of 3rd party packages, there is no need to mention them.
actually all of the removed packages are available in ubuntu focal.
some of them have different package names though.
the same applies to el8.
Signed-off-by: Kefu Chai <kchai@redhat.com>
Point to pacific for downloading the cephadm script
Co-authored-by: Kefu Chai <tchaikov@gmail.com>
Signed-off-by: Sandro Bonazzola <sbonazzo@redhat.com>
Following the instructions as is results in the following error at step 15:
$ sudo -u ceph ceph-mon --mkfs -i node1 --monmap /tmp/monmap --keyring /tmp/ceph.mon.keyring
global_init: error reading config file.
Signed-off-by: Abutalib Aghayev <agayev@cs.cmu.edu>
'ninja' replaced 'make' in this PR[1]. This patch updates the doc about it.
[1] https://github.com/ceph/ceph/pull/39826
Signed-off-by: Varsha Rao <varao@redhat.com>
Instead of having a direct download link, we'll point to the
download page, which will eventually contain other MSI versions as
well (e.g. Quincy).
While at it, we're simplifying the document a bit, dropping
information that's also included in the manual install guide.
Signed-off-by: Lucian Petrut <lpetrut@cloudbasesolutions.com>
Most of the Windows documentation is currently included in the
README.windows.rst file.
To make it more accessible, we're moving most of it to the
"doc/" folder, adding the following pages:
* Installing Ceph on Windows
* RBD on Windows
* Windows troubleshooting
We'll keep the build and manual install instructions in
README.windows.rst. Note that ceph-dokan already has a separate
doc page.
Signed-off-by: Lucian Petrut <lpetrut@cloudbasesolutions.com>