chore: docs for ceph (#440)
- add maintenance mode, how to bootstrap an osd, remove an osd Reviewed-on: #440
This commit is contained in:
parent
1077bdcbc1
commit
3c114371e0
@ -58,3 +58,78 @@ this will overwrite the current capabilities of a given client.user
|
|||||||
mon 'allow r' \
|
mon 'allow r' \
|
||||||
mds 'allow rw path=/' \
|
mds 'allow rw path=/' \
|
||||||
osd 'allow rw pool=media_data'
|
osd 'allow rw pool=media_data'
|
||||||
|
|
||||||
|
## adding a new osd on new node
|
||||||
|
|
||||||
|
create the ceph conf (automate this?)
|
||||||
|
|
||||||
|
cat <<EOF | sudo tee /etc/ceph/ceph.conf
|
||||||
|
[global]
|
||||||
|
auth_client_required = cephx
|
||||||
|
auth_cluster_required = cephx
|
||||||
|
auth_service_required = cephx
|
||||||
|
fsid = de96a98f-3d23-465a-a899-86d3d67edab8
|
||||||
|
mon_allow_pool_delete = true
|
||||||
|
mon_initial_members = prodnxsr0009,prodnxsr0010,prodnxsr0011,prodnxsr0012,prodnxsr0013
|
||||||
|
mon_host = 198.18.23.9,198.18.23.10,198.18.23.11,198.18.23.12,198.18.23.13
|
||||||
|
ms_bind_ipv4 = true
|
||||||
|
ms_bind_ipv6 = false
|
||||||
|
osd_crush_chooseleaf_type = 1
|
||||||
|
osd_pool_default_min_size = 2
|
||||||
|
osd_pool_default_size = 3
|
||||||
|
osd_pool_default_pg_num = 128
|
||||||
|
public_network = 198.18.23.1/32,198.18.23.2/32,198.18.23.3/32,198.18.23.4/32,198.18.23.5/32,198.18.23.6/32,198.18.23.7/32,198.18.23.8/32,198.18.23.9/32,198.18.23.10/32,198.18.23.11/32,198.18.23.12/32,198.18.23.13/32
|
||||||
|
EOF
|
||||||
|
|
||||||
|
ssh to one of the monitor hosts, then transfer the keys required
|
||||||
|
|
||||||
|
sudo cat /etc/ceph/ceph.client.admin.keyring | ssh prodnxsr0003 'sudo tee /etc/ceph/ceph.client.admin.keyring'
|
||||||
|
sudo cat /var/lib/ceph/bootstrap-osd/ceph.keyring | ssh prodnxsr0003 'sudo tee /var/lib/ceph/bootstrap-osd/ceph.keyring'
|
||||||
|
|
||||||
|
assuming we are adding /dev/sda to the cluster, first zap the disk to remove partitions/lvm/metadata
|
||||||
|
|
||||||
|
sudo ceph-volume lvm zap /dev/sda --destroy
|
||||||
|
|
||||||
|
then add it to the cluster
|
||||||
|
|
||||||
|
sudo ceph-volume lvm create --data /dev/sda
|
||||||
|
|
||||||
|
## removing an osd
|
||||||
|
|
||||||
|
check what OSD IDs were on this host (if you know it)
|
||||||
|
|
||||||
|
sudo ceph osd tree
|
||||||
|
|
||||||
|
or check for any DOWN osds
|
||||||
|
|
||||||
|
sudo ceph osd stat
|
||||||
|
sudo ceph health detail
|
||||||
|
|
||||||
|
once you identify the old OSD ID, remove it with these steps, replace X with the actual OSD ID:
|
||||||
|
|
||||||
|
sudo ceph osd out osd.X
|
||||||
|
sudo ceph osd down osd.X
|
||||||
|
sudo ceph osd crush remove osd.X
|
||||||
|
sudo ceph auth del osd.X
|
||||||
|
sudo ceph osd rm osd.X
|
||||||
|
|
||||||
|
|
||||||
|
## maintenance mode for the cluster
|
||||||
|
|
||||||
|
from one node in the cluster disable recovery
|
||||||
|
|
||||||
|
sudo ceph osd set noout
|
||||||
|
sudo ceph osd set nobackfill
|
||||||
|
sudo ceph osd set norecover
|
||||||
|
sudo ceph osd set norebalance
|
||||||
|
sudo ceph osd set nodown
|
||||||
|
sudo ceph osd set pause
|
||||||
|
|
||||||
|
to undo the change, use unset
|
||||||
|
|
||||||
|
sudo ceph osd unset noout
|
||||||
|
sudo ceph osd unset nobackfill
|
||||||
|
sudo ceph osd unset norecover
|
||||||
|
sudo ceph osd unset norebalance
|
||||||
|
sudo ceph osd unset nodown
|
||||||
|
sudo ceph osd unset pause
|
||||||
|
|||||||
Loading…
Reference in New Issue
Block a user