Let's review some initial cluster state commands to help determine where we need to being our troubleshooting:
- Validate the overall health of the Ceph storage cluster :
# ceph health detail
- If the cluster is in a HEALTH_OK state, this command really does not tell you much. But if there are PGs or OSDs in the error state, this command can provide further details about the state of the cluster. Let's stop one of our OSD processes on ceph-node2 and then rerun the ceph health detail to see what it reports:
root@ceph-node2 # systemctl stop ceph-osd@<id> # ceph health detail
- With an OSD process stopped, you can see that ...