How to replace a failed disk on ceph node
We need to know how to replace a failed OSD disk in a Ceph cluster deployed by Fuel 6.1. All our ceph nodes uses SSD drives for Ceph journals.
The firsts steps are quite simple but not complete :
-identify the failed osd : ceph osd tree | grep down
-identify the journal partition used by the osd : ceph-disk list (don't work as the partition are listed but not the relationship to the osd)
-identify the mount point of the failed osd : mount (we figured out the ceph partition is the third one)
-unmount the partition : umount ${CEPH_OSD_PART}
-replace the disk and setup the controller so it recognize the new disk
-out the osd : ceph osd out ${CEPH_OSD} && stop ceph-osd id=${CEPH_OSD_NUM}
-remove the crush map : ceph osd crush remove ${CEPH_OSD} && ceph auth del ${CEPH_OSD} && ceph osd rm ${CEPH_OSD}
What would be the next steps to recreate the osd and reuse the existing journal partition?
Question information
- Language:
- English Edit question
- Status:
- Expired
- Assignee:
- No assignee Edit question
- Last query:
- Last reply: