Re: Filestore to Bluestore migration question

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Wow, after all of this, everything went well and I was able to convert osd.120-129 from Filestore to Bluestore. 

***
root@osd2:~# ls -l /var/lib/ceph/osd/ceph-120
total 48
-rw-r--r-- 1 ceph ceph 384 Nov  7 14:34 activate.monmap
lrwxrwxrwx 1 ceph ceph  19 Nov  7 14:38 block -> /dev/hdd120/data120
lrwxrwxrwx 1 ceph ceph  15 Nov  7 14:38 block.db -> /dev/ssd0/db120
-rw-r--r-- 1 ceph ceph   2 Nov  7 14:34 bluefs
-rw-r--r-- 1 ceph ceph  37 Nov  7 14:38 ceph_fsid
-rw-r--r-- 1 ceph ceph  37 Nov  7 14:38 fsid
-rw------- 1 ceph ceph  57 Nov  7 14:38 keyring
-rw-r--r-- 1 ceph ceph   8 Nov  7 14:34 kv_backend
-rw-r--r-- 1 ceph ceph  21 Nov  7 14:34 magic
-rw-r--r-- 1 ceph ceph   4 Nov  7 14:34 mkfs_done
-rw-r--r-- 1 ceph ceph  41 Nov  7 14:34 osd_key
-rw-r--r-- 1 ceph ceph   6 Nov  7 14:38 ready
-rw-r--r-- 1 ceph ceph  10 Nov  7 14:38 type
-rw-r--r-- 1 ceph ceph   4 Nov  7 14:38 whoami
***

and df -h showing
tmpfs           126G   48K  126G   1% /var/lib/ceph/osd/ceph-120
tmpfs           126G   48K  126G   1% /var/lib/ceph/osd/ceph-121
tmpfs           126G   48K  126G   1% /var/lib/ceph/osd/ceph-122
tmpfs           126G   48K  126G   1% /var/lib/ceph/osd/ceph-123 ....

******
It seems like wipefs did delete all the remnants of the filestore partition correctly since I did not have to do any additional clean-up this time. I basically followed all the steps that I wrote out (with a few minor edits Hector suggested).   THANK YOU SO MUCH!!!  After I work on the rest of this node, I will go back to the previous node and see if I can zap it and start all over again.

On Wed, Nov 7, 2018 at 12:21 PM, Hector Martin <hector@xxxxxxxxxxxxxx> wrote:
On 11/8/18 2:15 AM, Hayashida, Mami wrote:
> Thank you very much.  Yes, I am aware that zapping the SSD and
> converting it to LVM requires stopping all the FileStore OSDs whose
> journals are on that SSD first.  I will add in the `hdparm` to my steps.
> I did run into remnants of gpt information lurking around when trying to
> re-use osd disks in the past -- so that's probably a good preemptive move.

Just for reference, "ceph-volume lvm zap" runs wipefs and also wipes the
beginning of the device separately. It should get rid of the GPT
partition table. hdparm -z just tells the kernel to re-read it (which
should remove any device nodes associated with now-gone partitions).

I just checked the wipefs manpage and it seems it does trigger a
partition table re-read itself, which would make the hdparm unnecessary.
It might be useful if you can check that the partition devices (sda1
etc) exist before the zap command and disappear after it, confirming
that hdparm is not necessary. And if they still exist, then run hdparm,
and if they persist after that too, something's wrong and you should
investigate. GPT partition tables can be notoriously annoying to wipe
because there is a backup at the end of the device, but wipefs *should*
know about that as far as I know.

--
Hector Martin (hector@xxxxxxxxxxxxxx)
Public Key: https://mrcn.st/pub



--
Mami Hayashida
Research Computing Associate

Research Computing Infrastructure
University of Kentucky Information Technology Services
301 Rose Street | 102 James F. Hardymon Building
Lexington, KY 40506-0495
mami.hayashida@xxxxxxx
(859)323-7521
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux