ceph-disk is not available in Nautilus. why scrub first? It is a new disk not having any data yet. Scrubbing is verifying pg's not? I just created a vm on the ceph node where I want to add this osd. Did a passthru of the disk and installed a few rpm's with nodeps to get the ceph-disk command. -----Original Message----- Sent: 01 March 2020 17:47 Subject: Re: Is it ok to add a luminous ceph-disk osd to nautilus still? Ensure that it gets scrubbed at least once by Luminous first. But how and why are you doing this ? Why not use Nautilus binaries ? > On Mar 1, 2020, at 8:36 AM, Marc Roos <M.Roos@xxxxxxxxxxxxxxxxx> wrote: > > > If I create and osd with luminous 12.0.3 binaries, can I just add it > to an existing Nautilus cluster? > > I sort of did this already, just wondered if there are any drawbacks. > > > [@test2 software]# ceph-disk prepare --bluestore --zap-disk /dev/sdb > Creating new GPT entries. > GPT data structures destroyed! You may now partition the disk using > fdisk or other utilities. > Creating new GPT entries. > The operation has completed successfully. > Setting name! > partNum is 0 > REALLY setting name! > The operation has completed successfully. > Setting name! > partNum is 1 > REALLY setting name! > The operation has completed successfully. > The operation has completed successfully. > meta-data=/dev/sdb1 isize=2048 agcount=4, agsize=6400 > blks > = sectsz=512 attr=2, projid32bit=1 > = crc=1 finobt=0, sparse=0 > data = bsize=4096 blocks=25600, imaxpct=25 > = sunit=0 swidth=0 blks > naming =version 2 bsize=4096 ascii-ci=0 ftype=1 > log =internal log bsize=4096 blocks=864, version=2 > = sectsz=512 sunit=0 blks, lazy-count=1 > realtime =none extsz=4096 blocks=0, rtextents=0 > The operation has completed successfully. > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an > email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx