Hi, You should erase any partitions or LVM groups on the disks and restart OSD hosts so CEPH would be able to detect drives. I usually just do 'dd if=/dev/zero of=/dev/<sd*> bs=1M count=1024' and then reboot host to make sure it will definitely be clean. Or, alternatively, you can zap the drives, or you can just remove LVM groups using pvremove or remove patitions using fdisk. Regards, Yury. On Fri, 5 Nov 2021, 07:24 Zach Heise, <heise@xxxxxxxxxxxx> wrote: > Hi Carsten, > > When I had problems on my physical hosts (recycled systems that we wanted > to > just use in a test cluster) I found that I needed to use sgdisk --zap-all > /dev/sd{letter} to clean all partition maps off the disks before ceph would > recognize them as available. Worth a shot in your case, even though as > fresh > virtual volumes they shouldn't have anything on them (yet) anyway. > > -----Original Message----- > From: Scharfenberg, Carsten <c.scharfenberg@xxxxxxxxxxxxx> > Sent: Thursday, November 4, 2021 12:59 PM > To: ceph-users@xxxxxxx > Subject: fresh pacific installation does not detect available > disks > > Hello everybody, > > as ceph newbie I've tried out setting up ceph pacific according to the > official documentation: https://docs.ceph.com/en/latest/cephadm/install/ > The intention was to setup a single node "cluster" with radosgw to feature > local S3 storage. > This failed because my ceph "cluster" would not detect OSDs. > I started from a Debain 11.1 (bullseye) VM hosted on VMware workstation. Of > course I've added some additional disk images to be used as OSDs. > These are the steps I've performed: > > curl --silent --remote-name --location > https://github.com/ceph/ceph/raw/pacific/src/cephadm/cephadm > chmod +x cephadm > ./cephadm add-repo --release pacific > ./cephadm install > > apt install -y cephadm > > cephadm bootstrap --mon-ip <my_ip> > > cephadm add-repo --release pacific > > cephadm install ceph-common > > ceph orch apply osd --all-available-devices > > > The last command would have no effect. Its sole output is: > > Scheduled osd.all-available-devices update... > > > > Also ceph -s shows that no OSDs were added: > > cluster: > > id: 655a7a32-3bbf-11ec-920e-000c29da2e6a > > health: HEALTH_WARN > > OSD count 0 < osd_pool_default_size 1 > > > > services: > > mon: 1 daemons, quorum terraformdemo (age 2d) > > mgr: terraformdemo.aylzbb(active, since 2d) > > osd: 0 osds: 0 up, 0 in (since 2d) > > > > data: > > pools: 0 pools, 0 pgs > > objects: 0 objects, 0 B > > usage: 0 B used, 0 B / 0 B avail > > pgs: > > > To find out what may be going wrong I've also tried out this: > > cephadm install ceph-osd > > ceph-volume inventory > This results in a list that makes more sense: > > Device Path Size rotates available Model name > > /dev/sdc 20.00 GB True True VMware Virtual S > > /dev/sde 20.00 GB True True VMware Virtual S > > /dev/sda 20.00 GB True False VMware Virtual S > > /dev/sdb 20.00 GB True False VMware Virtual S > > /dev/sdd 20.00 GB True False VMware Virtual S > > > So how can I convince cephadm to use the available devices? > > Regards, > Carsten > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email > to ceph-users-leave@xxxxxxx > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx