Hi,
I've had the best zapping experience with ceph-volume. Have you tried this:
# ceph orch device zap host1 /dev/sdc --force
This worked quite well for me, or you can also try:
# ceph-volume lvm zap --destroy /dev/sdc
Using virtual disks should be just fine.
Zitat von "Scharfenberg, Carsten" <c.scharfenberg@xxxxxxxxxxxxx>:
Thanks for your support, guys.
Unfortunately I do not know the tool sgdisk. It’s also not available
from the standard Debian package repository.
So I’ve tried out Yury’s approach to use dd… without success:
root@terraformdemo:~# dd if=/dev/zero of=/dev/sdc bs=1M count=1024
1024+0 records in
1024+0 records out
1073741824 bytes (1.1 GB, 1.0 GiB) copied, 1.05943 s, 1.0 GB/s
root@terraformdemo:~# dd if=/dev/zero of=/dev/sde bs=1M count=1024
1024+0 records in
1024+0 records out
1073741824 bytes (1.1 GB, 1.0 GiB) copied, 0.618606 s, 1.7 GB/s
root@terraformdemo:~# ceph-volume inventory
Device Path Size rotates available Model name
/dev/sdc 20.00 GB True True VMware Virtual S
/dev/sde 20.00 GB True True VMware Virtual S
/dev/sda 20.00 GB True False VMware Virtual S
/dev/sdb 20.00 GB True False VMware Virtual S
/dev/sdd 20.00 GB True False VMware Virtual S
root@terraformdemo:~# ceph orch device ls
root@terraformdemo:~#
Do you have any other ideas? Could it be that ceph is not usable
with this kind of virtual harddisk?
--
Carsten
From: Сергей Процун <prosergey07@xxxxxxxxx>
Sent: Donnerstag, 4. November 2021 21:36
To: Yury Kirsanov <y.kirsanov@xxxxxxxxx>
Cc: Zach Heise <heise@xxxxxxxxxxxx>; Scharfenberg, Carsten
<c.scharfenberg@xxxxxxxxxxxxx>; ceph-users <ceph-users@xxxxxxx>
Subject: Re: Re: fresh pacific installation does not
detect available disks
ACHTUNG: Diese Mail kommt von einem EXTERNEN ABSENDER - bitte
VORSICHT beim Öffnen von ANHÄNGEN und im Umgang mit LINKS.
Hello,
I agree with that point. When ceph creates lvm volumes it adds lvm
tags to them. Thats how ceph finds that those they are occupied by
ceph. So you should remove lvm volumes and even better clean all
data on those lvm volumes. Usually its enough to clean just the head
of lvm partition where it stores information of the volumes itself.
---
Sergey Protsun
чт, 4 лист. 2021, 22:29 користувач Yury Kirsanov
<y.kirsanov@xxxxxxxxx<mailto:y.kirsanov@xxxxxxxxx>> пише:
Hi,
You should erase any partitions or LVM groups on the disks and restart OSD
hosts so CEPH would be able to detect drives. I usually just do 'dd
if=/dev/zero of=/dev/<sd*> bs=1M count=1024' and then reboot host to make
sure it will definitely be clean. Or, alternatively, you can zap the
drives, or you can just remove LVM groups using pvremove or remove
patitions using fdisk.
Regards,
Yury.
On Fri, 5 Nov 2021, 07:24 Zach Heise,
<heise@xxxxxxxxxxxx<mailto:heise@xxxxxxxxxxxx>> wrote:
Hi Carsten,
When I had problems on my physical hosts (recycled systems that we wanted
to
just use in a test cluster) I found that I needed to use sgdisk --zap-all
/dev/sd{letter} to clean all partition maps off the disks before ceph would
recognize them as available. Worth a shot in your case, even though as
fresh
virtual volumes they shouldn't have anything on them (yet) anyway.
-----Original Message-----
From: Scharfenberg, Carsten
<c.scharfenberg@xxxxxxxxxxxxx<mailto:c.scharfenberg@xxxxxxxxxxxxx>>
Sent: Thursday, November 4, 2021 12:59 PM
To: ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx>
Subject: fresh pacific installation does not detect available
disks
Hello everybody,
as ceph newbie I've tried out setting up ceph pacific according to the
official documentation: https://docs.ceph.com/en/latest/cephadm/install/
The intention was to setup a single node "cluster" with radosgw to feature
local S3 storage.
This failed because my ceph "cluster" would not detect OSDs.
I started from a Debain 11.1 (bullseye) VM hosted on VMware workstation. Of
course I've added some additional disk images to be used as OSDs.
These are the steps I've performed:
curl --silent --remote-name --location
https://github.com/ceph/ceph/raw/pacific/src/cephadm/cephadm
chmod +x cephadm
./cephadm add-repo --release pacific
./cephadm install
apt install -y cephadm
cephadm bootstrap --mon-ip <my_ip>
cephadm add-repo --release pacific
cephadm install ceph-common
ceph orch apply osd --all-available-devices
The last command would have no effect. Its sole output is:
Scheduled osd.all-available-devices update...
Also ceph -s shows that no OSDs were added:
cluster:
id: 655a7a32-3bbf-11ec-920e-000c29da2e6a
health: HEALTH_WARN
OSD count 0 < osd_pool_default_size 1
services:
mon: 1 daemons, quorum terraformdemo (age 2d)
mgr: terraformdemo.aylzbb(active, since 2d)
osd: 0 osds: 0 up, 0 in (since 2d)
data:
pools: 0 pools, 0 pgs
objects: 0 objects, 0 B
usage: 0 B used, 0 B / 0 B avail
pgs:
To find out what may be going wrong I've also tried out this:
cephadm install ceph-osd
ceph-volume inventory
This results in a list that makes more sense:
Device Path Size rotates available Model name
/dev/sdc 20.00 GB True True VMware Virtual S
/dev/sde 20.00 GB True True VMware Virtual S
/dev/sda 20.00 GB True False VMware Virtual S
/dev/sdb 20.00 GB True False VMware Virtual S
/dev/sdd 20.00 GB True False VMware Virtual S
So how can I convince cephadm to use the available devices?
Regards,
Carsten
_______________________________________________
ceph-users mailing list --
ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx> To unsubscribe send
an email
to ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx>
To unsubscribe send an email to
ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx>
To unsubscribe send an email to
ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx