after wal device crash can not recreate osd

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello together,

i need to recreate the osd´s on one ceph-node, because the nvme wal device
has died. I replaced the nvme to a brand new one and i try now to recreate
the osd´s on this node, but i get an error while re-creating them.
Can somebody tell me why i get this error? i never saw before

this is what i do:

root@cd88-ceph-osdh-01:~# ceph osd out osd.2
marked out osd.2.

root@cd88-ceph-osdh-01:~# systemctl stop ceph-osd@osd.2

root@cd88-ceph-osdh-01:~# ceph osd crush remove osd.2
removed item id 2 name 'osd.2' from crush map

root@cd88-ceph-osdh-01:~# ceph auth del osd.2
updated

root@cd88-ceph-osdh-01:~# ceph osd rm osd.2
removed osd.2

root@cd88-ceph-osdh-01:~# ceph-volume lvm zap /dev/ceph-block-0/block-0
--> Zapping: /dev/ceph-block-0/block-0
Running command: /usr/bin/dd if=/dev/zero of=/dev/ceph-block-0/block-0
bs=1M count=10 conv=fsync
 stderr: 10+0 records in
10+0 records out
10485760 bytes (10 MB, 10 MiB) copied, 0.0878055 s, 119 MB/s
 stderr:
--> Zapping successful for: <LV: /dev/ceph-block-0/block-0>
root@cd88-ceph-osdh-01:~# ceph-volume lvm zap /dev/ceph-db-0/db-0
--> Zapping: /dev/ceph-db-0/db-0
Running command: /usr/bin/dd if=/dev/zero of=/dev/ceph-db-0/db-0 bs=1M
count=10 conv=fsync
 stderr: 10+0 records in
10+0 records out
 stderr: 10485760 bytes (10 MB, 10 MiB) copied, 0.0376431 s, 279 MB/s
--> Zapping successful for: <LV: /dev/ceph-db-0/db-0>

root@cd88-ceph-osdh-01:/# vi /var/lib/ceph/bootstrap-osd/ceph.keyring

root@cd88-ceph-osdh-01:/# ceph-volume lvm prepare --no-systemd --bluestore
--data ceph-block-0/block-0 --block.wal ceph-wal-0/waldb-0 --block.db
ceph-db-0/db-0
Running command: /usr/bin/ceph-authtool --gen-print-key
Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd
--keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new
f2dcdce7-0b88-4a83-b9ca-2f4546c9a718
Running command: /usr/bin/ceph-authtool --gen-print-key
Running command: /usr/bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-2
Running command: /usr/bin/chown -h ceph:ceph /dev/ceph-block-0/block-0
Running command: /usr/bin/chown -R ceph:ceph /dev/dm-27
Running command: /usr/bin/ln -s /dev/ceph-block-0/block-0
/var/lib/ceph/osd/ceph-2/block
Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd
--keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o
/var/lib/ceph/osd/ceph-2/activate.monmap
 stderr: got monmap epoch 3
Running command: /usr/bin/ceph-authtool /var/lib/ceph/osd/ceph-2/keyring
--create-keyring --name osd.2 --add-key
AQCFXa5hn+TBJxAArSTZFWMavJqWGISwSR8Xow==
 stdout: creating /var/lib/ceph/osd/ceph-2/keyring
added entity osd.2 auth(key=AQCFXa5hn+TBJxAArSTZFWMavJqWGISwSR8Xow==)
Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-2/keyring
Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-2/
Running command: /usr/bin/chown -h ceph:ceph /dev/ceph-wal-0/waldb-0
Running command: /usr/bin/chown -R ceph:ceph /dev/dm-40
Running command: /usr/bin/chown -h ceph:ceph /dev/ceph-db-0/db-0
Running command: /usr/bin/chown -R ceph:ceph /dev/dm-2
Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore
bluestore --mkfs -i 2 --monmap /var/lib/ceph/osd/ceph-2/activate.monmap
--keyfile - --bluestore-block-wal-path /dev/ceph-wal-0/waldb-0
--bluestore-block-db-path /dev/ceph-db-0/db-0 --osd-data
/var/lib/ceph/osd/ceph-2/ --osd-uuid f2dcdce7-0b88-4a83-b9ca-2f4546c9a718
--setuser ceph --setgroup ceph
 stderr: 2021-12-06T18:59:19.652+0000 7f3ea7a84080 -1
bluestore(/var/lib/ceph/osd/ceph-2/) _read_fsid unparsable uuid
 stderr:
/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/16.2.5/rpm/el8/BUILD/ceph-16.2.5/src/common/PriorityCache.cc:
In function 'void PriorityCache::Manager::balance()' thread 7f3e835a3700
time 2021-12-06T18:59:21.531373+0000
 stderr:
/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/16.2.5/rpm/el8/BUILD/ceph-16.2.5/src/common/PriorityCache.cc:
301: FAILED ceph_assert(mem_avail >= 0)
 stderr: ceph version 16.2.5 (0883bdea7337b95e4b611c768c0279868462204a)
pacific (stable)
 stderr: 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char
const*)+0x158) [0x558fe9b69eba]
 stderr: 2: /usr/bin/ceph-osd(+0x56a0d4) [0x558fe9b6a0d4]
 stderr: 3: (PriorityCache::Manager::balance()+0x4e1) [0x558fea6f1a61]
 stderr: 4: (BlueStore::MempoolThread::entry()+0x889) [0x558fea17f259]
 stderr: 5: /lib64/libpthread.so.0(+0x814a) [0x7f3ea57db14a]
 stderr: 6: clone()
 stderr: 2021-12-06T18:59:21.528+0000 7f3e835a3700 -1
/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/16.2.5/rpm/el8/BUILD/ceph-16.2.5/src/common/PriorityCache.cc:
In function 'void PriorityCache::Manager::balance()' thread 7f3e835a3700
time 2021-12-06T18:59:21.531373+0000
 stderr:
/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/16.2.5/rpm/el8/BUILD/ceph-16.2.5/src/common/PriorityCache.cc:
301: FAILED ceph_assert(mem_avail >= 0)
 stderr: ceph version 16.2.5 (0883bdea7337b95e4b611c768c0279868462204a)
pacific (stable)
 stderr: 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char
const*)+0x158) [0x558fe9b69eba]
 stderr: 2: /usr/bin/ceph-osd(+0x56a0d4) [0x558fe9b6a0d4]
 stderr: 3: (PriorityCache::Manager::balance()+0x4e1) [0x558fea6f1a61]
 stderr: 4: (BlueStore::MempoolThread::entry()+0x889) [0x558fea17f259]
 stderr: 5: /lib64/libpthread.so.0(+0x814a) [0x7f3ea57db14a]
 stderr: 6: clone()
 stderr: *** Caught signal (Aborted) **
 stderr: in thread 7f3e835a3700 thread_name:bstore_mempool
 stderr: ceph version 16.2.5 (0883bdea7337b95e4b611c768c0279868462204a)
pacific (stable)
 stderr: 1: /lib64/libpthread.so.0(+0x12b20) [0x7f3ea57e5b20]
 stderr: 2: gsignal()
 stderr: 3: abort()
 stderr: 4: (ceph::__ceph_assert_fail(char const*, char const*, int, char
const*)+0x1a9) [0x558fe9b69f0b]
 stderr: 5: /usr/bin/ceph-osd(+0x56a0d4) [0x558fe9b6a0d4]
 stderr: 6: (PriorityCache::Manager::balance()+0x4e1) [0x558fea6f1a61]
 stderr: 7: (BlueStore::MempoolThread::entry()+0x889) [0x558fea17f259]
 stderr: 8: /lib64/libpthread.so.0(+0x814a) [0x7f3ea57db14a]
 stderr: 9: clone()
 stderr: 2021-12-06T18:59:21.532+0000 7f3e835a3700 -1 *** Caught signal
(Aborted) **
 stderr: in thread 7f3e835a3700 thread_name:bstore_mempool
 stderr: ceph version 16.2.5 (0883bdea7337b95e4b611c768c0279868462204a)
pacific (stable)
 stderr: 1: /lib64/libpthread.so.0(+0x12b20) [0x7f3ea57e5b20]
 stderr: 2: gsignal()
 stderr: 3: abort()
 stderr: 4: (ceph::__ceph_assert_fail(char const*, char const*, int, char
const*)+0x1a9) [0x558fe9b69f0b]
 stderr: 5: /usr/bin/ceph-osd(+0x56a0d4) [0x558fe9b6a0d4]
 stderr: 6: (PriorityCache::Manager::balance()+0x4e1) [0x558fea6f1a61]
 stderr: 7: (BlueStore::MempoolThread::entry()+0x889) [0x558fea17f259]
 stderr: 8: /lib64/libpthread.so.0(+0x814a) [0x7f3ea57db14a]
 stderr: 9: clone()
 stderr: NOTE: a copy of the executable, or `objdump -rdS <executable>` is
needed to interpret this.
 stderr: -2953> 2021-12-06T18:59:19.652+0000 7f3ea7a84080 -1
bluestore(/var/lib/ceph/osd/ceph-2/) _read_fsid unparsable uuid
 stderr: -1> 2021-12-06T18:59:21.528+0000 7f3e835a3700 -1
/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/16.2.5/rpm/el8/BUILD/ceph-16.2.5/src/common/PriorityCache.cc:
In function 'void PriorityCache::Manager::balance()' thread 7f3e835a3700
time 2021-12-06T18:59:21.531373+0000
 stderr:
/home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos8/DIST/centos8/MACHINE_SIZE/gigantic/release/16.2.5/rpm/el8/BUILD/ceph-16.2.5/src/common/PriorityCache.cc:
301: FAILED ceph_assert(mem_avail >= 0)
 stderr: ceph version 16.2.5 (0883bdea7337b95e4b611c768c0279868462204a)
pacific (stable)
 stderr: 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char
const*)+0x158) [0x558fe9b69eba]
 stderr: 2: /usr/bin/ceph-osd(+0x56a0d4) [0x558fe9b6a0d4]
 stderr: 3: (PriorityCache::Manager::balance()+0x4e1) [0x558fea6f1a61]
 stderr: 4: (BlueStore::MempoolThread::entry()+0x889) [0x558fea17f259]
 stderr: 5: /lib64/libpthread.so.0(+0x814a) [0x7f3ea57db14a]
 stderr: 6: clone()
 stderr: 0> 2021-12-06T18:59:21.532+0000 7f3e835a3700 -1 *** Caught signal
(Aborted) **
 stderr: in thread 7f3e835a3700 thread_name:bstore_mempool
 stderr: ceph version 16.2.5 (0883bdea7337b95e4b611c768c0279868462204a)
pacific (stable)
 stderr: 1: /lib64/libpthread.so.0(+0x12b20) [0x7f3ea57e5b20]
 stderr: 2: gsignal()
 stderr: 3: abort()
 stderr: 4: (ceph::__ceph_assert_fail(char const*, char const*, int, char
const*)+0x1a9) [0x558fe9b69f0b]
 stderr: 5: /usr/bin/ceph-osd(+0x56a0d4) [0x558fe9b6a0d4]
 stderr: 6: (PriorityCache::Manager::balance()+0x4e1) [0x558fea6f1a61]
 stderr: 7: (BlueStore::MempoolThread::entry()+0x889) [0x558fea17f259]
 stderr: 8: /lib64/libpthread.so.0(+0x814a) [0x7f3ea57db14a]
 stderr: 9: clone()
 stderr: NOTE: a copy of the executable, or `objdump -rdS <executable>` is
needed to interpret this.
 stderr: *** Caught signal (Segmentation fault) **
 stderr: in thread 7f3e835a3700 thread_name:bstore_mempool
 stderr: ceph version 16.2.5 (0883bdea7337b95e4b611c768c0279868462204a)
pacific (stable)
 stderr: 1: /lib64/libpthread.so.0(+0x12b20) [0x7f3ea57e5b20]
 stderr: 2: pthread_getname_np()
 stderr: 3: (ceph::logging::Log::dump_recent()+0x4b3) [0x558fea4ae233]
 stderr: 4: /usr/bin/ceph-osd(+0xcab9fb) [0x558fea2ab9fb]
 stderr: 5: /lib64/libpthread.so.0(+0x12b20) [0x7f3ea57e5b20]
 stderr: 6: gsignal()
 stderr: 7: abort()
 stderr: 8: (ceph::__ceph_assert_fail(char const*, char const*, int, char
const*)+0x1a9) [0x558fe9b69f0b]
 stderr: 9: /usr/bin/ceph-osd(+0x56a0d4) [0x558fe9b6a0d4]
 stderr: 10: (PriorityCache::Manager::balance()+0x4e1) [0x558fea6f1a61]
 stderr: 11: (BlueStore::MempoolThread::entry()+0x889) [0x558fea17f259]
 stderr: 12: /lib64/libpthread.so.0(+0x814a) [0x7f3ea57db14a]
 stderr: 13: clone()
--> Was unable to complete a new OSD, will rollback changes
Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd
--keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd purge-new osd.2
--yes-i-really-mean-it
 stderr: purged osd.2
-->  RuntimeError: Command failed with exit code 250: /usr/bin/ceph-osd
--cluster ceph --osd-objectstore bluestore --mkfs -i 2 --monmap
/var/lib/ceph/osd/ceph-2/activate.monmap --keyfile -
--bluestore-block-wal-path /dev/ceph-wal-0/waldb-0
--bluestore-block-db-path /dev/ceph-db-0/db-0 --osd-data
/var/lib/ceph/osd/ceph-2/ --osd-uuid f2dcdce7-0b88-4a83-b9ca-2f4546c9a718
--setuser ceph --setgroup ceph
root@cd88-ceph-osdh-01:/#
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux