Re: Bug in OSD cache management?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Jan,

it would be nice to see how the state evolves. Is there a chance to get
a log with `debug_prioritycache=20` and `debug_bluestore=20`?

> How I can create osd?

As a workaround you might want to poke with `osd_memory_cache_min`
which, by default, is just 128 MB.

CCing BlueStore folks.

Regards,
Radek

On Fri, Aug 4, 2023 at 12:58 AM Jan Marek <jmarek@xxxxxx> wrote:
>
> Hello,
>
> I would like to add OSD to cluster, but I've got this error:
>
> # ceph-volume --cluster ceph lvm create --bluestore --data /dev/sda --block.wal /dev/ceph/wal.a --block.db /dev/ceph/db.a
> Running command: /usr/bin/ceph-authtool --gen-print-key
> Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 459cc86e-999a-4398-bd20-d03031f98a10
> Running command: vgcreate --force --yes ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb /dev/sda
>  stdout: Physical volume "/dev/sda" successfully created.
>  stdout: Volume group "ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb" successfully created
> Running command: lvcreate --yes -l 1907721 -n osd-block-459cc86e-999a-4398-bd20-d03031f98a10 ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb
>  stdout: Logical volume "osd-block-459cc86e-999a-4398-bd20-d03031f98a10" created.
> Running command: /usr/bin/ceph-authtool --gen-print-key
> Running command: /usr/bin/mount -t tmpfs tmpfs /var/lib/ceph/osd/ceph-0
> --> Executable selinuxenabled not in PATH: /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
> Running command: /usr/bin/chown -h ceph:ceph /dev/ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb/osd-block-459cc86e-999a-4398-bd20-d03031f98a10
> Running command: /usr/bin/chown -R ceph:ceph /dev/dm-24
> Running command: /usr/bin/ln -s /dev/ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb/osd-block-459cc86e-999a-4398-bd20-d03031f98a10 /var/lib/ceph/osd/ceph-0/block
> Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring mon getmap -o /var/lib/ceph/osd/ceph-0/activate.monmap
>  stderr: 2023-02-09T13:10:37.252+0100 7fdc2b98a700 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin: (2) No such file or directory
> 2023-02-09T13:10:37.252+0100 7fdc2b98a700 -1 AuthRegistry(0x7fdc240607d0) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin, disabling cephx
>  stderr: got monmap epoch 254
> --> Creating keyring file for osd.0
> Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/keyring
> Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0/
> Running command: /usr/bin/chown -h ceph:ceph /dev/ceph/wal.a
> Running command: /usr/bin/chown -R ceph:ceph /dev/dm-22
> Running command: /usr/bin/chown -h ceph:ceph /dev/ceph/db.a
> Running command: /usr/bin/chown -R ceph:ceph /dev/dm-23
> Running command: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap --keyfile - --bluestore-block-wal-path /dev/ceph/wal.a --bluestore-block-db-path /dev/ceph/db.a --osd-data /var/lib/ceph/osd/ceph-0/ --osd-uuid 459cc86e-999a-4398-bd20-d03031f98a10 --setuser ceph --setgroup ceph
>  stderr: 2023-02-09T13:10:37.728+0100 7f54531aa240 -1 bluestore(/var/lib/ceph/osd/ceph-0/) _read_fsid unparsable uuid
>  stderr: ./src/common/PriorityCache.cc: In function 'void PriorityCache::Manager::balance()' thread 7f5446944700 time 2023-02-09T13:10:39.843937+0100
>  stderr: ./src/common/PriorityCache.cc: 300: FAILED ceph_assert(mem_avail >= 0)
>  stderr: ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy (stable)
>  stderr: 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x124) [0x563da783bf70]
>  stderr: 2: /usr/bin/ceph-osd(+0xc2310e) [0x563da783c10e]
>  stderr: 3: (PriorityCache::Manager::balance()+0x478) [0x563da8459938]
>  stderr: 4: (BlueStore::MempoolThread::entry()+0x82a) [0x563da7e9e27a]
>  stderr: 5: /lib/x86_64-linux-gnu/libpthread.so.0(+0x7ea7) [0x7f5453804ea7]
>  stderr: 6: clone()
>  stderr: *** Caught signal (Aborted) **
>  stderr: in thread 7f5446944700 thread_name:bstore_mempool
>  stderr: 2023-02-09T13:10:39.844+0100 7f5446944700 -1 ./src/common/PriorityCache.cc: In function 'void PriorityCache::Manager::balance()' thread 7f5446944700 time 2023-02-09T13:10:39.843937+0100
>  stderr: ./src/common/PriorityCache.cc: 300: FAILED ceph_assert(mem_avail >= 0)
>  stderr: ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy (stable)
>  stderr: 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x124) [0x563da783bf70]
>  stderr: 2: /usr/bin/ceph-osd(+0xc2310e) [0x563da783c10e]
>  stderr: 3: (PriorityCache::Manager::balance()+0x478) [0x563da8459938]
>  stderr: 4: (BlueStore::MempoolThread::entry()+0x82a) [0x563da7e9e27a]
>  stderr: 5: /lib/x86_64-linux-gnu/libpthread.so.0(+0x7ea7) [0x7f5453804ea7]
>  stderr: 6: clone()
>  stderr: ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy (stable)
>  stderr: 1: /lib/x86_64-linux-gnu/libpthread.so.0(+0x13140) [0x7f5453810140]
>  stderr: 2: gsignal()
>  stderr: 3: abort()
>  stderr: 4: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x17e) [0x563da783bfca]
>  stderr: 5: /usr/bin/ceph-osd(+0xc2310e) [0x563da783c10e]
>  stderr: 6: (PriorityCache::Manager::balance()+0x478) [0x563da8459938]
>  stderr: 7: (BlueStore::MempoolThread::entry()+0x82a) [0x563da7e9e27a]
>  stderr: 8: /lib/x86_64-linux-gnu/libpthread.so.0(+0x7ea7) [0x7f5453804ea7]
>  stderr: 9: clone()
>  stderr: 2023-02-09T13:10:39.852+0100 7f5446944700 -1 *** Caught signal (Aborted) **
>  stderr: in thread 7f5446944700 thread_name:bstore_mempool
>  stderr: ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy (stable)
>  stderr: 1: /lib/x86_64-linux-gnu/libpthread.so.0(+0x13140) [0x7f5453810140]
>  stderr: 2: gsignal()
>  stderr: 3: abort()
>  stderr: 4: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x17e) [0x563da783bfca]
>  stderr: 5: /usr/bin/ceph-osd(+0xc2310e) [0x563da783c10e]
>  stderr: 6: (PriorityCache::Manager::balance()+0x478) [0x563da8459938]
>  stderr: 7: (BlueStore::MempoolThread::entry()+0x82a) [0x563da7e9e27a]
>  stderr: 8: /lib/x86_64-linux-gnu/libpthread.so.0(+0x7ea7) [0x7f5453804ea7]
>  stderr: 9: clone()
>  stderr: NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.
>  stderr: -3212> 2023-02-09T13:10:37.728+0100 7f54531aa240 -1 bluestore(/var/lib/ceph/osd/ceph-0/) _read_fsid unparsable uuid
>  stderr: -1> 2023-02-09T13:10:39.844+0100 7f5446944700 -1 ./src/common/PriorityCache.cc: In function 'void PriorityCache::Manager::balance()' thread 7f5446944700 time 2023-02-09T13:10:39.843937+0100
>  stderr: ./src/common/PriorityCache.cc: 300: FAILED ceph_assert(mem_avail >= 0)
>  stderr: ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy (stable)
>  stderr: 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x124) [0x563da783bf70]
>  stderr: 2: /usr/bin/ceph-osd(+0xc2310e) [0x563da783c10e]
>  stderr: 3: (PriorityCache::Manager::balance()+0x478) [0x563da8459938]
>  stderr: 4: (BlueStore::MempoolThread::entry()+0x82a) [0x563da7e9e27a]
>  stderr: 5: /lib/x86_64-linux-gnu/libpthread.so.0(+0x7ea7) [0x7f5453804ea7]
>  stderr: 6: clone()
>  stderr: 0> 2023-02-09T13:10:39.852+0100 7f5446944700 -1 *** Caught signal (Aborted) **
>  stderr: in thread 7f5446944700 thread_name:bstore_mempool
>  stderr: ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy (stable)
>  stderr: 1: /lib/x86_64-linux-gnu/libpthread.so.0(+0x13140) [0x7f5453810140]
>  stderr: 2: gsignal()
>  stderr: 3: abort()
>  stderr: 4: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x17e) [0x563da783bfca]
>  stderr: 5: /usr/bin/ceph-osd(+0xc2310e) [0x563da783c10e]
>  stderr: 6: (PriorityCache::Manager::balance()+0x478) [0x563da8459938]
>  stderr: 7: (BlueStore::MempoolThread::entry()+0x82a) [0x563da7e9e27a]
>  stderr: 8: /lib/x86_64-linux-gnu/libpthread.so.0(+0x7ea7) [0x7f5453804ea7]
>  stderr: 9: clone()
>  stderr: NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.
> --> Was unable to complete a new OSD, will rollback changes
> Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd purge-new osd.0 --yes-i-really-mean-it
>  stderr: 2023-02-09T13:10:40.072+0100 7f6963aec700 -1 auth: unable to find a keyring on /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin: (2) No such file or directory
> 2023-02-09T13:10:40.072+0100 7f6963aec700 -1 AuthRegistry(0x7f695c0607d0) no keyring found at /etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc/ceph/keyring.bin, disabling cephx
>  stderr: purged osd.0
> --> Zapping: /dev/ceph/db.a
> --> Unmounting /var/lib/ceph/osd/ceph-0
> Running command: /usr/bin/umount -v /var/lib/ceph/osd/ceph-0
>  stderr: umount: /var/lib/ceph/osd/ceph-0 odpojen
> Running command: /usr/bin/dd if=/dev/zero of=/dev/ceph/db.a bs=1M count=10 conv=fsync
>  stderr: 10+0 záznamů přečteno
> 10+0 záznamů zapsáno
> 10485760 bajtů (10 MB, 10 MiB) zkopírováno, 0,0144712 s, 725 MB/s
> --> More than 1 LV left in VG, will proceed to destroy LV only
> --> Removing LV because --destroy was given: /dev/ceph/db.a
> Running command: lvremove -v -f /dev/ceph/db.a
>  stdout: Logical volume "db.a" successfully removed
>  stderr: Removing ceph-db.a (254:23)
>  stderr: Archiving volume group "ceph" metadata (seqno 85).
>  stderr: Releasing logical volume "db.a"
>  stderr: Creating volume group backup "/etc/lvm/backup/ceph" (seqno 86).
> --> Zapping: /dev/ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb/osd-block-459cc86e-999a-4398-bd20-d03031f98a10
> Running command: /usr/bin/dd if=/dev/zero of=/dev/ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb/osd-block-459cc86e-999a-4398-bd20-d03031f98a10 bs=1M count=10 conv=fsync
>  stderr: 10+0 záznamů přečteno
> 10+0 záznamů zapsáno
>  stderr: 10485760 bajtů (10 MB, 10 MiB) zkopírováno, 0,096139 s, 109 MB/s
> --> Only 1 LV left in VG, will proceed to destroy volume group ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb
> Running command: vgremove -v -f ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb
>  stderr: Removing ceph--8f15ecd8--3b3f--404f--a851--8bbba15e30bb-osd--block--459cc86e--999a--4398--bd20--d03031f98a10 (254:24)
>  stderr: Archiving volume group "ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb" metadata (seqno 5).
>  stderr: Releasing logical volume "osd-block-459cc86e-999a-4398-bd20-d03031f98a10"
>  stderr: Creating volume group backup "/etc/lvm/backup/ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb" (seqno 6).
>  stdout: Logical volume "osd-block-459cc86e-999a-4398-bd20-d03031f98a10" successfully removed
>  stderr: Removing physical volume "/dev/sda" from volume group "ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb"
>  stdout: Volume group "ceph-8f15ecd8-3b3f-404f-a851-8bbba15e30bb" successfully removed
> Running command: pvremove -v -f -f /dev/sda
>  stdout: Labels on physical volume "/dev/sda" successfully wiped.
> --> Zapping: /dev/ceph/wal.a
> Running command: /usr/bin/dd if=/dev/zero of=/dev/ceph/wal.a bs=1M count=10 conv=fsync
>  stderr: 10+0 záznamů přečteno
> 10+0 záznamů zapsáno
> 10485760 bajtů (10 MB, 10 MiB) zkopírováno, 0,014648 s, 716 MB/s
> --> More than 1 LV left in VG, will proceed to destroy LV only
> --> Removing LV because --destroy was given: /dev/ceph/wal.a
> Running command: lvremove -v -f /dev/ceph/wal.a
>  stdout: Logical volume "wal.a" successfully removed
>  stderr: Removing ceph-wal.a (254:22)
>  stderr: Archiving volume group "ceph" metadata (seqno 86).
>  stderr: Releasing logical volume "wal.a"
>  stderr: Creating volume group backup "/etc/lvm/backup/ceph" (seqno 87).
> --> Zapping successful for OSD: 0
> -->  RuntimeError: Command failed with exit code 250: /usr/bin/ceph-osd --cluster ceph --osd-objectstore bluestore --mkfs -i 0 --monmap /var/lib/ceph/osd/ceph-0/activate.monmap --keyfile - --bluestore-block-wal-path /dev/ceph/wal.a --bluestore-block-db-path /dev/ceph/db.a --osd-data /var/lib/ceph/osd/ceph-0/ --osd-uuid 459cc86e-999a-4398-bd20-d03031f98a10 --setuser ceph --setgroup ceph
>
> My server have 64GB RAM.
>
> Version of ceph package is 17.2.5
>
> ceph config dump | grep osd
>
> global                   dev       osd_max_pg_log_entries                    10
> global                   advanced  osd_memory_target_autotune                true
> global                   dev       osd_min_pg_log_entries                    10
> global                   dev       osd_pg_log_dups_tracked                   10
> global                   dev       osd_pg_log_trim_min                       10
> global                   advanced  osd_pool_default_min_size                 2
> global                   advanced  osd_pool_default_size                     3
> osd                      dev       bluefs_allocator                          bitmap                                                                                     *
> osd                      advanced  bluestore_allocator                       bitmap                                                                                     *
> osd                      advanced  bluestore_default_buffered_write          true
> osd                      dev       bluestore_fsck_quick_fix_on_mount         true
> osd                      dev       osd_memory_cache_min                      134217728
> osd         host:c-osd2  basic     osd_memory_target                         3927826739
> osd         host:c-osd3  basic     osd_memory_target                         3927839880
> osd         host:c-osd4  basic     osd_memory_target                         3927826739
> osd         host:c-osd5  basic     osd_memory_target                         3927826739
> osd         host:c-osd6  basic     osd_memory_target                         3927826739
> osd                      basic     osd_memory_target                         1073741824
>
> How I can create osd?
>
> I've tried `ceph orch`, but its got the same error...
>
> Sincerely
> Jan Marek
> --
> Ing. Jan Marek
> University of South Bohemia
> Academic Computer Centre
> Phone: +420389032080
> http://www.gnu.org/philosophy/no-word-attachments.cs.html
> _______________________________________________
> Dev mailing list -- dev@xxxxxxx
> To unsubscribe send an email to dev-leave@xxxxxxx
_______________________________________________
Dev mailing list -- dev@xxxxxxx
To unsubscribe send an email to dev-leave@xxxxxxx




[Index of Archives]     [CEPH Users]     [Ceph Devel]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux