kvm die with assert(m_seed < old_pg_num)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I just experienced the same behavior as described in http://tracker.ceph.com/issues/10399

I know that Ubuntu 15.04 is not supported, but I thought it might be interesting for you.

I have 2 physical hosts, one with lots of disks and one with only the boot/root disk. On both systems I run kvm. I have 3 VMs running ceph mon, 6 VMs each running 2 ceph osds and some other VMs. The mon and osd VMs boot from virtual disks hosted on local disks of the kvm host. The osds access physical disks on the host using RDM. All other VMs boot from rbd hosted on the ceph cluster.

Here a list of installed software:
root@kvmhost01:~# dpkg -l | grep ceph
ii  ceph                                0.94.2-0ubuntu0.15.04.1             amd64        distributed storage and file system
ii  ceph-common                         0.94.2-0ubuntu0.15.04.1             amd64        common utilities to mount and interact with a ceph storage cluster
ii  ceph-fs-common                      0.94.2-0ubuntu0.15.04.1             amd64        common utilities to mount and interact with a ceph file system
ii  ceph-mds                            0.94.2-0ubuntu0.15.04.1             amd64        metadata server for the ceph distributed file system
ii  libcephfs1                          0.94.2-0ubuntu0.15.04.1             amd64        Ceph distributed file system client library
ii  python-cephfs                       0.94.2-0ubuntu0.15.04.1             amd64        Python libraries for the Ceph libcephfs library
root@kvmhost01:~# dpkg -l | grep -E 'kvm|qemu|libvirt'
ii  ipxe-qemu                           1.0.0+git-20141004.86285d1-1ubuntu3 all          PXE boot firmware - ROM images for qemu
ii  libvirt-bin                         1.2.12-0ubuntu14.1                  amd64        programs for the libvirt library
ii  libvirt0                            1.2.12-0ubuntu14.1                  amd64        library for interfacing with different virtualization systems
ii  qemu-kvm                            1:2.2+dfsg-5expubuntu9.2            amd64        QEMU Full virtualization
ii  qemu-system-common                  1:2.2+dfsg-5expubuntu9.2            amd64        QEMU full system emulation binaries (common files)
ii  qemu-system-x86                     1:2.2+dfsg-5expubuntu9.2            amd64        QEMU full system emulation binaries (x86)
ii  qemu-utils                          1:2.2+dfsg-5expubuntu9.2            amd64        QEMU utilities

Here the error logged by libvirt/qemu:
osd/osd_types.cc: In function 'bool pg_t::is_split(unsigned int, unsigned int, std::set<pg_t>*) const' thread 7f29c
8415700 time 2015-07-26 14:53:17.118199
osd/osd_types.cc: 459: FAILED assert(m_seed < old_pg_num)
ceph version 0.94.2 (5fb85614ca8f354284c713a2f9c610860720bbf3)
1: (()+0x12cc72) [0x7f29d87a3c72]
2: (()+0x20b6f7) [0x7f29d88826f7]
3: (()+0x20b7c0) [0x7f29d88827c0]
4: (()+0x93b81) [0x7f29d870ab81]
5: (()+0xac47f) [0x7f29d872347f]
6: (()+0xacd58) [0x7f29d8723d58]
7: (()+0xae389) [0x7f29d8725389]
8: (()+0xb485f) [0x7f29d872b85f]
9: (()+0x2be40c) [0x7f29d893540c]
10: (()+0x2eee2d) [0x7f29d8965e2d]
11: (()+0x76aa) [0x7f29d42676aa]
12: (clone()+0x6d) [0x7f29d3f9ceed]
NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this.
terminate called after throwing an instance of 'ceph::FailedAssertion'
2015-07-26 12:53:48.440+0000: shutting down


and here some info on ceph and what happend:

root@kvmhost01:~# ceph df
GLOBAL:
    SIZE       AVAIL      RAW USED     %RAW USED
    13333G     13087G         240G          1.80
POOLS:
    NAME                   ID     USED       %USED     MAX AVAIL     OBJECTS
    rbd                    0           0         0         2147G           0
    libvirt-pool           1      69316M      0.51         2147G       18042
    cephfs_data            3       4519M      0.03         2205G        1346
    cephfs_metadata        4       8499k         0         2205G          22
    .rgw.root              6         848         0         2205G           3
    .rgw.control           7           0         0         2205G           8
    .rgw                   8         706         0         2205G           4
    .rgw.gc                9           0         0         2205G          32
    .users.uid             10        675         0         2205G           4
    .users.email           11          8         0         2205G           1
    .users                 12         19         0         2205G           2
    .rgw.buckets.index     13          0         0         2205G           2
    .rgw.buckets           14      7935M      0.06         2205G        4573
root@kvmhost01:~# ceph health
HEALTH_WARN too many PGs per OSD (738 > max 300); pool libvirt-pool has too few pgs
root@kvmhost01:~# rados df
pool name                 KB      objects       clones     degraded      unfound           rd        rd KB           wr        wr KB
.rgw                       1            4            0            0           0           90           70           30            8
.rgw.buckets         8126126         4573            0            0           0         5108         2664        35376      8127189
.rgw.buckets.index         0            2            0            0           0         9968         9962         9934            0
.rgw.control               0            8            0            0           0            0            0            0            0
.rgw.gc                    0           32            0            0           0         4928         4896         3576            0
.rgw.root                  1            3            0            0           0           65           43            3            3
.users                     1            2            0            0           0            0            0            2            2
.users.email               1            1            0            0           0            0            0            1            1
.users.uid                 1            4            0            0           0           70           68           60            2
cephfs_data          4628365         1346            0            0           0         4212       159015         4017      4663669
cephfs_metadata         8500           22            0            0           0           76        11168         4079        10581
libvirt-pool        70980227        18042            0            0           0      4544014    124869457      7658948    185182989
rbd                        0            0            0            0           0            0            0            0            0
  total used       251477028        24039
  total avail    13723737372
  total space    13981283784
root@kvmhost01:~# ceph osd pool set libvirt-pool pg_num 256
set pool 1 pg_num to 256
root@kvmhost01:~# ceph osd pool set libvirt-pool pgp_num 256
Error EBUSY: currently creating pgs, wait
root@chbsskvmtst01:~# ceph osd pool get libvirt-pool pgp_num
pgp_num: 128



a few minutes later



root@kvmhost01:~# ceph osd pool set libvirt-pool pgp_num 256
set pool 1 pgp_num to 256
root@kvmhost01:~# ceph osd pool get libvirt-pool pgp_num
pgp_num: 256
root@kvmhost01:~# ceph osd pool get libvirt-pool pg_num
pg_num: 256
root@kvmhost01:~# rados df
pool name                 KB      objects       clones     degraded      unfound           rd        rd KB           wr        wr KB
.rgw                       1            4            0            0           0           90           70           30            8
.rgw.buckets         8126126         4573            0            0           0         5108         2664        35376      8127189
.rgw.buckets.index         0            2            0            0           0         9968         9962         9934            0
.rgw.control               0            8            0            0           0            0            0            0            0
.rgw.gc                    0           32            0            0           0         4928         4896         3576            0
.rgw.root                  1            3            0            0           0           65           43            3            3
.users                     1            2            0            0           0            0            0            2            2
.users.email               1            1            0            0           0            0            0            1            1
.users.uid                 1            4            0            0           0           70           68           60            2
cephfs_data          4628365         1346            0            0           0         4212       159015         4017      4663669
cephfs_metadata         8500           22            0            0           0           76        11168         4079        10581
libvirt-pool        70923657        18049            0           92           0      2016645     54587609      3461290     81148392
rbd                        0            0            0            0           0            0            0            0            0
  total used       258125320        24046
  total avail    13717090168
  total space    13981283784
root@kvmhost01:~# ceph health
HEALTH_WARN too many PGs per OSD (770 > max 300); mds cluster is degraded; mds cephmds02 is laggy
root@chbsskvmtst01:~# ceph mds stat
e33: 1/1/1 up {0=cephmds02=up:replay(laggy or crashed)}
root@chbsskvmtst01:~# virsh list --all
Id    Name                           State
----------------------------------------------------
3     cephmon01                      running
4     cephmon02                      running
11    cephosd01                      running
12    cephosd02                      running
13    cephosd03                      running
14    cephosd04                      running
16    cephosd06                      running
17    cephosd05                      running
20    cephadm01                      running
-     cephmds01                      shut off
 


after I started the mds VMs everything went back to normal (except too many PGs per OSD)

Let me know if you need more info
Bernhard

____________________________________________________________
FREE 3D EARTH SCREENSAVER - Watch the Earth right on your desktop!
Check it out at http://www.inbox.com/earth


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux