Ceph pool size and OSD data distribution

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I have a 7 node cluster which is complaining that:
root@s1:~# ceph -s

  cluster:
    id:     a6092407-216f-41ff-bccb-9bed78587ac3
    health: HEALTH_WARN
            1 nearfull osd(s)
            4 pool(s) nearfull

  services:
    mon: 3 daemons, quorum sm1,2,s5
    mgr: s1(active), standbys: s5, sm1
    mds: cephfs-1/1/1 up  {0=s1=up:active}, 2 up:standby
    osd: 23 osds: 23 up, 23 in

  data:
    pools:   4 pools, 1312 pgs
    objects: 1.26M objects, 4.64TiB
    usage:   11.4TiB used, 8.48TiB / 19.8TiB avail
    pgs:     1312 active+clean

  io:
    client:   542KiB/s wr, 0op/s rd, 32op/s wr


I see that the distribution of data over the OSD is very uneven.
Particularly on host S1 there are 6 SAS 300GB SAS drives that are identical
in spec, yet one is more than 89% in use, while another is on jus tover 40%
in use. What causes this?

root@s1:~# ceph osd df tree

ID  CLASS WEIGHT   REWEIGHT SIZE    USE     DATA    OMAP    META
AVAIL   %USE  VAR  PGS TYPE NAME
 -1       19.82628        - 19.8TiB 11.4TiB 11.3TiB 2.18GiB 32.9GiB
8.48TiB 57.25 1.00   - root default
 -2        6.36676        - 6.37TiB 3.04TiB 3.03TiB  653MiB 11.3GiB
3.33TiB 47.76 0.83   -     host hp1
  3   hdd  0.90959  1.00000  931GiB  422GiB  420GiB 84.6MiB 2.21GiB
509GiB 45.35 0.79 143         osd.3
  4   hdd  0.68210  1.00000  699GiB  265GiB  264GiB 66.7MiB  957MiB
433GiB 37.95 0.66  94         osd.4
  6   hdd  0.68210  1.00000  699GiB  308GiB  307GiB 64.7MiB  988MiB
390GiB 44.15 0.77  99         osd.6
  7   hdd  0.68210  1.00000  699GiB  346GiB  345GiB 74.4MiB  988MiB
353GiB 49.51 0.86 109         osd.7
 16   hdd  0.90959  1.00000  931GiB  461GiB  460GiB  103MiB 1.13GiB
470GiB 49.51 0.86 145         osd.16
 19   hdd  0.90959  1.00000  931GiB  516GiB  514GiB 96.2MiB 2.06GiB
415GiB 55.40 0.97 140         osd.19
 22   hdd  0.68210  1.00000  699GiB  290GiB  288GiB 68.9MiB 1.91GiB
408GiB 41.55 0.73  98         osd.22
 24   hdd  0.90959  1.00000  931GiB  505GiB  504GiB 94.8MiB 1.17GiB
426GiB 54.21 0.95 150         osd.24
 -3        1.63440        - 1.63TiB 1.07TiB 1.06TiB  236MiB 5.77GiB
582GiB 65.22 1.14   -     host s1
 10   hdd  0.27240  1.00000  279GiB  152GiB  151GiB 19.9MiB 1004MiB
127GiB 54.35 0.95  44         osd.10
 11   hdd  0.27240  1.00000  279GiB  114GiB  113GiB 43.3MiB  981MiB
165GiB 40.91 0.71  63         osd.11
 12   hdd  0.27240  1.00000  279GiB  180GiB  179GiB 41.4MiB  983MiB
98.6GiB 64.66 1.13  58         osd.12
 13   hdd  0.27240  1.00000  279GiB  190GiB  189GiB 33.8MiB  990MiB
89.4GiB 67.96 1.19  52         osd.13
 14   hdd  0.27240  1.00000  279GiB  249GiB  248GiB 48.6MiB  975MiB
30.0GiB 89.26 1.56  67         osd.14
 15   hdd  0.27240  1.00000  279GiB  207GiB  206GiB 49.2MiB  975MiB
72.0GiB 74.17 1.30  60         osd.15
 -4        2.72888        - 2.73TiB 1.71TiB 1.70TiB  279MiB 4.47GiB
1.02TiB 62.64 1.09   -     host s2
  9   hdd  1.81929  1.00000 1.82TiB 1.15TiB 1.15TiB  196MiB 2.35GiB
685GiB 63.21 1.10 390         osd.9
 17   hdd  0.90959  1.00000  931GiB  573GiB  571GiB 83.3MiB 2.12GiB
359GiB 61.50 1.07 181         osd.17
 -6        1.81929        - 1.82TiB 1.24TiB 1.24TiB  203MiB 2.34GiB
594GiB 68.12 1.19   -     host s4
 18   hdd  1.81929  1.00000 1.82TiB 1.24TiB 1.24TiB  203MiB 2.34GiB
594GiB 68.12 1.19 407         osd.18
 -7        2.72888        - 2.73TiB 1.73TiB 1.72TiB  341MiB 3.48GiB
1.00TiB 63.25 1.10   -     host s5
  2   hdd  1.81929  1.00000 1.82TiB 1.09TiB 1.09TiB  203MiB 2.06GiB
747GiB 59.89 1.05 368         osd.2
 20   hdd  0.90959  1.00000  931GiB  652GiB  650GiB  138MiB 1.42GiB
280GiB 69.96 1.22 215         osd.20
-15        2.72888        - 2.73TiB 1.41TiB 1.41TiB  307MiB 2.98GiB
1.32TiB 51.76 0.90   -     host s6
  0   hdd  1.81929  1.00000 1.82TiB  923GiB  921GiB  182MiB 1.81GiB
940GiB 49.56 0.87 358         osd.0
  1   hdd  0.90959  1.00000  931GiB  523GiB  522GiB  125MiB 1.18GiB
408GiB 56.18 0.98 187         osd.1
 -5        1.81918        - 1.82TiB 1.16TiB 1.15TiB  211MiB 2.56GiB
679GiB 63.56 1.11   -     host sm1
  5   hdd  0.90959  1.00000  931GiB  558GiB  557GiB  116MiB 1.23GiB
373GiB 59.94 1.05 182         osd.5
  8   hdd  0.90959  1.00000  931GiB  626GiB  624GiB 95.5MiB 1.33GiB
306GiB 67.18 1.17 198         osd.8
                      TOTAL 19.8TiB 11.4TiB 11.3TiB 2.18GiB 32.9GiB
8.48TiB 57.25
MIN/MAX VAR: 0.66/1.56  STDDEV: 12.03


How does this work?

thanks

Roland
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux