Crush not deliverying data uniformly -> HEALTH_ERR full osd

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Ceph community

We are at the stage of performance capacity testing, where significant amounts of backup data is being written to Ceph.

The issue we have, is that the underlying HDD's are not being populated (roughly) uniformly, and our Ceph system hits a brick wall after a couple of days our 30TB storage system is no longer able to operate after having only stored ~7TB.

Basically, despite HDD's (1:1 ratio between OSD and HDD) all being the same storage size and weighting in the Crushmap, we have disks either:
a) using 1% space;
b) using 48%; or
c) using 96%
Too precise a split to be an accident.  See below for more detail (osd11-22 not expected to get data, per our crushmap):


ceph pg dump
<snip>
pool 0  2442    0       0       0       10240000000     7302520 7302520
pool 1  57      0       0       0       127824767       5603518 5603518
pool 2  0       0       0       0       0       0       0
pool 3  1808757 0       0       0       7584377697985   1104048 1104048
 sum    1811256 0       0       0       7594745522752   14010086        14010086
osdstat kbused  kbavail kb      hb in   hb out
0       930606904       1021178408      1953514584      [11,12,13,14,15,16,17,18,19,20,21,22]   []
1       1874428 1949525164      1953514584      [11,12,13,14,15,16,17,18,19,20,21,22]   []
2       928811428       1022963676      1953514584      [11,12,13,14,15,16,17,18,19,20,21,22]   []
3       929733676       1022051996      1953514584      [11,12,13,14,15,16,17,18,19,20,21,22]   []
4       1719124 1949678844      1953514584      [11,12,13,14,15,16,17,18,19,20,21,22]   []
5       1853452 1949545892      1953514584      [11,12,13,14,15,16,17,18,19,20,21,22]   []
6       930979476       1020807132      1953514584      [11,12,13,14,15,16,17,18,19,20,21,22]   []
7       1808968 1949590496      1953514584      [11,12,13,14,15,16,17,18,19,20,21,22]   []
8       934035924       1017759100      1953514584      [11,12,13,14,15,16,17,18,19,20,21,22]   []
9       1855955384      94927432        1953514584      [11,12,13,14,15,16,17,18,19,20,21,22]   []
10      933572004       1018232340      1953514584      [11,12,13,14,15,16,17,18,19,20,21,22]   []
11      2057096 953060760       957230808       [0,1,2,3,4,5,6,7,8,9,10,17,18,19,20,21,22]      []
12      2053512 953064656       957230808       [0,1,2,3,4,5,6,7,8,9,10,17,18,19,20,21,22]      []
13      2148732 972501316       976762584       [0,1,2,3,4,5,6,7,8,9,10,17,18,19,20,21,22]      []
14      2064640 972585104       976762584       [0,1,2,3,4,5,6,7,8,9,10,17,18,19,20,21,22]      []
15      1945388 972703468       976762584       [0,1,2,3,4,5,6,7,8,9,10,17,18,19,20,21] []
16      2051708 972599412       976762584       [0,1,2,3,4,6,7,8,9,10,17,18,19,20,21]   []
17      2137632 952980216       957230808       [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16]      []
18      2000124 953117508       957230808       [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16]      []
19      2095124 972554492       976762584       [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16]      []
20      1986800 972662640       976762584       [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16]      []
21      2035204 972615332       976762584       [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16]      []
22      1961412 972687788       976762584       [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16]      []
 sum    7475488140      25609393172     33131684328

2012-08-06 10:03:58.964716 7f06783bb700  0 -- 10.32.0.10:0/15147 send_keepalive con 0x223f690, no pipe.


root@dsanb1-coy:~# df -h
Filesystem                               Size  Used Avail Use% Mounted on
/dev/md0                                 462G   12G  446G   3% /
udev                                      12G  4.0K   12G   1% /dev
tmpfs                                    4.8G  448K  4.8G   1% /run
none                                     5.0M     0  5.0M   0% /run/lock
none                                      12G     0   12G   0% /run/shm
/dev/sdc                                 1.9T  888G  974G  48% /ceph-data/osd.0
/dev/sdd                                 1.9T  1.8G  1.9T   1% /ceph-data/osd.1
/dev/sdp                                 1.9T  891G  972G  48% /ceph-data/osd.10
/dev/sde                                 1.9T  886G  976G  48% /ceph-data/osd.2
/dev/sdf                                 1.9T  887G  975G  48% /ceph-data/osd.3
/dev/sdg                                 1.9T  1.7G  1.9T   1% /ceph-data/osd.4
/dev/sdh                                 1.9T  1.8G  1.9T   1% /ceph-data/osd.5
/dev/sdi                                 1.9T  888G  974G  48% /ceph-data/osd.6
/dev/sdm                                 1.9T  1.8G  1.9T   1% /ceph-data/osd.7
/dev/sdn                                 1.9T  891G  971G  48% /ceph-data/osd.8
/dev/sdo                                 1.9T  1.8T   91G  96% /ceph-data/osd.9
10.32.0.10,10.32.0.25,10.32.0.11:6789:/   31T  7.1T   24T  23% /mnt/ceph


We are writing via fstab based cephfs mounts, and the above is going to pool3, which is a "backup" pool where we are testing replication level of 1x only. This should not have any effect though? Below will illustrate the layout we are using (above data writing issue is only going to the first node per our testing design):

root@dsanb1-coy:~# ceph osd tree
dumped osdmap tree epoch 136
# id    weight  type name       up/down reweight
-7      23      zone bak
-6      23              rack 1nrack
-2      11                      host dsanb1-coy
0       2                               osd.0   up      1
1       2                               osd.1   up      1
10      2                               osd.10  up      1
2       2                               osd.2   up      1
3       2                               osd.3   up      1
4       2                               osd.4   up      1
5       2                               osd.5   up      1
6       2                               osd.6   up      1
7       2                               osd.7   up      1
8       2                               osd.8   up      1
9       2                               osd.9   up      1
-1      23      zone default
-3      23              rack 2nrack
-2      11                      host dsanb1-coy
0       2                               osd.0   up      1
1       2                               osd.1   up      1
10      2                               osd.10  up      1
2       2                               osd.2   up      1
3       2                               osd.3   up      1
4       2                               osd.4   up      1
5       2                               osd.5   up      1
6       2                               osd.6   up      1
7       2                               osd.7   up      1
8       2                               osd.8   up      1
9       2                               osd.9   up      1
-4      6                       host dsanb2-coy
11      1                               osd.11  up      1
12      1                               osd.12  up      1
13      1                               osd.13  up      1
14      1                               osd.14  up      1
15      1                               osd.15  up      1
16      1                               osd.16  up      1
-5      6                       host dsanb3-coy
17      1                               osd.17  up      1
18      1                               osd.18  up      1
19      1                               osd.19  up      1
20      1                               osd.20  up      1
21      1                               osd.21  up      1
22      1                               osd.22  up      1


Has anybody got any suggestions?

Many thanks everybody..........
Paul


--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux