ceph df showing wrong MAX AVAIL for hybrid CRUSH Rule

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

 

I have the following configuration of OSDs:

 

ID CLASS WEIGHT  REWEIGHT SIZE  USE    AVAIL %USE  VAR  PGS

0   hdd 5.45599  1.00000 5587G  2259G 3327G 40.45 1.10 234

1   hdd 5.45599  1.00000 5587G  2295G 3291G 41.08 1.11 231

2   hdd 5.45599  1.00000 5587G  2321G 3265G 41.56 1.13 232

3   hdd 5.45599  1.00000 5587G  2313G 3273G 41.42 1.12 234

4   hdd 5.45599  1.00000 5587G  2105G 3481G 37.68 1.02 212

5   hdd 5.45599  1.00000 5587G  2231G 3355G 39.94 1.08 218

41   ssd 0.87299  1.00000  894G  9637M  884G  1.05 0.03  31

42   ssd 0.87299  1.00000  894G 13361M  881G  1.46 0.04  41

6   hdd 5.45599  1.00000 5587G  2404G 3182G 43.03 1.17 239

7   hdd 5.45599  1.00000 5587G  2226G 3360G 39.85 1.08 222

8   hdd 5.45599  1.00000 5587G  2668G 2918G 47.76 1.30 256

9   hdd 5.45599  1.00000 5587G  2366G 3220G 42.36 1.15 236

10   hdd 5.45599  1.00000 5587G  2454G 3132G 43.92 1.19 253

11   hdd 5.45599  1.00000 5587G  2405G 3181G 43.06 1.17 245

43   ssd 0.87299  1.00000  894G 15498M  879G  1.69 0.05  47

44   ssd 0.87299  1.00000  894G 10104M  884G  1.10 0.03  27

12   hdd 5.45599  1.00000 5587G  2242G 3344G 40.14 1.09 229

13   hdd 5.45599  1.00000 5587G  2551G 3035G 45.67 1.24 247

14   hdd 5.45599  1.00000 5587G  2513G 3074G 44.98 1.22 245

15   hdd 5.45599  1.00000 5587G  2014G 3572G 36.06 0.98 209

16   hdd 5.45599  1.00000 5587G  2586G 3000G 46.29 1.26 249

17   hdd 5.45599  1.00000 5587G  2459G 3127G 44.02 1.19 243

45   ssd 0.87299  1.00000  894G  9697M  884G  1.06 0.03  35

46   ssd 0.87299  1.00000  894G 12975M  881G  1.42 0.04  37

18   hdd 3.63699  1.00000 3724G  1595G 2128G 42.84 1.16 156

19   hdd 3.63699  1.00000 3724G  1387G 2336G 37.25 1.01 147

20   hdd 3.63699  1.00000 3724G  1551G 2172G 41.67 1.13 157

21   hdd 3.63699  1.00000 3724G  1535G 2189G 41.22 1.12 155

22   hdd 3.63699  1.00000 3724G  1459G 2264G 39.20 1.06 155

23   hdd 3.63699  1.00000 3724G  1395G 2329G 37.46 1.02 147

24   hdd 3.63699  1.00000 3724G  1489G 2234G 40.00 1.08 160

25   hdd 3.63699  1.00000 3724G  1634G 2090G 43.88 1.19 159

26   hdd 3.63699  1.00000 3724G  1566G 2157G 42.06 1.14 154

49   ssd 0.87299  1.00000  894G  9385M  884G  1.03 0.03  32

50   ssd 0.87299  1.00000  894G 12757M  881G  1.39 0.04  36

27   hdd 5.45599  1.00000 5587G  2462G 3124G 44.08 1.20 244

28   hdd 5.45599  1.00000 5587G  2314G 3272G 41.43 1.12 237

29   hdd 5.45599  1.00000 5587G  2166G 3420G 38.79 1.05 221

30   hdd 5.45599  1.00000 5587G  2484G 3102G 44.47 1.21 242

31   hdd 5.45599  1.00000 5587G  2292G 3294G 41.03 1.11 225

32   hdd 5.45599  1.00000 5587G  1982G 3604G 35.49 0.96 208

47   ssd 0.87299  1.00000  894G 12015M  882G  1.31 0.04  39

48   ssd 0.87299  1.00000  894G 14820M  879G  1.62 0.04  46

33   hdd 5.45599  1.00000 5587G  2002G 3584G 35.85 0.97 205

34   hdd 5.45599  1.00000 5587G  2069G 3517G 37.04 1.00 209

35   hdd 5.45599  1.00000 5587G  2187G 3399G 39.16 1.06 226

36   hdd 5.45599  1.00000 5587G  1821G 3765G 32.60 0.88 185

37   hdd 5.45599  1.00000 5587G  2123G 3463G 38.01 1.03 205

38   hdd 5.45599  1.00000 5587G  2197G 3390G 39.32 1.07 228

39   hdd 5.45599  1.00000 5587G  2180G 3406G 39.02 1.06 217

40   hdd 5.45599  1.00000 5587G  2232G 3354G 39.97 1.08 228

51   ssd 0.87320  1.00000  894G 14747M  879G  1.61 0.04  38

52   ssd 0.87320  1.00000  894G  7716M  886G  0.84 0.02  18

53   ssd 0.87320  1.00000  894G 12660M  881G  1.38 0.04  33

54   ssd 0.87320  1.00000  894G 11155M  883G  1.22 0.03  31

55   ssd 0.87320  1.00000  894G  9350M  885G  1.02 0.03  24

56   ssd 0.87320  1.00000  894G 13816M  880G  1.51 0.04  38

57   ssd 0.87320  1.00000  894G 10122M  884G  1.11 0.03  28

58   ssd 0.87320  1.00000  894G 10096M  884G  1.10 0.03  32

59   ssd 0.87320  1.00000  894G 13750M  880G  1.50 0.04  36

60   ssd 0.87320  1.00000  894G 16168M  878G  1.77 0.05  35

61   ssd 0.87320  1.00000  894G 11401M  883G  1.25 0.03  31

62   ssd 0.87320  1.00000  894G 12105M  882G  1.32 0.04  32

63   ssd 0.87320  1.00000  894G 13998M  880G  1.53 0.04  40

64   ssd 0.87320  1.00000  894G 17468M  877G  1.91 0.05  42

65   ssd 0.87320  1.00000  894G  9540M  884G  1.04 0.03  25

66   ssd 0.87320  1.00000  894G 15109M  879G  1.65 0.04  42

                    TOTAL  230T 86866G  145T 36.88

MIN/MAX VAR: 0.02/1.30  STDDEV: 22.48

 

With the following crush ruleset:

# begin crush map

tunable choose_local_tries 0

tunable choose_local_fallback_tries 0

tunable choose_total_tries 50

tunable chooseleaf_descend_once 1

tunable chooseleaf_vary_r 1

tunable chooseleaf_stable 1

tunable straw_calc_version 1

tunable allowed_bucket_algs 54

 

# devices

device 0 osd.0 class hdd

device 1 osd.1 class hdd

device 2 osd.2 class hdd

device 3 osd.3 class hdd

device 4 osd.4 class hdd

device 5 osd.5 class hdd

device 6 osd.6 class hdd

device 7 osd.7 class hdd

device 8 osd.8 class hdd

device 9 osd.9 class hdd

device 10 osd.10 class hdd

device 11 osd.11 class hdd

device 12 osd.12 class hdd

device 13 osd.13 class hdd

device 14 osd.14 class hdd

device 15 osd.15 class hdd

device 16 osd.16 class hdd

device 17 osd.17 class hdd

device 18 osd.18 class hdd

device 19 osd.19 class hdd

device 20 osd.20 class hdd

device 21 osd.21 class hdd

device 22 osd.22 class hdd

device 23 osd.23 class hdd

device 24 osd.24 class hdd

device 25 osd.25 class hdd

device 26 osd.26 class hdd

device 27 osd.27 class hdd

device 28 osd.28 class hdd

device 29 osd.29 class hdd

device 30 osd.30 class hdd

device 31 osd.31 class hdd

device 32 osd.32 class hdd

device 33 osd.33 class hdd

device 34 osd.34 class hdd

device 35 osd.35 class hdd

device 36 osd.36 class hdd

device 37 osd.37 class hdd

device 38 osd.38 class hdd

device 39 osd.39 class hdd

device 40 osd.40 class hdd

device 41 osd.41 class ssd

device 42 osd.42 class ssd

device 43 osd.43 class ssd

device 44 osd.44 class ssd

device 45 osd.45 class ssd

device 46 osd.46 class ssd

device 47 osd.47 class ssd

device 48 osd.48 class ssd

device 49 osd.49 class ssd

device 50 osd.50 class ssd

device 51 osd.51 class ssd

device 52 osd.52 class ssd

device 53 osd.53 class ssd

device 54 osd.54 class ssd

device 55 osd.55 class ssd

device 56 osd.56 class ssd

device 57 osd.57 class ssd

device 58 osd.58 class ssd

device 59 osd.59 class ssd

device 60 osd.60 class ssd

device 61 osd.61 class ssd

device 62 osd.62 class ssd

device 63 osd.63 class ssd

device 64 osd.64 class ssd

device 65 osd.65 class ssd

device 66 osd.66 class ssd

 

# types

type 0 osd

type 1 host

type 2 chassis

type 3 rack

type 4 row

type 5 pdu

type 6 pod

type 7 room

type 8 datacenter

type 9 region

type 10 root

 

# buckets

host node1 {

        id -2           # do not change unnecessarily

        id -8 class hdd         # do not change unnecessarily

        id -15 class ssd                # do not change unnecessarily

        # weight 34.482

        alg straw2

        hash 0  # rjenkins1

        item osd.0 weight 5.456

        item osd.1 weight 5.456

        item osd.2 weight 5.456

        item osd.3 weight 5.456

        item osd.4 weight 5.456

        item osd.5 weight 5.456

        item osd.41 weight 0.873

        item osd.42 weight 0.873

}

host node2 {

        id -3           # do not change unnecessarily

        id -9 class hdd         # do not change unnecessarily

        id -16 class ssd                # do not change unnecessarily

        # weight 34.482

        alg straw2

        hash 0  # rjenkins1

        item osd.6 weight 5.456

        item osd.7 weight 5.456

        item osd.8 weight 5.456

        item osd.9 weight 5.456

        item osd.10 weight 5.456

        item osd.11 weight 5.456

        item osd.43 weight 0.873

        item osd.44 weight 0.873

}

host node3 {

        id -4           # do not change unnecessarily

        id -10 class hdd                # do not change unnecessarily

        id -17 class ssd                # do not change unnecessarily

        # weight 34.482

        alg straw2

        hash 0  # rjenkins1

        item osd.12 weight 5.456

        item osd.13 weight 5.456

        item osd.14 weight 5.456

        item osd.15 weight 5.456

        item osd.16 weight 5.456

        item osd.17 weight 5.456

        item osd.45 weight 0.873

        item osd.46 weight 0.873

}

host node4 {

        id -5           # do not change unnecessarily

        id -11 class hdd                # do not change unnecessarily

        id -18 class ssd                # do not change unnecessarily

        # weight 34.479

        alg straw2

        hash 0  # rjenkins1

        item osd.18 weight 3.637

        item osd.21 weight 3.637

        item osd.24 weight 3.637

        item osd.25 weight 3.637

        item osd.22 weight 3.637

        item osd.26 weight 3.637

        item osd.19 weight 3.637

        item osd.23 weight 3.637

        item osd.20 weight 3.637

        item osd.49 weight 0.873

        item osd.50 weight 0.873

}

host node5 {

        id -6           # do not change unnecessarily

        id -12 class hdd                # do not change unnecessarily

        id -19 class ssd                # do not change unnecessarily

        # weight 34.482

        alg straw2

        hash 0  # rjenkins1

        item osd.27 weight 5.456

        item osd.28 weight 5.456

        item osd.29 weight 5.456

        item osd.30 weight 5.456

        item osd.31 weight 5.456

        item osd.32 weight 5.456

        item osd.47 weight 0.873

        item osd.48 weight 0.873

}

host node6 {

        id -7           # do not change unnecessarily

        id -13 class hdd                # do not change unnecessarily

        id -20 class ssd                # do not change unnecessarily

        # weight 43.648

        alg straw2

        hash 0  # rjenkins1

        item osd.33 weight 5.456

        item osd.34 weight 5.456

        item osd.35 weight 5.456

        item osd.36 weight 5.456

        item osd.37 weight 5.456

        item osd.38 weight 5.456

        item osd.39 weight 5.456

        item osd.40 weight 5.456

}

host node7 {

        id -22          # do not change unnecessarily

        id -23 class hdd                # do not change unnecessarily

        id -24 class ssd                # do not change unnecessarily

        # weight 6.986

        alg straw2

        hash 0  # rjenkins1

        item osd.51 weight 0.873

        item osd.52 weight 0.873

        item osd.53 weight 0.873

        item osd.54 weight 0.873

        item osd.55 weight 0.873

        item osd.56 weight 0.873

        item osd.57 weight 0.873

        item osd.58 weight 0.873

}

host node8 {

        id -25          # do not change unnecessarily

        id -26 class hdd                # do not change unnecessarily

        id -27 class ssd                # do not change unnecessarily

        # weight 6.986

        alg straw2

        hash 0  # rjenkins1

        item osd.59 weight 0.873

        item osd.60 weight 0.873

        item osd.61 weight 0.873

        item osd.62 weight 0.873

        item osd.63 weight 0.873

        item osd.64 weight 0.873

        item osd.65 weight 0.873

        item osd.66 weight 0.873

}

root default {

        id -1           # do not change unnecessarily

        id -14 class hdd                # do not change unnecessarily

        id -21 class ssd                # do not change unnecessarily

        # weight 230.027

        alg straw2

        hash 0  # rjenkins1

        item node1 weight 34.482

        item node2 weight 34.482

        item node3 weight 34.482

        item node4 weight 34.479

        item node5 weight 34.482

        item node6 weight 43.649

        item node7 weight 6.986

        item node8 weight 6.986

}

 

# rules

rule replicated_ruleset {

        id 0

        type replicated

        min_size 1

        max_size 10

        step take default class hdd

        step chooseleaf firstn 0 type host

        step emit

}

rule hybrid {

        id 1

        type replicated

        min_size 1

        max_size 10

        step take default class ssd

        step chooseleaf firstn 1 type host

        step emit

        step take default class hdd

        step chooseleaf firstn -1 type host

        step emit

}

rule flash {

        id 2

        type replicated

        min_size 1

        max_size 10

        step take default class ssd

        step chooseleaf firstn 0 type host

        step emit

}

 

# end crush map

 

And this is shown by ceph df, with the used CRUSH Rule added, all pools using a replication of 3:

GLOBAL:

    SIZE     AVAIL     RAW USED     %RAW USED

    230T      145T       86866G         36.88

POOLS:

    NAME                ID     USED       %USED     MAX AVAIL     OBJECTS

    pool1             7      10454G     23.82        33430G     2871159 -> replicated_ruleset

    pool2             8      13333G     28.51        33430G     3660994 -> replicated_ruleset

    rbd                 18      3729G     10.04        33430G     1021785 -> replicated_ruleset

    cephfs_data         24       270G      3.62         7213G       77568 -> hybrid

    cephfs_metadata     25     39781k         0         7213G          44 -> flash

 

I have 23244G in SSDs in total.

As you can see, the „replicated_ruleset“ chooses only HDDs, the „flash“ rule only SSDs and the „hybrid“ rule uses one SSD (as primary, since HDDs primary affinity is set to 0.5, while the SSDs is on 1.0) and the rest as HDDs.

 

ceph df shows MAX AVAIL for the „replicated_ruleset“ and „flash“ rule correctly for my pools with replication 3. However the „hybrid“ rule shows the available space as if it was using 3 SSDs, even though it’s only using one, even with a replication of 3.

 

Is this a display bug or is something wrong with my configuration? And if it’s a bug, will it break anything if MAX AVAIL or %USED fills up to over 100%?

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux