6
CEPH OSD Distribution:
ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-41 654.84045 root archive
-37 130.96848 host CEPH-ARCH-R03-07
100 archive 10.91399 osd.100 up 1.00000 1.00000
101 archive 10.91399 osd.101 up 1.00000 1.00000
102 archive 10.91399 osd.102 up 1.00000 1.00000
103 archive 10.91399 osd.103 up 1.00000 1.00000
104 archive 10.91399 osd.104 up 1.00000 1.00000
105 archive 10.91399 osd.105 up 1.00000 1.00000
106 archive 10.91409 osd.106 up 1.00000 1.00000
107 archive 10.91409 osd.107 up 1.00000 1.00000
108 archive 10.91409 osd.108 up 1.00000 1.00000
109 archive 10.91409 osd.109 up 1.00000 1.00000
110 archive 10.91409 osd.110 up 1.00000 1.00000
111 archive 10.91409 osd.111 up 1.00000 1.00000
-23 130.96800 host CEPH005
4 archive 10.91399 osd.4 up 1.00000 1.00000
41 archive 10.91399 osd.41 up 1.00000 1.00000
74 archive 10.91399 osd.74 up 1.00000 1.00000
75 archive 10.91399 osd.75 up 1.00000 1.00000
81 archive 10.91399 osd.81 up 1.00000 1.00000
82 archive 10.91399 osd.82 up 1.00000 1.00000
83 archive 10.91399 osd.83 up 1.00000 1.00000
84 archive 10.91399 osd.84 up 1.00000 1.00000
85 archive 10.91399 osd.85 up 1.00000 1.00000
86 archive 10.91399 osd.86 up 1.00000 1.00000
87 archive 10.91399 osd.87 up 1.00000 1.00000
88 archive 10.91399 osd.88 up 1.00000 1.00000
-17 130.96800 host CEPH006
7 archive 10.91399 osd.7 up 1.00000 1.00000
8 archive 10.91399 osd.8 up 1.00000 1.00000
9 archive 10.91399 osd.9 up 1.00000 1.00000
10 archive 10.91399 osd.10 up 1.00000 1.00000
12 archive 10.91399 osd.12 up 1.00000 1.00000
13 archive 10.91399 osd.13 up 1.00000 1.00000
42 archive 10.91399 osd.42 up 1.00000 1.00000
43 archive 10.91399 osd.43 up 1.00000 1.00000
51 archive 10.91399 osd.51 up 1.00000 1.00000
53 archive 10.91399 osd.53 up 1.00000 1.00000
76 archive 10.91399 osd.76 up 1.00000 1.00000
80 archive 10.91399 osd.80 up 1.00000 1.00000
-26 130.96800 host CEPH007
14 archive 10.91399 osd.14 up 1.00000 1.00000
15 archive 10.91399 osd.15 up 1.00000 1.00000
16 archive 10.91399 osd.16 up 1.00000 1.00000
39 archive 10.91399 osd.39 up 1.00000 1.00000
40 archive 10.91399 osd.40 up 1.00000 1.00000
44 archive 10.91399 osd.44 up 1.00000 1.00000
48 archive 10.91399 osd.48 up 1.00000 1.00000
49 archive 10.91399 osd.49 up 1.00000 1.00000
52 archive 10.91399 osd.52 up 1.00000 1.00000
77 archive 10.91399 osd.77 up 1.00000 1.00000
89 archive 10.91399 osd.89 up 1.00000 1.00000
90 archive 10.91399 osd.90 up 1.00000 1.00000
-31 130.96800 host CEPH008
5 archive 10.91399 osd.5 up 1.00000 1.00000
6 archive 10.91399 osd.6 up 1.00000 1.00000
11 archive 10.91399 osd.11 up 1.00000 1.00000
45 archive 10.91399 osd.45 up 1.00000 1.00000
46 archive 10.91399 osd.46 up 1.00000 1.00000
47 archive 10.91399 osd.47 up 1.00000 1.00000
55 archive 10.91399 osd.55 up 1.00000 1.00000
70 archive 10.91399 osd.70 up 1.00000 1.00000
71 archive 10.91399 osd.71 up 1.00000 1.00000
78 archive 10.91399 osd.78 up 1.00000 1.00000
79 archive 10.91399 osd.79 up 1.00000 1.00000
91 archive 10.91399 osd.91 up 1.00000 1.00000
-1 27.91296 root default
-30 6.98199 host CEPH-SSD-004
92 ssd 0.87299 osd.92 up 1.00000 1.00000
93 ssd 0.87299 osd.93 up 1.00000 1.00000
94 ssd 0.87299 osd.94 up 1.00000 1.00000
95 ssd 0.87299 osd.95 up 1.00000 1.00000
96 ssd 0.87299 osd.96 up 1.00000 1.00000
97 ssd 0.87299 osd.97 up 1.00000 1.00000
98 ssd 0.87299 osd.98 up 1.00000 1.00000
99 ssd 0.87299 osd.99 up 1.00000 1.00000
-3 6.97699 host CEPH001
1 ssd 0.43599 osd.1 up 1.00000 1.00000
17 ssd 0.43599 osd.17 up 1.00000 1.00000
18 ssd 0.43599 osd.18 up 1.00000 1.00000
19 ssd 0.43599 osd.19 up 1.00000 1.00000
20 ssd 0.43599 osd.20 up 1.00000 1.00000
21 ssd 0.43599 osd.21 up 1.00000 1.00000
22 ssd 0.43599 osd.22 up 1.00000 1.00000
23 ssd 0.43599 osd.23 up 1.00000 1.00000
37 ssd 0.87299 osd.37 up 1.00000 1.00000
54 ssd 0.87299 osd.54 up 1.00000 1.00000
56 ssd 0.43599 osd.56 up 1.00000 1.00000
60 ssd 0.43599 osd.60 up 1.00000 1.00000
61 ssd 0.43599 osd.61 up 1.00000 1.00000
62 ssd 0.43599 osd.62 up 1.00000 1.00000
-5 6.97699 host CEPH002
2 ssd 0.43599 osd.2 up 1.00000 1.00000
24 ssd 0 osd.24 up 1.00000 1.00000
25 ssd 0.43599 osd.25 up 1.00000 1.00000
26 ssd 0.43599 osd.26 up 1.00000 1.00000
27 ssd 0.43599 osd.27 up 1.00000 1.00000
28 ssd 0.43599 osd.28 up 1.00000 1.00000
29 ssd 0.43599 osd.29 up 1.00000 1.00000
30 ssd 0.43599 osd.30 up 1.00000 1.00000
38 ssd 0.87299 osd.38 up 1.00000 1.00000
57 ssd 0.43599 osd.57 up 1.00000 1.00000
63 ssd 0.43599 osd.63 up 1.00000 1.00000
64 ssd 0.43599 osd.64 up 1.00000 1.00000
65 ssd 0.43599 osd.65 up 1.00000 1.00000
66 ssd 0.43599 osd.66 up 1.00000 1.00000
72 ssd 0.87299 osd.72 up 1.00000 1.00000
-7 6.97699 host CEPH003
0 ssd 0.43599 osd.0 up 1.00000 1.00000
3 ssd 0.43599 osd.3 up 1.00000 1.00000
31 ssd 0 osd.31 up 1.00000 1.00000
32 ssd 0.43599 osd.32 up 1.00000 1.00000
33 ssd 0.43599 osd.33 up 1.00000 1.00000
34 ssd 0.43599 osd.34 up 1.00000 1.00000
35 ssd 0.43599 osd.35 up 1.00000 1.00000
36 ssd 0.43599 osd.36 up 1.00000 1.00000
50 ssd 0.87299 osd.50 up 1.00000 1.00000
58 ssd 0.43599 osd.58 up 1.00000 1.00000
59 ssd 0.43599 osd.59 up 1.00000 1.00000
67 ssd 0.43599 osd.67 up 1.00000 1.00000
68 ssd 0.43599 osd.68 up 1.00000 1.00000
69 ssd 0.43599 osd.69 up 1.00000 1.00000
73 ssd 0.87299 osd.73 up 1.00000 1.00000
CEPH CONF:
[global]
#Normal-Memory 1/5
debug rgw = 1
#Disable
debug osd = 0
debug journal = 0
debug ms = 0
fsid = e1ee8086-7cce-43fd-a252-3d677af22428
mon_initial_members = CEPH001, CEPH002, CEPH003
mon_host = 172.16.2.10,172.16.2.11,172.16.2.12
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
osd pool default pg num = 1024
osd pool default pgp num = 1024
public network = 172.16.2.0/24
cluster network = 172.16.1.0/24
osd pool default size = 2
osd pool default min size = 1
rgw dynamic resharding = true
[osd]
osd mkfs type = xfs
osd op threads = 12
osd disk threads = 12
osd recovery threads = 4
osd recovery op priority = 1
osd recovery max active = 2
osd recovery max single start = 1
osd max backfills = 4
osd backfill scan max = 16
osd backfill scan min = 4
osd client op priority = 63
osd memory target = 2147483648
osd scrub begin hour = 23
osd scrub end hour = 6
osd scrub load threshold = 1 #low load scrubbing
osd scrub during recovery = false #scrub during recovery
[mon]
mon allow pool delete = true
mon osd min down reporters = 3
[mon.a]
host = CEPH001
public bind addr = 172.16.2.10
mon addr = 172.16.2.10:6789
mon allow pool delete = true
[mon.b]
host = CEPH002
public bind addr = 172.16.2.11
mon addr = 172.16.2.11:6789
mon allow pool delete = true
[mon.c]
host = CEPH003
public bind addr = 172.16.2.12
mon addr = 172.16.2.12:6789
mon allow pool delete = true
[client.rgw.ceph-rgw01]
host = ceph-rgw01
rgw dns name = eu-es-s3gateway.edh-services.com
rgw frontends = "beast endpoint=172.16.2.6 port=8080"
rgw resolve cname = false
rgw thread pool size = 4096
rgw op thread timeout = 600
rgw num rados handles = 1
rgw num control oids = 8
rgw cache enabled = true
rgw cache lru size = 10000
rgw enable usage log = true
rgw usage log tick interval = 30
rgw usage log flush threshold = 1024
rgw usage max shards = 32
rgw usage max user shards = 1
rgw log http headers = "http_x_forwarded_for"
[client.rgw.ceph-rgw03]
host = ceph-rgw03
rgw dns name = XXXXXXXXXXXXXX
rgw frontends = "beast endpoint=172.16.2.8 port=8080"
rgw resolve cname = false
rgw thread pool size = 4096
rgw op thread timeout = 600
rgw num rados handles = 1
rgw num control oids = 8
rgw cache enabled = true
rgw cache lru size = 100000
rgw enable usage log = true
rgw usage log tick interval = 30
rgw usage log flush threshold = 1024
rgw usage max shards = 32
rgw usage max user shards = 1
rgw log http headers = "http_x_forwarded_for"
[OSD HOST CROP]
BUCKET:
radosgw-admin bucket stats --bucket=XXXX
{
"bucket": "XXXX",
"zonegroup": "4d8c7c5f-ca40-4ee3-b5bb-b2cad90bd007",
"placement_rule": "default-placement",
"explicit_placement": {
"data_pool": "default.rgw.buckets.data",
"data_extra_pool": "default.rgw.buckets.non-ec",
"index_pool": "default.rgw.buckets.index"
},
"id": "48efb8c3-693c-4fe0-bbe4-fdc16f590a82.14987264.1",
"marker": "48efb8c3-693c-4fe0-bbe4-fdc16f590a82.3856921.7",
"index_type": "Normal",
"owner": "XXXXX",
"ver":
"0#38096,1#37913,2#37954,3#37836,4#38134,5#38236,6#37911,7#37879,8#378
04,9#38005,10#37584,11#38446,12#37957,13#37537,14#37775,15#37794,16#38
399,17#38162,18#37834,19#37633,20#37403,21#38173,22#37651,23#37303,24#
37992,25#38228,26#38441,27#37975,28#38095,29#37835,30#38264,31#37958,3
2#37666,33#37517,34#38260,35#38168,36#37689,37#37551,38#37700,39#38056
,40#38175,41#37765,42#37721,43#38472,44#37928,45#38451,46#37491,47#378
75,48#38405,49#38011,50#38025,51#37983,52#37940,53#38306,54#37908,55#3
8181,56#37721,57#38366,58#37834,59#38392,60#37928,61#38235,62#37837,63
#37940,64#38294,65#37610,66#37974,67#38304,68#37725,69#37301,70#37155,
71#37681,72#37463,73#37603,74#37323,75#37717,76#37111,77#37688,78#3747
3,79#37052,80#37413,81#37758,82#36971,83#37327,84#37056,85#37302,86#37
492,87#37250,88#37708,89#36891,90#38019,91#37485,92#37335,93#37712,94#
37754,95#37117,96#37085,97#37694,98#37386,99#37384,100#37668,101#37329
,102#37177,103#37494,104#37296,105#37366,106#37853,107#37234,108#36945
,109#37040,110#37389,111#37973,112#37092,113#37327,114#37505,115#37545
,116#37884,117#37325,118#37401,119#37208,120#37277,121#37087,122#37664
,123#37649,124#37517,125#37618,126#37145,127#37300",
"master_ver":
"0#0,1#0,2#0,3#0,4#0,5#0,6#0,7#0,8#0,9#0,10#0,11#0,12#0,13#0,14#0,15#0
,16#0,17#0,18#0,19#0,20#0,21#0,22#0,23#0,24#0,25#0,26#0,27#0,28#0,29#0
,30#0,31#0,32#0,33#0,34#0,35#0,36#0,37#0,38#0,39#0,40#0,41#0,42#0,43#0
,44#0,45#0,46#0,47#0,48#0,49#0,50#0,51#0,52#0,53#0,54#0,55#0,56#0,57#0
,58#0,59#0,60#0,61#0,62#0,63#0,64#0,65#0,66#0,67#0,68#0,69#0,70#0,71#0
,72#0,73#0,74#0,75#0,76#0,77#0,78#0,79#0,80#0,81#0,82#0,83#0,84#0,85#0
,86#0,87#0,88#0,89#0,90#0,91#0,92#0,93#0,94#0,95#0,96#0,97#0,98#0,99#0
,100#0,101#0,102#0,103#0,104#0,105#0,106#0,107#0,108#0,109#0,110#0,111
#0,112#0,113#0,114#0,115#0,116#0,117#0,118#0,119#0,120#0,121#0,122#0,1
23#0,124#0,125#0,126#0,127#0",
"mtime": "2019-04-30 15:14:20.152747",
"max_marker":
"0#,1#,2#,3#,4#,5#,6#,7#,8#,9#,10#,11#,12#,13#,14#,15#,16#,17#,18#,19#
,20#,21#,22#,23#,24#,25#,26#,27#,28#,29#,30#,31#,32#,33#,34#,35#,36#,3
7#,38#,39#,40#,41#,42#,43#,44#,45#,46#,47#,48#,49#,50#,51#,52#,53#,54#
,55#,56#,57#,58#,59#,60#,61#,62#,63#,64#,65#,66#,67#,68#,69#,70#,71#,7
2#,73#,74#,75#,76#,77#,78#,79#,80#,81#,82#,83#,84#,85#,86#,87#,88#,89#
,90#,91#,92#,93#,94#,95#,96#,97#,98#,99#,100#,101#,102#,103#,104#,105#
,106#,107#,108#,109#,110#,111#,112#,113#,114#,115#,116#,117#,118#,119#
,120#,121#,122#,123#,124#,125#,126#,127#",
"usage": {
"rgw.none": {
"size": 0,
"size_actual": 0,
"size_utilized": 0,
"size_kb": 0,
"size_kb_actual": 0,
"size_kb_utilized": 0,
"num_objects": 0
},
"rgw.main": {
"size": 95368837327307,
"size_actual": 95379421843456,
"size_utilized": 95370098064318,
"size_kb": 93133630203,
"size_kb_actual": 93143966644,
"size_kb_utilized": 93134861391,
"num_objects": 5872260
},
"rgw.multimeta": {
"size": 0,
"size_actual": 0,
"size_utilized": 0,
"size_kb": 0,
"size_kb_actual": 0,
"size_kb_utilized": 0,
"num_objects": 467
}
},
"bucket_quota": {
"enabled": false,
"check_on_raw": false,
"max_size": -1024,
"max_size_kb": 0,
"max_objects": -1
}
}
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com