On Wed, Feb 1, 2017 at 3:38 AM, Gregory Farnum <gfarnum@xxxxxxxxxx> wrote: > On Tue, Jan 31, 2017 at 9:06 AM, Muthusamy Muthiah > <muthiah.muthusamy@xxxxxxxxx> wrote: >> Hi Greg, >> >> the problem is in kraken, when a pool is created with EC profile , min_size >> equals erasure size. >> >> For 3+1 profile , following is the pool status , >> pool 2 'cdvr_ec' erasure size 4 min_size 4 crush_ruleset 1 object_hash >> rjenkins pg_num 1024 pgp_num 1024 last_change 234 flags hashpspool >> stripe_width 4128 >> >> For 4+1 profile: >> pool 5 'cdvr_ec' erasure size 5 min_size 5 crush_ruleset 1 object_hash >> rjenkins pg_num 4096 pgp_num 4096 >> >> For 3+2 profile : >> pool 3 'cdvr_ec' erasure size 5 min_size 4 crush_ruleset 1 object_hash >> rjenkins pg_num 1024 pgp_num 1024 last_change 412 flags hashpspool >> stripe_width 4128 >> >> Where as on Jewel release for EC 4+1: >> pool 30 'cdvr_ec' erasure size 5 min_size 4 crush_ruleset 1 object_hash >> rjenkins pg_num 4096 pgp_num 4096 >> >> Trying to modify min_size and verify the status. >> >> Is there any reason behind this change in ceph kraken or a bug. > > The change was made on purpose because running with k replicas on a > k+m pool is a bad idea. However, it definitely should have recovered > the missing shard and then gone active, which doesn't appear to have Yeah, that might be true. > happened in this case. > > It looks like we just screwed up and don't let EC pools do recovery on > min size. You can restore the old behavior by setting min_size equal > to k and we'll be fixing this for the next release. (In general, k+1 If it's true, we should stop users to set up this ratio of data / coding chunks. Or there should be any reasonable warning. Should it be feature request? > pools are not a good idea, which is why we didn't catch this in > testing.) > -Greg > >> >> Thanks, >> Muthu >> >> >> >> >> On 31 January 2017 at 18:17, Muthusamy Muthiah <muthiah.muthusamy@xxxxxxxxx> >> wrote: >>> >>> Hi Greg, >>> >>> Following are the test outcomes on EC profile ( n = k + m) >>> >>> >>> >>> 1. Kraken filestore and bluetore with m=1 , recovery does not start >>> . >>> >>> 2. Jewel filestore and bluestore with m=1 , recovery happens . >>> >>> 3. Kraken bluestore all default configuration and m=1, no recovery. >>> >>> 4. Kraken bluestore with m=2 , recovery happens when one OSD is down >>> and for 2 OSD fails. >>> >>> >>> >>> So, the issue seems to be on ceph-kraken release. Your views… >>> >>> >>> >>> Thanks, >>> >>> Muthu >>> >>> >>> >>> >>> On 31 January 2017 at 14:18, Muthusamy Muthiah >>> <muthiah.muthusamy@xxxxxxxxx> wrote: >>>> >>>> Hi Greg, >>>> >>>> Now we could see the same problem exists for kraken-filestore also. >>>> Attached the requested osdmap and crushmap. >>>> >>>> OSD.1 was stopped in this following procedure and OSD map for a PG is >>>> displayed. >>>> >>>> ceph osd dump | grep cdvr_ec >>>> 2017-01-31 08:39:44.827079 7f323d66c700 -1 WARNING: the following >>>> dangerous and experimental features are enabled: bluestore,rocksdb >>>> 2017-01-31 08:39:44.848901 7f323d66c700 -1 WARNING: the following >>>> dangerous and experimental features are enabled: bluestore,rocksdb >>>> pool 2 'cdvr_ec' erasure size 4 min_size 4 crush_ruleset 1 object_hash >>>> rjenkins pg_num 1024 pgp_num 1024 last_change 234 flags hashpspool >>>> stripe_width 4128 >>>> >>>> [root@ca-cn2 ~]# ceph osd getmap -o /tmp/osdmap >>>> >>>> >>>> [root@ca-cn2 ~]# osdmaptool --pool 2 --test-map-object object1 >>>> /tmp/osdmap >>>> osdmaptool: osdmap file '/tmp/osdmap' >>>> object 'object1' -> 2.2bc -> [20,47,1,36] >>>> >>>> [root@ca-cn2 ~]# ceph osd map cdvr_ec object1 >>>> osdmap e402 pool 'cdvr_ec' (2) object 'object1' -> pg 2.bac5debc (2.2bc) >>>> -> up ([20,47,1,36], p20) acting ([20,47,1,36], p20) >>>> >>>> [root@ca-cn2 ~]# systemctl stop ceph-osd@1.service >>>> >>>> [root@ca-cn2 ~]# ceph osd getmap -o /tmp/osdmap1 >>>> >>>> >>>> [root@ca-cn2 ~]# osdmaptool --pool 2 --test-map-object object1 >>>> /tmp/osdmap1 >>>> osdmaptool: osdmap file '/tmp/osdmap1' >>>> object 'object1' -> 2.2bc -> [20,47,2147483647,36] >>>> >>>> >>>> [root@ca-cn2 ~]# ceph osd map cdvr_ec object1 >>>> osdmap e406 pool 'cdvr_ec' (2) object 'object1' -> pg 2.bac5debc (2.2bc) >>>> -> up ([20,47,39,36], p20) acting ([20,47,NONE,36], p20) >>>> >>>> >>>> [root@ca-cn2 ~]# ceph osd tree >>>> 2017-01-31 08:42:19.606876 7f4ed856a700 -1 WARNING: the following >>>> dangerous and experimental features are enabled: bluestore,rocksdb >>>> 2017-01-31 08:42:19.628358 7f4ed856a700 -1 WARNING: the following >>>> dangerous and experimental features are enabled: bluestore,rocksdb >>>> ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY >>>> -1 327.47314 root default >>>> -2 65.49463 host ca-cn4 >>>> 3 5.45789 osd.3 up 1.00000 1.00000 >>>> 5 5.45789 osd.5 up 1.00000 1.00000 >>>> 10 5.45789 osd.10 up 1.00000 1.00000 >>>> 16 5.45789 osd.16 up 1.00000 1.00000 >>>> 21 5.45789 osd.21 up 1.00000 1.00000 >>>> 27 5.45789 osd.27 up 1.00000 1.00000 >>>> 30 5.45789 osd.30 up 1.00000 1.00000 >>>> 35 5.45789 osd.35 up 1.00000 1.00000 >>>> 42 5.45789 osd.42 up 1.00000 1.00000 >>>> 47 5.45789 osd.47 up 1.00000 1.00000 >>>> 51 5.45789 osd.51 up 1.00000 1.00000 >>>> 53 5.45789 osd.53 up 1.00000 1.00000 >>>> -3 65.49463 host ca-cn3 >>>> 2 5.45789 osd.2 up 1.00000 1.00000 >>>> 6 5.45789 osd.6 up 1.00000 1.00000 >>>> 11 5.45789 osd.11 up 1.00000 1.00000 >>>> 15 5.45789 osd.15 up 1.00000 1.00000 >>>> 20 5.45789 osd.20 up 1.00000 1.00000 >>>> 25 5.45789 osd.25 up 1.00000 1.00000 >>>> 29 5.45789 osd.29 up 1.00000 1.00000 >>>> 33 5.45789 osd.33 up 1.00000 1.00000 >>>> 38 5.45789 osd.38 up 1.00000 1.00000 >>>> 40 5.45789 osd.40 up 1.00000 1.00000 >>>> 45 5.45789 osd.45 up 1.00000 1.00000 >>>> 49 5.45789 osd.49 up 1.00000 1.00000 >>>> -4 65.49463 host ca-cn5 >>>> 0 5.45789 osd.0 up 1.00000 1.00000 >>>> 7 5.45789 osd.7 up 1.00000 1.00000 >>>> 12 5.45789 osd.12 up 1.00000 1.00000 >>>> 17 5.45789 osd.17 up 1.00000 1.00000 >>>> 23 5.45789 osd.23 up 1.00000 1.00000 >>>> 26 5.45789 osd.26 up 1.00000 1.00000 >>>> 32 5.45789 osd.32 up 1.00000 1.00000 >>>> 34 5.45789 osd.34 up 1.00000 1.00000 >>>> 41 5.45789 osd.41 up 1.00000 1.00000 >>>> 46 5.45789 osd.46 up 1.00000 1.00000 >>>> 52 5.45789 osd.52 up 1.00000 1.00000 >>>> 56 5.45789 osd.56 up 1.00000 1.00000 >>>> -5 65.49463 host ca-cn1 >>>> 4 5.45789 osd.4 up 1.00000 1.00000 >>>> 9 5.45789 osd.9 up 1.00000 1.00000 >>>> 14 5.45789 osd.14 up 1.00000 1.00000 >>>> 19 5.45789 osd.19 up 1.00000 1.00000 >>>> 24 5.45789 osd.24 up 1.00000 1.00000 >>>> 36 5.45789 osd.36 up 1.00000 1.00000 >>>> 43 5.45789 osd.43 up 1.00000 1.00000 >>>> 50 5.45789 osd.50 up 1.00000 1.00000 >>>> 55 5.45789 osd.55 up 1.00000 1.00000 >>>> 57 5.45789 osd.57 up 1.00000 1.00000 >>>> 58 5.45789 osd.58 up 1.00000 1.00000 >>>> 59 5.45789 osd.59 up 1.00000 1.00000 >>>> -6 65.49463 host ca-cn2 >>>> 1 5.45789 osd.1 down 0 1.00000 >>>> 8 5.45789 osd.8 up 1.00000 1.00000 >>>> 13 5.45789 osd.13 up 1.00000 1.00000 >>>> 18 5.45789 osd.18 up 1.00000 1.00000 >>>> 22 5.45789 osd.22 up 1.00000 1.00000 >>>> 28 5.45789 osd.28 up 1.00000 1.00000 >>>> 31 5.45789 osd.31 up 1.00000 1.00000 >>>> 37 5.45789 osd.37 up 1.00000 1.00000 >>>> 39 5.45789 osd.39 up 1.00000 1.00000 >>>> 44 5.45789 osd.44 up 1.00000 1.00000 >>>> 48 5.45789 osd.48 up 1.00000 1.00000 >>>> 54 5.45789 osd.54 up 1.00000 1.00000 >>>> >>>> health HEALTH_ERR >>>> 69 pgs are stuck inactive for more than 300 seconds >>>> 69 pgs incomplete >>>> 69 pgs stuck inactive >>>> 69 pgs stuck unclean >>>> 512 requests are blocked > 32 sec >>>> monmap e2: 5 mons at >>>> {ca-cn1=10.50.5.117:6789/0,ca-cn2=10.50.5.118:6789/0,ca-cn3=10.50.5.119:6789/0,ca-cn4=10.50.5.120:6789/0,ca-cn5=10.50.5.121:6789/0} >>>> election epoch 8, quorum 0,1,2,3,4 >>>> ca-cn1,ca-cn2,ca-cn3,ca-cn4,ca-cn5 >>>> mgr active: ca-cn4 standbys: ca-cn2, ca-cn5, ca-cn3, ca-cn1 >>>> osdmap e406: 60 osds: 59 up, 59 in; 69 remapped pgs >>>> flags sortbitwise,require_jewel_osds,require_kraken_osds >>>> pgmap v23018: 1024 pgs, 1 pools, 3892 GB data, 7910 kobjects >>>> 6074 GB used, 316 TB / 322 TB avail >>>> 955 active+clean >>>> 69 remapped+incomplete >>>> >>>> Thanks, >>>> Muthu >>>> >>>> >>>> On 31 January 2017 at 02:54, Gregory Farnum <gfarnum@xxxxxxxxxx> wrote: >>>>> >>>>> You might also check out "ceph osd tree" and crush dump and make sure >>>>> they look the way you expect. >>>>> >>>>> On Mon, Jan 30, 2017 at 1:23 PM, Gregory Farnum <gfarnum@xxxxxxxxxx> >>>>> wrote: >>>>> > On Sun, Jan 29, 2017 at 6:40 AM, Muthusamy Muthiah >>>>> > <muthiah.muthusamy@xxxxxxxxx> wrote: >>>>> >> Hi All, >>>>> >> >>>>> >> Also tried EC profile 3+1 on 5 node cluster with bluestore enabled . >>>>> >> When >>>>> >> an OSD is down the cluster goes to ERROR state even when the cluster >>>>> >> is n+1 >>>>> >> . No recovery happening. >>>>> >> >>>>> >> health HEALTH_ERR >>>>> >> 75 pgs are stuck inactive for more than 300 seconds >>>>> >> 75 pgs incomplete >>>>> >> 75 pgs stuck inactive >>>>> >> 75 pgs stuck unclean >>>>> >> monmap e2: 5 mons at >>>>> >> >>>>> >> {ca-cn1=10.50.5.117:6789/0,ca-cn2=10.50.5.118:6789/0,ca-cn3=10.50.5.119:6789/0,ca-cn4=10.50.5.120:6789/0,ca-cn5=10.50.5.121:6789/0} >>>>> >> election epoch 10, quorum 0,1,2,3,4 >>>>> >> ca-cn1,ca-cn2,ca-cn3,ca-cn4,ca-cn5 >>>>> >> mgr active: ca-cn1 standbys: ca-cn4, ca-cn3, ca-cn5, ca-cn2 >>>>> >> osdmap e264: 60 osds: 59 up, 59 in; 75 remapped pgs >>>>> >> flags sortbitwise,require_jewel_osds,require_kraken_osds >>>>> >> pgmap v119402: 1024 pgs, 1 pools, 28519 GB data, 21548 kobjects >>>>> >> 39976 GB used, 282 TB / 322 TB avail >>>>> >> 941 active+clean >>>>> >> 75 remapped+incomplete >>>>> >> 8 active+clean+scrubbing >>>>> >> >>>>> >> this seems to be an issue with bluestore , recovery not happening >>>>> >> properly >>>>> >> with EC . >>>>> > >>>>> > It's possible but it seems a lot more likely this is some kind of >>>>> > config issue. Can you share your osd map ("ceph osd getmap")? >>>>> > -Greg >>>>> > >>>>> >> >>>>> >> Thanks, >>>>> >> Muthu >>>>> >> >>>>> >> On 24 January 2017 at 12:57, Muthusamy Muthiah >>>>> >> <muthiah.muthusamy@xxxxxxxxx> >>>>> >> wrote: >>>>> >>> >>>>> >>> Hi Greg, >>>>> >>> >>>>> >>> We use EC:4+1 on 5 node cluster in production deployments with >>>>> >>> filestore >>>>> >>> and it does recovery and peering when one OSD goes down. After few >>>>> >>> mins , >>>>> >>> other OSD from a node where the fault OSD exists will take over the >>>>> >>> PGs >>>>> >>> temporarily and all PGs goes to active + clean state . Cluster also >>>>> >>> does not >>>>> >>> goes down during this recovery process. >>>>> >>> >>>>> >>> Only on bluestore we see cluster going to error state when one OSD >>>>> >>> is >>>>> >>> down. >>>>> >>> We are still validating this and let you know additional findings. >>>>> >>> >>>>> >>> Thanks, >>>>> >>> Muthu >>>>> >>> >>>>> >>> On 21 January 2017 at 02:06, Shinobu Kinjo <skinjo@xxxxxxxxxx> >>>>> >>> wrote: >>>>> >>>> >>>>> >>>> `ceph pg dump` should show you something like: >>>>> >>>> >>>>> >>>> * active+undersized+degraded ... [NONE,3,2,4,1] 3 >>>>> >>>> [NONE,3,2,4,1] >>>>> >>>> >>>>> >>>> Sam, >>>>> >>>> >>>>> >>>> Am I wrong? Or is it up to something else? >>>>> >>>> >>>>> >>>> >>>>> >>>> On Sat, Jan 21, 2017 at 4:22 AM, Gregory Farnum >>>>> >>>> <gfarnum@xxxxxxxxxx> >>>>> >>>> wrote: >>>>> >>>> > I'm pretty sure the default configs won't let an EC PG go active >>>>> >>>> > with >>>>> >>>> > only "k" OSDs in its PG; it needs at least k+1 (or possibly more? >>>>> >>>> > Not >>>>> >>>> > certain). Running an "n+1" EC config is just not a good idea. >>>>> >>>> > For testing you could probably adjust this with the equivalent of >>>>> >>>> > min_size for EC pools, but I don't know the parameters off the >>>>> >>>> > top of >>>>> >>>> > my head. >>>>> >>>> > -Greg >>>>> >>>> > >>>>> >>>> > On Fri, Jan 20, 2017 at 2:15 AM, Muthusamy Muthiah >>>>> >>>> > <muthiah.muthusamy@xxxxxxxxx> wrote: >>>>> >>>> >> Hi , >>>>> >>>> >> >>>>> >>>> >> We are validating kraken 11.2.0 with bluestore on 5 node >>>>> >>>> >> cluster with >>>>> >>>> >> EC >>>>> >>>> >> 4+1. >>>>> >>>> >> >>>>> >>>> >> When an OSD is down , the peering is not happening and ceph >>>>> >>>> >> health >>>>> >>>> >> status >>>>> >>>> >> moved to ERR state after few mins. This was working in previous >>>>> >>>> >> development >>>>> >>>> >> releases. Any additional configuration required in v11.2.0 >>>>> >>>> >> >>>>> >>>> >> Following is our ceph configuration: >>>>> >>>> >> >>>>> >>>> >> mon_osd_down_out_interval = 30 >>>>> >>>> >> mon_osd_report_timeout = 30 >>>>> >>>> >> mon_osd_down_out_subtree_limit = host >>>>> >>>> >> mon_osd_reporter_subtree_level = host >>>>> >>>> >> >>>>> >>>> >> and the recovery parameters set to default. >>>>> >>>> >> >>>>> >>>> >> [root@ca-cn1 ceph]# ceph osd crush show-tunables >>>>> >>>> >> >>>>> >>>> >> { >>>>> >>>> >> "choose_local_tries": 0, >>>>> >>>> >> "choose_local_fallback_tries": 0, >>>>> >>>> >> "choose_total_tries": 50, >>>>> >>>> >> "chooseleaf_descend_once": 1, >>>>> >>>> >> "chooseleaf_vary_r": 1, >>>>> >>>> >> "chooseleaf_stable": 1, >>>>> >>>> >> "straw_calc_version": 1, >>>>> >>>> >> "allowed_bucket_algs": 54, >>>>> >>>> >> "profile": "jewel", >>>>> >>>> >> "optimal_tunables": 1, >>>>> >>>> >> "legacy_tunables": 0, >>>>> >>>> >> "minimum_required_version": "jewel", >>>>> >>>> >> "require_feature_tunables": 1, >>>>> >>>> >> "require_feature_tunables2": 1, >>>>> >>>> >> "has_v2_rules": 1, >>>>> >>>> >> "require_feature_tunables3": 1, >>>>> >>>> >> "has_v3_rules": 0, >>>>> >>>> >> "has_v4_buckets": 0, >>>>> >>>> >> "require_feature_tunables5": 1, >>>>> >>>> >> "has_v5_rules": 0 >>>>> >>>> >> } >>>>> >>>> >> >>>>> >>>> >> ceph status: >>>>> >>>> >> >>>>> >>>> >> health HEALTH_ERR >>>>> >>>> >> 173 pgs are stuck inactive for more than 300 seconds >>>>> >>>> >> 173 pgs incomplete >>>>> >>>> >> 173 pgs stuck inactive >>>>> >>>> >> 173 pgs stuck unclean >>>>> >>>> >> monmap e2: 5 mons at >>>>> >>>> >> >>>>> >>>> >> >>>>> >>>> >> {ca-cn1=10.50.5.117:6789/0,ca-cn2=10.50.5.118:6789/0,ca-cn3=10.50.5.119:6789/0,ca-cn4=10.50.5.120:6789/0,ca-cn5=10.50.5.121:6789/0} >>>>> >>>> >> election epoch 106, quorum 0,1,2,3,4 >>>>> >>>> >> ca-cn1,ca-cn2,ca-cn3,ca-cn4,ca-cn5 >>>>> >>>> >> mgr active: ca-cn1 standbys: ca-cn2, ca-cn4, ca-cn5, >>>>> >>>> >> ca-cn3 >>>>> >>>> >> osdmap e1128: 60 osds: 59 up, 59 in; 173 remapped pgs >>>>> >>>> >> flags >>>>> >>>> >> sortbitwise,require_jewel_osds,require_kraken_osds >>>>> >>>> >> pgmap v782747: 2048 pgs, 1 pools, 63133 GB data, 46293 >>>>> >>>> >> kobjects >>>>> >>>> >> 85199 GB used, 238 TB / 322 TB avail >>>>> >>>> >> 1868 active+clean >>>>> >>>> >> 173 remapped+incomplete >>>>> >>>> >> 7 active+clean+scrubbing >>>>> >>>> >> >>>>> >>>> >> MON log: >>>>> >>>> >> >>>>> >>>> >> 2017-01-20 09:25:54.715684 7f55bcafb700 0 log_channel(cluster) >>>>> >>>> >> log >>>>> >>>> >> [INF] : >>>>> >>>> >> osd.54 out (down for 31.703786) >>>>> >>>> >> 2017-01-20 09:25:54.725688 7f55bf4d5700 0 >>>>> >>>> >> mon.ca-cn1@0(leader).osd >>>>> >>>> >> e1120 >>>>> >>>> >> crush map has features 288250512065953792, adjusting msgr >>>>> >>>> >> requires >>>>> >>>> >> 2017-01-20 09:25:54.729019 7f55bf4d5700 0 log_channel(cluster) >>>>> >>>> >> log >>>>> >>>> >> [INF] : >>>>> >>>> >> osdmap e1120: 60 osds: 59 up, 59 in >>>>> >>>> >> 2017-01-20 09:25:54.735987 7f55bf4d5700 0 log_channel(cluster) >>>>> >>>> >> log >>>>> >>>> >> [INF] : >>>>> >>>> >> pgmap v781993: 2048 pgs: 1869 active+clean, 173 incomplete, 6 >>>>> >>>> >> active+clean+scrubbing; 63159 GB data, 85201 GB used, 238 TB / >>>>> >>>> >> 322 TB >>>>> >>>> >> avail; >>>>> >>>> >> 21825 B/s rd, 163 MB/s wr, 2046 op/s >>>>> >>>> >> 2017-01-20 09:25:55.737749 7f55bf4d5700 0 >>>>> >>>> >> mon.ca-cn1@0(leader).osd >>>>> >>>> >> e1121 >>>>> >>>> >> crush map has features 288250512065953792, adjusting msgr >>>>> >>>> >> requires >>>>> >>>> >> 2017-01-20 09:25:55.744338 7f55bf4d5700 0 log_channel(cluster) >>>>> >>>> >> log >>>>> >>>> >> [INF] : >>>>> >>>> >> osdmap e1121: 60 osds: 59 up, 59 in >>>>> >>>> >> 2017-01-20 09:25:55.749616 7f55bf4d5700 0 log_channel(cluster) >>>>> >>>> >> log >>>>> >>>> >> [INF] : >>>>> >>>> >> pgmap v781994: 2048 pgs: 29 remapped+incomplete, 1869 >>>>> >>>> >> active+clean, >>>>> >>>> >> 144 >>>>> >>>> >> incomplete, 6 active+clean+scrubbing; 63159 GB data, 85201 GB >>>>> >>>> >> used, >>>>> >>>> >> 238 TB / >>>>> >>>> >> 322 TB avail; 44503 B/s rd, 45681 kB/s wr, 518 op/s >>>>> >>>> >> 2017-01-20 09:25:56.768721 7f55bf4d5700 0 log_channel(cluster) >>>>> >>>> >> log >>>>> >>>> >> [INF] : >>>>> >>>> >> pgmap v781995: 2048 pgs: 47 remapped+incomplete, 1869 >>>>> >>>> >> active+clean, >>>>> >>>> >> 126 >>>>> >>>> >> incomplete, 6 active+clean+scrubbing; 63159 GB data, 85201 GB >>>>> >>>> >> used, >>>>> >>>> >> 238 TB / >>>>> >>>> >> 322 TB avail; 20275 B/s rd, 72742 kB/s wr, 665 op/s >>>>> >>>> >> >>>>> >>>> >> Thanks, >>>>> >>>> >> Muthu >>>>> >>>> >> >>>>> >>>> >> >>>>> >>>> >> _______________________________________________ >>>>> >>>> >> ceph-users mailing list >>>>> >>>> >> ceph-users@xxxxxxxxxxxxxx >>>>> >>>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>> >>>> >> >>>>> >>>> > _______________________________________________ >>>>> >>>> > ceph-users mailing list >>>>> >>>> > ceph-users@xxxxxxxxxxxxxx >>>>> >>>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>> >>> >>>>> >>> >>>>> >> >>>>> >> >>>>> >> _______________________________________________ >>>>> >> ceph-users mailing list >>>>> >> ceph-users@xxxxxxxxxxxxxx >>>>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>> >> >>>> >>>> >>> >> _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com