Ceph Not getting into a clean state

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Right,

I've run into the situation where the system seems reluctant to 
reorganise after changing all the pool sizes - until the osds are 
restarted (essentially I just rebooted each host in turn) *then* the 
health went to OK. This was a while ago (pre 0.72), so something else 
may be going on with current versions - but maybe try restarting 'em.

Regards

Mark

On 09/05/14 18:53, Georg H?llrigl wrote:
> Hello,
>
> I've already thought about that - but even after changing the 
> replication level (size) I'm not getting a clean cluster (there are 
> only the default pools ATM):
>
> root at ceph-m-02:~#ceph -s
>     cluster b04fc583-9e71-48b7-a741-92f4dff4cfef
>      health HEALTH_WARN 232 pgs stuck unclean; recovery 26/126 objects 
> degraded (20.635%)
>      monmap e2: 3 mons at 
> {ceph-m-01=10.0.0.100:6789/0,ceph-m-02=10.0.1.101:6789/0,ceph-m-03=10.0.1.102:6789/0}, 
> election epoch 8, quorum 0,1,2 ceph-m-01,ceph-m-02,ceph-m-03
>      osdmap e56: 9 osds: 9 up, 9 in
>       pgmap v287: 232 pgs, 8 pools, 822 bytes data, 43 objects
>             9342 MB used, 78317 GB / 78326 GB avail
>             26/126 objects degraded (20.635%)
>                  119 active
>                  113 active+remapped
> root at ceph-m-02:~#ceph osd dump | grep size
> pool 0 'data' replicated size 2 min_size 2 crush_ruleset 0 object_hash 
> rjenkins pg_num 64 pgp_num 64 last_change 48 owner 0 flags hashpspool 
> crash_replay_interval 45 stripe_width 0
> pool 1 'metadata' replicated size 2 min_size 2 crush_ruleset 0 
> object_hash rjenkins pg_num 64 pgp_num 64 last_change 49 owner 0 flags 
> hashpspool stripe_width 0
> pool 2 'rbd' replicated size 2 min_size 2 crush_ruleset 0 object_hash 
> rjenkins pg_num 64 pgp_num 64 last_change 50 owner 0 flags hashpspool 
> stripe_width 0
> pool 3 '.rgw.root' replicated size 2 min_size 2 crush_ruleset 0 
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 52 owner 0 flags 
> hashpspool stripe_width 0
> pool 4 '.rgw.control' replicated size 2 min_size 2 crush_ruleset 0 
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 53 owner 0 flags 
> hashpspool stripe_width 0
> pool 5 '.rgw' replicated size 2 min_size 2 crush_ruleset 0 object_hash 
> rjenkins pg_num 8 pgp_num 8 last_change 54 owner 18446744073709551615 
> flags hashpspool stripe_width 0
> pool 6 '.rgw.gc' replicated size 2 min_size 2 crush_ruleset 0 
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 55 owner 0 flags 
> hashpspool stripe_width 0
> pool 7 '.users.uid' replicated size 2 min_size 2 crush_ruleset 0 
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 56 owner 
> 18446744073709551615 flags hashpspool stripe_width 0
>
>
> Kind Regards,
> Georg
>
>
> On 09.05.2014 08:29, Mark Kirkwood wrote:
>> So that's two hosts - if this is a new cluster chances are the pools
>> have replication size=3, and won't place replica pgs on the same host...
>> 'ceph osd dump' will let you know if this is the case. If it is ether
>> reduce size to 2, add another host or edit your crush rules to allow
>> replica pgs on the same host.
>>
>> Cheers
>>
>> Mark
>>
>> On 09/05/14 18:20, Georg H?llrigl wrote:
>>> #ceph osd tree
>>> # id    weight  type name       up/down reweight
>>> -1      76.47   root default
>>> -2      32.72           host ceph-s-01
>>> 0       7.27                    osd.0   up      1
>>> 1       7.27                    osd.1   up      1
>>> 2       9.09                    osd.2   up      1
>>> 3       9.09                    osd.3   up      1
>>> -3      43.75           host ceph-s-02
>>> 4       10.91                   osd.4   up      1
>>> 5       0.11                    osd.5   up      1
>>> 6       10.91                   osd.6   up      1
>>> 7       10.91                   osd.7   up      1
>>> 8       10.91                   osd.8   up      1
>>>
>>>
>>> On 08.05.2014 19:11, Craig Lewis wrote:
>>>> What does `ceph osd tree` output?
>>>>
>>>> On 5/8/14 07:30 , Georg H?llrigl wrote:
>>>>> Hello,
>>>>>
>>>>> We've a fresh cluster setup - with Ubuntu 14.04 and ceph firefly. By
>>>>> now I've tried this multiple times - but the result keeps the same 
>>>>> and
>>>>> shows me lots of troubles (the cluster is empty, no client has
>>>>> accessed it)
>>>>>
>>>>> #ceph -s
>>>>>     cluster b04fc583-9e71-48b7-a741-92f4dff4cfef
>>>>>      health HEALTH_WARN 470 pgs stale; 470 pgs stuck stale; 18 pgs
>>>>> stuck unclean; 26 requests are blocked > 32 sec
>>>>>      monmap e2: 3 mons at
>>>>> {ceph-m-01=10.0.0.100:6789/0,ceph-m-02=10.0.1.101:6789/0,ceph-m-03=10.0.1.102:6789/0}, 
>>>>>
>>>>>
>>>>> election epoch 8, quorum 0,1,2 ceph-m-01,ceph-m-02,ceph-m-03
>>>>>      osdmap e409: 9 osds: 9 up, 9 in
>>>>>       pgmap v1231: 480 pgs, 9 pools, 822 bytes data, 43 objects
>>>>>             9373 MB used, 78317 GB / 78326 GB avail
>>>>>                  451 stale+active+clean
>>>>>                    1 stale+active+clean+scrubbing
>>>>>                   10 active+clean
>>>>>                   18 stale+active+remapped
>>>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users at lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> _______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux