Ceph Not getting into a clean state

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

I've already thought about that - but even after changing the 
replication level (size) I'm not getting a clean cluster (there are only 
the default pools ATM):

root at ceph-m-02:~#ceph -s
     cluster b04fc583-9e71-48b7-a741-92f4dff4cfef
      health HEALTH_WARN 232 pgs stuck unclean; recovery 26/126 objects 
degraded (20.635%)
      monmap e2: 3 mons at 
{ceph-m-01=10.0.0.100:6789/0,ceph-m-02=10.0.1.101:6789/0,ceph-m-03=10.0.1.102:6789/0}, 
election epoch 8, quorum 0,1,2 ceph-m-01,ceph-m-02,ceph-m-03
      osdmap e56: 9 osds: 9 up, 9 in
       pgmap v287: 232 pgs, 8 pools, 822 bytes data, 43 objects
             9342 MB used, 78317 GB / 78326 GB avail
             26/126 objects degraded (20.635%)
                  119 active
                  113 active+remapped
root at ceph-m-02:~#ceph osd dump | grep size
pool 0 'data' replicated size 2 min_size 2 crush_ruleset 0 object_hash 
rjenkins pg_num 64 pgp_num 64 last_change 48 owner 0 flags hashpspool 
crash_replay_interval 45 stripe_width 0
pool 1 'metadata' replicated size 2 min_size 2 crush_ruleset 0 
object_hash rjenkins pg_num 64 pgp_num 64 last_change 49 owner 0 flags 
hashpspool stripe_width 0
pool 2 'rbd' replicated size 2 min_size 2 crush_ruleset 0 object_hash 
rjenkins pg_num 64 pgp_num 64 last_change 50 owner 0 flags hashpspool 
stripe_width 0
pool 3 '.rgw.root' replicated size 2 min_size 2 crush_ruleset 0 
object_hash rjenkins pg_num 8 pgp_num 8 last_change 52 owner 0 flags 
hashpspool stripe_width 0
pool 4 '.rgw.control' replicated size 2 min_size 2 crush_ruleset 0 
object_hash rjenkins pg_num 8 pgp_num 8 last_change 53 owner 0 flags 
hashpspool stripe_width 0
pool 5 '.rgw' replicated size 2 min_size 2 crush_ruleset 0 object_hash 
rjenkins pg_num 8 pgp_num 8 last_change 54 owner 18446744073709551615 
flags hashpspool stripe_width 0
pool 6 '.rgw.gc' replicated size 2 min_size 2 crush_ruleset 0 
object_hash rjenkins pg_num 8 pgp_num 8 last_change 55 owner 0 flags 
hashpspool stripe_width 0
pool 7 '.users.uid' replicated size 2 min_size 2 crush_ruleset 0 
object_hash rjenkins pg_num 8 pgp_num 8 last_change 56 owner 
18446744073709551615 flags hashpspool stripe_width 0


Kind Regards,
Georg


On 09.05.2014 08:29, Mark Kirkwood wrote:
> So that's two hosts - if this is a new cluster chances are the pools
> have replication size=3, and won't place replica pgs on the same host...
> 'ceph osd dump' will let you know if this is the case. If it is ether
> reduce size to 2, add another host or edit your crush rules to allow
> replica pgs on the same host.
>
> Cheers
>
> Mark
>
> On 09/05/14 18:20, Georg H?llrigl wrote:
>> #ceph osd tree
>> # id    weight  type name       up/down reweight
>> -1      76.47   root default
>> -2      32.72           host ceph-s-01
>> 0       7.27                    osd.0   up      1
>> 1       7.27                    osd.1   up      1
>> 2       9.09                    osd.2   up      1
>> 3       9.09                    osd.3   up      1
>> -3      43.75           host ceph-s-02
>> 4       10.91                   osd.4   up      1
>> 5       0.11                    osd.5   up      1
>> 6       10.91                   osd.6   up      1
>> 7       10.91                   osd.7   up      1
>> 8       10.91                   osd.8   up      1
>>
>>
>> On 08.05.2014 19:11, Craig Lewis wrote:
>>> What does `ceph osd tree` output?
>>>
>>> On 5/8/14 07:30 , Georg H?llrigl wrote:
>>>> Hello,
>>>>
>>>> We've a fresh cluster setup - with Ubuntu 14.04 and ceph firefly. By
>>>> now I've tried this multiple times - but the result keeps the same and
>>>> shows me lots of troubles (the cluster is empty, no client has
>>>> accessed it)
>>>>
>>>> #ceph -s
>>>>     cluster b04fc583-9e71-48b7-a741-92f4dff4cfef
>>>>      health HEALTH_WARN 470 pgs stale; 470 pgs stuck stale; 18 pgs
>>>> stuck unclean; 26 requests are blocked > 32 sec
>>>>      monmap e2: 3 mons at
>>>> {ceph-m-01=10.0.0.100:6789/0,ceph-m-02=10.0.1.101:6789/0,ceph-m-03=10.0.1.102:6789/0},
>>>>
>>>> election epoch 8, quorum 0,1,2 ceph-m-01,ceph-m-02,ceph-m-03
>>>>      osdmap e409: 9 osds: 9 up, 9 in
>>>>       pgmap v1231: 480 pgs, 9 pools, 822 bytes data, 43 objects
>>>>             9373 MB used, 78317 GB / 78326 GB avail
>>>>                  451 stale+active+clean
>>>>                    1 stale+active+clean+scrubbing
>>>>                   10 active+clean
>>>>                   18 stale+active+remapped
>>>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users at lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux