Re: Can't recover pgs degraded/stuck unclean/undersized

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



sure, as requested:

cephfs was created using the following command:

ceph osd pool create cephfs_metadata 128 128
ceph osd pool create cephfs_data 128 128
ceph fs new cephfs cephfs_metadata cephfs_data

ceph.conf:
https://paste.debian.net/895841/

# ceph osd crush tree
https://paste.debian.net/895839/

# ceph osd crush rule list
[
    "replicated_ruleset",
    "replicated_ruleset_ssd"
]

# ceph osd crush rule dump
https://paste.debian.net/895842/

# ceph osd tree
ID WEIGHT   TYPE NAME                     UP/DOWN REWEIGHT PRIMARY-AFFINITY 
-3  0.07999 root default-ssd                                                
-5  0.03999     host dc1-master-ds02-ssd                                   
11  0.03999         osd.11                     up  1.00000          1.00000 
-6  0.03999     host dc1-master-ds03-ssd                                   
13  0.03999         osd.13                     up  1.00000          1.00000 
-1 31.39999 root default                                                    
-2 31.39999     host dc1-master-ds01                                       
 0  3.70000         osd.0                      up  1.00000          1.00000 
 1  3.70000         osd.1                      up  1.00000          1.00000 
 2  4.00000         osd.2                      up  1.00000          1.00000 
 3  4.00000         osd.3                      up  1.00000          1.00000 
 4  4.00000         osd.4                      up  1.00000          1.00000 
 5  4.00000         osd.5                      up  1.00000          1.00000 
 6  4.00000         osd.6                      up  1.00000          1.00000 
 7  4.00000         osd.7                      up  1.00000          1.00000 


# ceph osd pool ls
.rgw.root
master.rgw.control
master.rgw.data.root
master.rgw.gc
master.rgw.log
master.rgw.intent-log
master.rgw.usage
master.rgw.users.keys
master.rgw.users.email
master.rgw.users.swift
master.rgw.users.uid
master.rgw.buckets.index
master.rgw.buckets.data
master.rgw.meta
master.rgw.buckets.non-ec
rbd
cephfs_metadata
cephfs_data

# ceph osd pool stats
https://paste.debian.net/895840/




On Tue, Nov 15, 2016 at 10:33 AM Burkhard Linke <Burkhard.Linke@xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx> wrote:
Hi,


On 11/15/2016 01:27 PM, Webert de Souza Lima wrote:
> Not that I know of. On 5 other clusters it works just fine and
> configuration is the same for all.
> On this cluster I was using only radosgw, but cephfs was not in use
> but it had been already created following our procedures.
>
> This happened right after mounting it.
Do you use any different setup for one of the pools?
active+undersized+degraded means that the crush rules for a PG cannot be
satisfied, and 128 PGs sounds like the default setup for the number of PGs.

With 10 OSDs I would suspect that you do not have enough host to satisfy
all crush requirements. Can you post your crush tree, the crush rules
and the detailed pool configuration?

Regards,
Burkhard
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux