Hello, I have a bright new cluster with 2 pools, but cluster keeps pgs in inactive state. I have 3 OSDs and 1 Mon… all seems ok except I could not have pgs in clean+active state ! I might miss something obvious but I really don’t know what…. Someone could help me ? I tried to seek answers among the list mail threads, but no luck, other situation seems different Thank you a lot for your help Vincent # ceph -v ceph version 13.2.4 (b10be4d44915a4d78a8e06aa31919e74927b142e) mimic (stable) # ceph -s cluster: id: ff4c91fb-3c29-4d9f-a26f-467d6b6a712e health: HEALTH_WARN Reduced data availability: 200 pgs inactive services: mon: 1 daemons, quorum ip-10-8-66-123.eu-west-2.compute.internal mgr: ip-10-8-66-123.eu-west-2.compute.internal(active) osd: 3 osds: 3 up, 3 in data: pools: 2 pools, 200 pgs objects: 0 objects, 0 B usage: 3.0 GiB used, 2.9 TiB / 2.9 TiB avail pgs: 100.000% pgs unknown 200 unknown # ceph osd tree -f json-pretty { "nodes": [ { "id": -1, "name": "default", "type": "root", "type_id": 10, "children": [ -3, -5, -7 ] }, { "id": -7, "name": "ip-10-8-10-108", "type": "host", "type_id": 1, "pool_weights": {}, "children": [ 2 ] }, { "id": 2, "device_class": "hdd", "name": "osd.2", "type": "osd", "type_id": 0, "crush_weight": 0.976593, "depth": 2, "pool_weights": {}, "exists": 1, "status": "up", "reweight": 1.000000, "primary_affinity": 1.000000 }, { "id": -5, "name": "ip-10-8-22-148", "type": "host", "type_id": 1, "pool_weights": {}, "children": [ 1 ] }, { "id": 1, "device_class": "hdd", "name": "osd.1", "type": "osd", "type_id": 0, "crush_weight": 0.976593, "depth": 2, "pool_weights": {}, "exists": 1, "status": "up", "reweight": 1.000000, "primary_affinity": 1.000000 }, { "id": -3, "name": "ip-10-8-5-246", "type": "host", "type_id": 1, "pool_weights": {}, "children": [ 0 ] }, { "id": 0, "device_class": "hdd", "name": "osd.0", "type": "osd", "type_id": 0, "crush_weight": 0.976593, "depth": 2, "pool_weights": {}, "exists": 1, "status": "up", "reweight": 1.000000, "primary_affinity": 1.000000 } ], "stray": [] } # cat /etc/ceph/ceph.conf [global] fsid = ff4c91fb-3c29-4d9f-a26f-467d6b6a712e mon initial members = ip-10-8-66-123 mon host = 10.8.66.123 auth_cluster_required = cephx auth_service_required = cephx auth_client_required = cephx pid file = /var/run/$cluster/$type.pid #Choose reasonable numbers for number of replicas and placement groups. osd pool default size = 3 # Write an object 3 times osd pool default min size = 2 # Allow writing 2 copy in a degraded state osd pool default pg num = 100 osd pool default pgp num = 100 #Choose a reasonable crush leaf type #0 for a 1-node cluster. #1 for a multi node cluster in a single rack #2 for a multi node, multi chassis cluster with multiple hosts in a chassis #3 for a multi node cluster with hosts across racks, etc. osd crush chooseleaf type = 2 [mon] debug mon = 20 # ceph health detail HEALTH_WARN Reduced data availability: 200 pgs inactive PG_AVAILABILITY Reduced data availability: 200 pgs inactive pg 1.46 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.47 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.48 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.49 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.4a is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.4b is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.4c is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.4d is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.4e is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.4f is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.50 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.51 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.52 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.53 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.54 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.55 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.56 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.57 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.58 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.59 is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.5a is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.5b is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.5c is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.5d is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.5e is stuck inactive for 10848.068201, current state unknown, last acting [] pg 1.5f is stuck inactive for 10848.068201, current state unknown, last acting [] pg 2.44 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.48 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.49 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.4a is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.4b is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.4c is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.4d is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.4e is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.4f is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.50 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.51 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.52 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.53 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.54 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.55 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.56 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.57 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.58 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.59 is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.5a is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.5b is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.5c is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.5d is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.5e is stuck inactive for 10846.400420, current state unknown, last acting [] pg 2.5f is stuck inactive for 10846.400420, current state unknown, last acting [] This email and any attachments are intended solely for the use of the individual or entity to whom it is addressed and may be confidential and/or privileged. If you are not one of the named recipients or have received this email in error, (i) you should not read, disclose, or copy it, (ii) please notify sender of your receipt by reply email and delete this email and all attachments, (iii) Dassault Systèmes does not accept or assume any liability or responsibility for any use of or reliance on this email. Please be informed that your personal data are processed according to our data privacy policy as described on our website. Should you have any questions related to personal data protection, please contact 3DS Data Protection Officer at 3DS.compliance-privacy@xxxxxxx
For other languages, go to https://www.3ds.com/terms/email-disclaimer |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com