Hi, How about this "osd: 7 osds: 6 up (since 3h), 6 in (since 6w)" 1 osd is missing? ------------------ Original ------------------ From: "Konold, Martin"<martin.konold@xxxxxxxxxx>; Date: Fri, Apr 1, 2022 05:56 PM To: "Janne Johansson"<icepic.dz@xxxxxxxxx>; Cc: "ceph-users"<ceph-users@xxxxxxx>; Subject: Re: PGs and OSDs unknown Hi, restarting ceph managers did not change anything. # systemctl status ceph-mgr@hbase10.service ● ceph-mgr@hbase10.service - Ceph cluster manager daemon Loaded: loaded (/lib/systemd/system/ceph-mgr@.service; enabled; vendor preset: enabled) Drop-In: /usr/lib/systemd/system/ceph-mgr@.service.d └─ceph-after-pve-cluster.conf Active: active (running) since Fri 2022-04-01 11:23:45 CEST; 2min 56s ago Main PID: 124618 (ceph-mgr) Tasks: 21 (limit: 154429) Memory: 200.0M CPU: 1.975s CGroup: /system.slice/system-ceph\x2dmgr.slice/ceph-mgr@hbase10.service └─124618 /usr/bin/ceph-mgr -f --cluster ceph --id hbase10 --setuser ceph --setgroup ceph Apr 01 11:23:45 hbase10.h.konsec.com systemd[1]: Started Ceph cluster manager daemon. Apr 01 11:23:47 hbase10.h.konsec.com ceph-mgr[124618]: context.c:56: warning: mpd_setminalloc: ignoring request to set MPD_MINALLOC a second time root@hbase10:~# ceph -s cluster: id: 0393d3c0-8788-4b9f-8572-65826aae2ee4 health: HEALTH_WARN Reduced data availability: 448 pgs inactive services: mon: 6 daemons, quorum hbase10,hbase11,hbase13,hbase16,hbase17,hbase18 (age 3h) mgr: hbase14(active, since 31m), standbys: hbase17, hbase11, hbase12, hbase18, hbase13, hbase15, hbase16 mds: 1/1 daemons up, 5 standby osd: 7 osds: 6 up (since 3h), 6 in (since 6w) data: volumes: 1/1 healthy pools: 4 pools, 448 pgs objects: 0 objects, 0 B usage: 0 B used, 0 B / 0 B avail pgs: 100.000% pgs unknown 448 unknown --- Viele Grüße ppa. Martin Konold -- Martin Konold - Prokurist, CTO KONSEC GmbH - make things real Amtsgericht Stuttgart, HRB 23690 Geschäftsführer: Andreas Mack Im Köller 3, 70794 Filderstadt, Germany On 2022-04-01 11:17, Janne Johansson wrote: > Den fre 1 apr. 2022 kl 11:15 skrev Konold, Martin > <martin.konold@xxxxxxxxxx>: >> Hi, >> running Ceph 16.2.7 on a pure NVME Cluster with 9 nodes I am >> experiencing "Reduced data availability: 448 pgs inactive". >> >> I cannot see any statistics or pool information with "ceph -s". > > Since the cluster seems operational, chances are high the MGR(s) are > just stuck, try failing over and/or restart mgr and see if that > doesn't fix it. > >> The RBDs are still operational and "ceph report" shows the osds as >> expected. >> _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx