Hi Herbert, could you please run "ceph osd df"? Cheers, Vadim On 12.06.2018 11:06, Steininger, Herbert wrote:
Hi Guys, i've inherited a CephFS-Cluster, I'm fairly new to CephFS. The Cluster was down and I managed somehow to bring it up again. But now there are some Problems that I can't fix that easily. This is what 'ceph -s' is giving me as Info: [root@pcl241 ceph]# ceph -s cluster cde1487e-f930-417a-9403-28e9ebf406b8 health HEALTH_WARN 2 pgs backfill_toofull 1 pgs degraded 1 pgs stuck degraded 2 pgs stuck unclean 1 pgs stuck undersized 1 pgs undersized recovery 260/29731463 objects degraded (0.001%) recovery 798/29731463 objects misplaced (0.003%) 2 near full osd(s) crush map has legacy tunables (require bobtail, min is firefly) crush map has straw_calc_version=0 monmap e8: 3 mons at {cephcontrol=172.22.12.241:6789/0,slurmbackup=172.22.20.4:6789/0,slurmmaster=172.22.20.3:6789/0} election epoch 48, quorum 0,1,2 cephcontrol,slurmmaster,slurmbackup fsmap e2288: 1/1/1 up {0=pcl241=up:active} osdmap e10865: 6 osds: 6 up, 6 in; 2 remapped pgs flags nearfull pgmap v14103169: 320 pgs, 3 pools, 30899 GB data, 9678 kobjects 92929 GB used, 22139 GB / 112 TB avail 260/29731463 objects degraded (0.001%) 798/29731463 objects misplaced (0.003%) 316 active+clean 2 active+clean+scrubbing+deep 1 active+undersized+degraded+remapped+backfill_toofull 1 active+remapped+backfill_toofull [root@pcl241 ceph]# [root@pcl241 ceph]# ceph osd tree ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY -1 109.13992 root default -2 0 host A1214-2950-01 -3 0 host A1214-2950-02 -4 0 host A1214-2950-04 -5 0 host A1214-2950-05 -6 0 host A1214-2950-03 -7 18.18999 host cuda002 1 18.18999 osd.1 up 1.00000 1.00000 -8 18.18999 host cuda001 0 18.18999 osd.0 up 1.00000 1.00000 -9 18.18999 host cuda005 3 18.18999 osd.3 up 1.00000 1.00000 -10 18.18999 host cuda003 4 18.18999 osd.4 up 1.00000 1.00000 -11 18.18999 host cuda004 2 18.18999 osd.2 up 1.00000 1.00000 -12 18.18999 host A1214-2950-06 5 18.18999 osd.5 up 1.00000 1.00000 -13 0 host pe9 Could someone please put me in the right Direction about what to do to fix the Problems? It seems that two OSD are full, but how can I solve that, if I don't have additionally hardware available? Also it seems that the Cluster has different ceph-versions running (Hammer and Jewel), how to solve that? Ceph-(mds/-mon/-osd) is running on Scientific Linux. If more Info is needed, just let me know. Thanks in Advance, Steininger Herbert --- Herbert Steininger Leiter EDV Administrator Max-Planck-Institut für Psychiatrie - EDV Kraepelinstr. 2-10 80804 München Tel +49 (0)89 / 30622-368 Mail herbert_steininger@xxxxxxxxxxxx Web http://www.psych.mpg.de _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
-- Vadim Bulst Universität Leipzig / URZ 04109 Leipzig, Augustusplatz 10 phone: ++49-341-97-33380 mail: vadim.bulst@xxxxxxxxxxxxxx
Attachment:
smime.p7s
Description: S/MIME Cryptographic Signature
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com