Re: Problems with CephFS

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Herbert,

could you please run "ceph osd df"?

Cheers,

Vadim


On 12.06.2018 11:06, Steininger, Herbert wrote:
Hi Guys,

i've inherited a CephFS-Cluster, I'm fairly new to CephFS.
The Cluster was down and I managed somehow to bring it up again.
But now there are some Problems that I can't fix that easily.
This is what 'ceph -s' is giving me as Info:
[root@pcl241 ceph]# ceph -s
     cluster cde1487e-f930-417a-9403-28e9ebf406b8
      health HEALTH_WARN
             2 pgs backfill_toofull
             1 pgs degraded
             1 pgs stuck degraded
             2 pgs stuck unclean
             1 pgs stuck undersized
             1 pgs undersized
             recovery 260/29731463 objects degraded (0.001%)
             recovery 798/29731463 objects misplaced (0.003%)
             2 near full osd(s)
             crush map has legacy tunables (require bobtail, min is firefly)
             crush map has straw_calc_version=0
      monmap e8: 3 mons at {cephcontrol=172.22.12.241:6789/0,slurmbackup=172.22.20.4:6789/0,slurmmaster=172.22.20.3:6789/0}
             election epoch 48, quorum 0,1,2 cephcontrol,slurmmaster,slurmbackup
       fsmap e2288: 1/1/1 up {0=pcl241=up:active}
      osdmap e10865: 6 osds: 6 up, 6 in; 2 remapped pgs
             flags nearfull
       pgmap v14103169: 320 pgs, 3 pools, 30899 GB data, 9678 kobjects
             92929 GB used, 22139 GB / 112 TB avail
             260/29731463 objects degraded (0.001%)
             798/29731463 objects misplaced (0.003%)
                  316 active+clean
                    2 active+clean+scrubbing+deep
                    1 active+undersized+degraded+remapped+backfill_toofull
                    1 active+remapped+backfill_toofull
[root@pcl241 ceph]#


[root@pcl241 ceph]# ceph osd tree
ID  WEIGHT    TYPE NAME              UP/DOWN REWEIGHT PRIMARY-AFFINITY
  -1 109.13992 root default
  -2         0     host A1214-2950-01
  -3         0     host A1214-2950-02
  -4         0     host A1214-2950-04
  -5         0     host A1214-2950-05
  -6         0     host A1214-2950-03
  -7  18.18999     host cuda002
   1  18.18999         osd.1               up  1.00000          1.00000
  -8  18.18999     host cuda001
   0  18.18999         osd.0               up  1.00000          1.00000
  -9  18.18999     host cuda005
   3  18.18999         osd.3               up  1.00000          1.00000
-10  18.18999     host cuda003
   4  18.18999         osd.4               up  1.00000          1.00000
-11  18.18999     host cuda004
   2  18.18999         osd.2               up  1.00000          1.00000
-12  18.18999     host A1214-2950-06
   5  18.18999         osd.5               up  1.00000          1.00000
-13         0     host pe9




Could someone please put me in the right Direction about what to do to fix the Problems?
It seems that two OSD are full, but how can I solve that, if I don't have additionally hardware available?
Also it seems that the Cluster has different ceph-versions running (Hammer and Jewel), how to solve that?
Ceph-(mds/-mon/-osd) is running on Scientific Linux.
If more Info is needed, just let me know.

Thanks in Advance,
Steininger Herbert

---
Herbert Steininger
Leiter EDV
Administrator
Max-Planck-Institut für Psychiatrie - EDV
Kraepelinstr.  2-10
80804 München
Tel      +49 (0)89 / 30622-368
Mail   herbert_steininger@xxxxxxxxxxxx
Web  http://www.psych.mpg.de


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

--
Vadim Bulst

Universität Leipzig / URZ
04109  Leipzig, Augustusplatz 10

phone: ++49-341-97-33380
mail:    vadim.bulst@xxxxxxxxxxxxxx


Attachment: smime.p7s
Description: S/MIME Cryptographic Signature

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux