Re: one PG is stuck on single OSD and does not sync back

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,
sure

ID  CLASS WEIGHT    TYPE NAME       STATUS REWEIGHT PRI-AFF
 -1       672.68457 root default
 -2        58.20561     host s3db1
 23   hdd  14.55269         osd.23      up  1.00000 1.00000
 69   hdd  14.55269         osd.69      up  1.00000 1.00000
 73   hdd  14.55269         osd.73      up  1.00000 1.00000
 79   hdd   3.63689         osd.79      up  1.00000 1.00000
 80   hdd   3.63689         osd.80      up  1.00000 1.00000
 81   hdd   3.63689         osd.81      up  1.00000 1.00000
 82   hdd   3.63689         osd.82      up  1.00000 1.00000
-11        50.94173     host s3db10
 63   hdd   7.27739         osd.63      up  1.00000 1.00000
 64   hdd   7.27739         osd.64      up  1.00000 1.00000
 65   hdd   7.27739         osd.65      up  1.00000 1.00000
 66   hdd   7.27739         osd.66      up  1.00000 1.00000
 67   hdd   7.27739         osd.67      up  1.00000 1.00000
 68   hdd   7.27739         osd.68      up  1.00000 1.00000
 70   hdd   7.27739         osd.70      up  1.00000 1.00000
-12        50.94173     host s3db11
 46   hdd   7.27739         osd.46      up  1.00000 1.00000
 47   hdd   7.27739         osd.47      up  1.00000 1.00000
 48   hdd   7.27739         osd.48      up  1.00000 1.00000
 49   hdd   7.27739         osd.49      up  1.00000 1.00000
 50   hdd   7.27739         osd.50      up  1.00000 1.00000
 51   hdd   7.27739         osd.51      up  1.00000 1.00000
 72   hdd   7.27739         osd.72      up  1.00000 1.00000
-37        58.55478     host s3db12
 19   hdd   3.68750         osd.19      up  1.00000 1.00000
 71   hdd   3.68750         osd.71      up  1.00000 1.00000
 75   hdd   3.68750         osd.75      up  1.00000 1.00000
 76   hdd   3.68750         osd.76      up  1.00000 1.00000
 77   hdd  14.60159         osd.77      up  1.00000 1.00000
 78   hdd  14.60159         osd.78      up  1.00000 1.00000
 83   hdd  14.60159         osd.83      up  1.00000 1.00000
 -3        58.49872     host s3db2
  1   hdd  14.65039         osd.1       up  1.00000 1.00000
  3   hdd   3.63689         osd.3       up  1.00000 1.00000
  4   hdd   3.63689         osd.4       up  1.00000 1.00000
  5   hdd   3.63689         osd.5       up  1.00000 1.00000
  6   hdd   3.63689         osd.6       up  1.00000 1.00000
  7   hdd  14.65039         osd.7       up  1.00000 1.00000
 74   hdd  14.65039         osd.74      up  1.00000 1.00000
 -4        58.49872     host s3db3
  2   hdd  14.65039         osd.2       up  1.00000 1.00000
  9   hdd  14.65039         osd.9       up  1.00000 1.00000
 10   hdd  14.65039         osd.10      up  1.00000 1.00000
 12   hdd   3.63689         osd.12      up  1.00000 1.00000
 13   hdd   3.63689         osd.13      up  1.00000 1.00000
 14   hdd   3.63689         osd.14      up  1.00000       0
 15   hdd   3.63689         osd.15      up  1.00000 1.00000
 -5        58.49872     host s3db4
 11   hdd  14.65039         osd.11      up  1.00000 1.00000
 17   hdd  14.65039         osd.17      up  1.00000 1.00000
 18   hdd  14.65039         osd.18      up  1.00000 1.00000
 20   hdd   3.63689         osd.20      up  1.00000 1.00000
 21   hdd   3.63689         osd.21      up  1.00000 1.00000
 22   hdd   3.63689         osd.22      up  1.00000 1.00000
 24   hdd   3.63689         osd.24      up  1.00000 1.00000
 -6        58.89636     host s3db5
  0   hdd   3.73630         osd.0       up  1.00000 1.00000
 25   hdd   3.73630         osd.25      up  1.00000 1.00000
 26   hdd   3.73630         osd.26      up  1.00000 1.00000
 27   hdd   3.73630         osd.27      up  1.00000 1.00000
 28   hdd  14.65039         osd.28      up  1.00000 1.00000
 29   hdd  14.65039         osd.29      up  1.00000 1.00000
 30   hdd  14.65039         osd.30      up  1.00000 1.00000
 -7        58.89636     host s3db6
 32   hdd   3.73630         osd.32      up  1.00000 1.00000
 33   hdd   3.73630         osd.33      up  1.00000 1.00000
 34   hdd   3.73630         osd.34      up  1.00000 1.00000
 35   hdd   3.73630         osd.35      up  1.00000 1.00000
 36   hdd  14.65039         osd.36      up  1.00000 1.00000
 37   hdd  14.65039         osd.37      up  1.00000 1.00000
 38   hdd  14.65039         osd.38      up  1.00000 1.00000
 -8        58.89636     host s3db7
 39   hdd   3.73630         osd.39      up  1.00000 1.00000
 40   hdd   3.73630         osd.40      up  1.00000 1.00000
 41   hdd   3.73630         osd.41      up  1.00000 1.00000
 42   hdd   3.73630         osd.42      up  1.00000 1.00000
 43   hdd  14.65039         osd.43      up  1.00000 1.00000
 44   hdd  14.65039         osd.44      up  1.00000 1.00000
 45   hdd  14.65039         osd.45      up  1.00000 1.00000
 -9        50.92773     host s3db8
  8   hdd   7.27539         osd.8       up  1.00000 1.00000
 16   hdd   7.27539         osd.16      up  1.00000 1.00000
 31   hdd   7.27539         osd.31      up  1.00000 1.00000
 52   hdd   7.27539         osd.52      up  1.00000 1.00000
 53   hdd   7.27539         osd.53      up  1.00000 1.00000
 54   hdd   7.27539         osd.54      up  1.00000 1.00000
 55   hdd   7.27539         osd.55      up  1.00000 1.00000
-10        50.92773     host s3db9
 56   hdd   7.27539         osd.56      up  1.00000 1.00000
 57   hdd   7.27539         osd.57      up  1.00000 1.00000
 58   hdd   7.27539         osd.58      up  1.00000 1.00000
 59   hdd   7.27539         osd.59      up  1.00000 1.00000
 60   hdd   7.27539         osd.60      up  1.00000 1.00000
 61   hdd   7.27539         osd.61      up  1.00000 1.00000
 62   hdd   7.27539         osd.62      up  1.00000 1.00000

Am Mo., 8. Feb. 2021 um 14:42 Uhr schrieb Eugen Block <eblock@xxxxxx>:

> Can you share 'ceph osd tree'? Are the weights of this OSD
> appropriate? I've seen stuck PGs because of OSD weight imbalance. Is
> the OSD in the correct subtree?
>
>
> Zitat von Boris Behrens <bb@xxxxxxxxx>:
>
> > Hi Eugen,
> >
> > I've set it to 0 but the "degraded objects" count does not go down.
> >
> > Am Mo., 8. Feb. 2021 um 14:23 Uhr schrieb Eugen Block <eblock@xxxxxx>:
> >
> >> Hi,
> >>
> >> one option would be to decrease (or set to 0) the primary-affinity of
> >> osd.14 and see if that brings the pg back.
> >>
> >> Regards,
> >> Eugen
> >>
> >>
> >
> > --
> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
> > groüen Saal.
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>


-- 
Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
groüen Saal.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux