Re: one PG is stuck on single OSD and does not sync back

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I've outed osd.18 and osd.54 and let it sync for some time and now the
problem is gone.
*shrugs

Thank you for the hints.

Am Mo., 8. Feb. 2021 um 14:46 Uhr schrieb Boris Behrens <bb@xxxxxxxxx>:

> Hi,
> sure
>
> ID  CLASS WEIGHT    TYPE NAME       STATUS REWEIGHT PRI-AFF
>  -1       672.68457 root default
>  -2        58.20561     host s3db1
>  23   hdd  14.55269         osd.23      up  1.00000 1.00000
>  69   hdd  14.55269         osd.69      up  1.00000 1.00000
>  73   hdd  14.55269         osd.73      up  1.00000 1.00000
>  79   hdd   3.63689         osd.79      up  1.00000 1.00000
>  80   hdd   3.63689         osd.80      up  1.00000 1.00000
>  81   hdd   3.63689         osd.81      up  1.00000 1.00000
>  82   hdd   3.63689         osd.82      up  1.00000 1.00000
> -11        50.94173     host s3db10
>  63   hdd   7.27739         osd.63      up  1.00000 1.00000
>  64   hdd   7.27739         osd.64      up  1.00000 1.00000
>  65   hdd   7.27739         osd.65      up  1.00000 1.00000
>  66   hdd   7.27739         osd.66      up  1.00000 1.00000
>  67   hdd   7.27739         osd.67      up  1.00000 1.00000
>  68   hdd   7.27739         osd.68      up  1.00000 1.00000
>  70   hdd   7.27739         osd.70      up  1.00000 1.00000
> -12        50.94173     host s3db11
>  46   hdd   7.27739         osd.46      up  1.00000 1.00000
>  47   hdd   7.27739         osd.47      up  1.00000 1.00000
>  48   hdd   7.27739         osd.48      up  1.00000 1.00000
>  49   hdd   7.27739         osd.49      up  1.00000 1.00000
>  50   hdd   7.27739         osd.50      up  1.00000 1.00000
>  51   hdd   7.27739         osd.51      up  1.00000 1.00000
>  72   hdd   7.27739         osd.72      up  1.00000 1.00000
> -37        58.55478     host s3db12
>  19   hdd   3.68750         osd.19      up  1.00000 1.00000
>  71   hdd   3.68750         osd.71      up  1.00000 1.00000
>  75   hdd   3.68750         osd.75      up  1.00000 1.00000
>  76   hdd   3.68750         osd.76      up  1.00000 1.00000
>  77   hdd  14.60159         osd.77      up  1.00000 1.00000
>  78   hdd  14.60159         osd.78      up  1.00000 1.00000
>  83   hdd  14.60159         osd.83      up  1.00000 1.00000
>  -3        58.49872     host s3db2
>   1   hdd  14.65039         osd.1       up  1.00000 1.00000
>   3   hdd   3.63689         osd.3       up  1.00000 1.00000
>   4   hdd   3.63689         osd.4       up  1.00000 1.00000
>   5   hdd   3.63689         osd.5       up  1.00000 1.00000
>   6   hdd   3.63689         osd.6       up  1.00000 1.00000
>   7   hdd  14.65039         osd.7       up  1.00000 1.00000
>  74   hdd  14.65039         osd.74      up  1.00000 1.00000
>  -4        58.49872     host s3db3
>   2   hdd  14.65039         osd.2       up  1.00000 1.00000
>   9   hdd  14.65039         osd.9       up  1.00000 1.00000
>  10   hdd  14.65039         osd.10      up  1.00000 1.00000
>  12   hdd   3.63689         osd.12      up  1.00000 1.00000
>  13   hdd   3.63689         osd.13      up  1.00000 1.00000
>  14   hdd   3.63689         osd.14      up  1.00000       0
>  15   hdd   3.63689         osd.15      up  1.00000 1.00000
>  -5        58.49872     host s3db4
>  11   hdd  14.65039         osd.11      up  1.00000 1.00000
>  17   hdd  14.65039         osd.17      up  1.00000 1.00000
>  18   hdd  14.65039         osd.18      up  1.00000 1.00000
>  20   hdd   3.63689         osd.20      up  1.00000 1.00000
>  21   hdd   3.63689         osd.21      up  1.00000 1.00000
>  22   hdd   3.63689         osd.22      up  1.00000 1.00000
>  24   hdd   3.63689         osd.24      up  1.00000 1.00000
>  -6        58.89636     host s3db5
>   0   hdd   3.73630         osd.0       up  1.00000 1.00000
>  25   hdd   3.73630         osd.25      up  1.00000 1.00000
>  26   hdd   3.73630         osd.26      up  1.00000 1.00000
>  27   hdd   3.73630         osd.27      up  1.00000 1.00000
>  28   hdd  14.65039         osd.28      up  1.00000 1.00000
>  29   hdd  14.65039         osd.29      up  1.00000 1.00000
>  30   hdd  14.65039         osd.30      up  1.00000 1.00000
>  -7        58.89636     host s3db6
>  32   hdd   3.73630         osd.32      up  1.00000 1.00000
>  33   hdd   3.73630         osd.33      up  1.00000 1.00000
>  34   hdd   3.73630         osd.34      up  1.00000 1.00000
>  35   hdd   3.73630         osd.35      up  1.00000 1.00000
>  36   hdd  14.65039         osd.36      up  1.00000 1.00000
>  37   hdd  14.65039         osd.37      up  1.00000 1.00000
>  38   hdd  14.65039         osd.38      up  1.00000 1.00000
>  -8        58.89636     host s3db7
>  39   hdd   3.73630         osd.39      up  1.00000 1.00000
>  40   hdd   3.73630         osd.40      up  1.00000 1.00000
>  41   hdd   3.73630         osd.41      up  1.00000 1.00000
>  42   hdd   3.73630         osd.42      up  1.00000 1.00000
>  43   hdd  14.65039         osd.43      up  1.00000 1.00000
>  44   hdd  14.65039         osd.44      up  1.00000 1.00000
>  45   hdd  14.65039         osd.45      up  1.00000 1.00000
>  -9        50.92773     host s3db8
>   8   hdd   7.27539         osd.8       up  1.00000 1.00000
>  16   hdd   7.27539         osd.16      up  1.00000 1.00000
>  31   hdd   7.27539         osd.31      up  1.00000 1.00000
>  52   hdd   7.27539         osd.52      up  1.00000 1.00000
>  53   hdd   7.27539         osd.53      up  1.00000 1.00000
>  54   hdd   7.27539         osd.54      up  1.00000 1.00000
>  55   hdd   7.27539         osd.55      up  1.00000 1.00000
> -10        50.92773     host s3db9
>  56   hdd   7.27539         osd.56      up  1.00000 1.00000
>  57   hdd   7.27539         osd.57      up  1.00000 1.00000
>  58   hdd   7.27539         osd.58      up  1.00000 1.00000
>  59   hdd   7.27539         osd.59      up  1.00000 1.00000
>  60   hdd   7.27539         osd.60      up  1.00000 1.00000
>  61   hdd   7.27539         osd.61      up  1.00000 1.00000
>  62   hdd   7.27539         osd.62      up  1.00000 1.00000
>
> Am Mo., 8. Feb. 2021 um 14:42 Uhr schrieb Eugen Block <eblock@xxxxxx>:
>
>> Can you share 'ceph osd tree'? Are the weights of this OSD
>> appropriate? I've seen stuck PGs because of OSD weight imbalance. Is
>> the OSD in the correct subtree?
>>
>>
>> Zitat von Boris Behrens <bb@xxxxxxxxx>:
>>
>> > Hi Eugen,
>> >
>> > I've set it to 0 but the "degraded objects" count does not go down.
>> >
>> > Am Mo., 8. Feb. 2021 um 14:23 Uhr schrieb Eugen Block <eblock@xxxxxx>:
>> >
>> >> Hi,
>> >>
>> >> one option would be to decrease (or set to 0) the primary-affinity of
>> >> osd.14 and see if that brings the pg back.
>> >>
>> >> Regards,
>> >> Eugen
>> >>
>> >>
>> >
>> > --
>> > Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
>> > groüen Saal.
>> > _______________________________________________
>> > ceph-users mailing list -- ceph-users@xxxxxxx
>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
>>
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
>
>
> --
> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
> groüen Saal.
>


-- 
Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
groüen Saal.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux