Re: PG active+clean+remapped status

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Cary,

No, I didnt try to repair it.
I am comparatively new in ceph. Is it okay to try to repair it ?
Or should I take any precautions while doing it ? 

Karun Josy

On Sat, Dec 16, 2017 at 2:08 PM, Cary <dynamic.cary@xxxxxxxxx> wrote:
Karun,

 Did you attempt a "ceph pg repair <pgid>"? Replace <pgid> with the pg
ID that needs repaired, 3.4.

Cary
-D123

On Sat, Dec 16, 2017 at 8:24 AM, Karun Josy <karunjosy1@xxxxxxxxx> wrote:
> Hello,
>
> I added 1 disk to the cluster and after rebalancing, it shows 1 PG is in
> remapped state. How can I correct it ?
>
> (I had to restart some osds during the rebalancing as there were some slow
> requests)
>
> $ ceph pg dump | grep remapped
> dumped all
> 3.4         981                  0        0         0       0 2655009792
> 1535     1535 active+clean+remapped 2017-12-15 22:07:21.663964  2824'785115
> 2824:2297888 [0,2,NONE,4,12,10,5,1]          0   [0,2,1,4,12,10,5,1]
> 0  2288'767367 2017-12-14 11:00:15.576741      417'518549 2017-12-08
> 03:56:14.006982
>
> That PG belongs to an erasure pool with k=5, m =3 profile, failure domain is
> host.
>
> ===========
>
> $ ceph osd tree
> ID  CLASS WEIGHT   TYPE NAME                STATUS REWEIGHT PRI-AFF
>  -1       16.94565 root default
>  -3        2.73788     host ceph-a1
>   0   ssd  1.86469         osd.0                up  1.00000 1.00000
>  14   ssd  0.87320         osd.14               up  1.00000 1.00000
>  -5        2.73788     host ceph-a2
>   1   ssd  1.86469         osd.1                up  1.00000 1.00000
>  15   ssd  0.87320         osd.15               up  1.00000 1.00000
>  -7        1.86469     host ceph-a3
>   2   ssd  1.86469         osd.2                up  1.00000 1.00000
>  -9        1.74640     host ceph-a4
>   3   ssd  0.87320         osd.3                up  1.00000 1.00000
>   4   ssd  0.87320         osd.4                up  1.00000 1.00000
> -11        1.74640     host ceph-a5
>   5   ssd  0.87320         osd.5                up  1.00000 1.00000
>   6   ssd  0.87320         osd.6                up  1.00000 1.00000
> -13        1.74640     host ceph-a6
>   7   ssd  0.87320         osd.7                up  1.00000 1.00000
>   8   ssd  0.87320         osd.8                up  1.00000 1.00000
> -15        1.74640     host ceph-a7
>   9   ssd  0.87320         osd.9                up  1.00000 1.00000
>  10   ssd  0.87320         osd.10               up  1.00000 1.00000
> -17        2.61960     host ceph-a8
>  11   ssd  0.87320         osd.11               up  1.00000 1.00000
>  12   ssd  0.87320         osd.12               up  1.00000 1.00000
>  13   ssd  0.87320         osd.13               up  1.00000 1.00000
>
>
>
> Karun
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux