Re: How to trigger scrubbing in Ceph on-demand ?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Reto,

Thanks a lot for the instructions. I tried the same, but still couldn't
trigger scrubbing deterministically. The first time I initiated scrubbing,
I saw scrubbing status in ceph -s, but for subsequent times, I didn't see
any scrubbing status. Do you know what might be going on potentially? Any
ideas would be appreciated. Thanks.

Best Regards,
*Jayjeet Chakraborty*
Ph.D. Student
Department of Computer Science and Engineering
University of California, Santa Cruz
*Email: jayjeetc@xxxxxxxx <jayjeetc@xxxxxxxx>*


On Wed, Oct 18, 2023 at 7:47 AM Reto Gysi <rlgysi@xxxxxxxxx> wrote:

> Hi
>
> I haven't updated to reef yet. I've tried this on quincy.
>
> # create a testfile on cephfs.rgysi.data pool
> root@zephir:/home/rgysi/misc# echo cephtest123 > cephtest.txt
>
> #list inode of new file
> root@zephir:/home/rgysi/misc# ls -i cephtest.txt
> 1099518867574 cephtest.txt
>
> convert inode value to hex value
> root@zephir:/home/rgysi/misc# printf "%x" 1099518867574
> 100006e7876
>
> # search for this value in the rados pool cephfs.rgysi.data, to find
> object(s)
> root@zephir:/home/rgysi/misc# rados -p cephfs.rgysi.data ls | grep
> 100006e7876
> 100006e7876.00000000
>
> # find pg for the object
> root@zephir:/home/rgysi/misc# ceph osd map cephfs.rgysi.data
> 100006e7876.00000000
> osdmap e105365 pool 'cephfs.rgysi.data' (25) object '100006e7876.00000000'
> -> pg 25.ee1befa1 (25.1) -> up ([0,2,8], p0) acting ([0,2,8], p0)
>
> #Initiate a deep-scrub for this pg
> root@zephir:/home/rgysi/misc# ceph pg deep-scrub 25.1
> instructing pg 25.1 on osd.0 to deep-scrub
>
> # check status of scrubbing
> root@zephir:/home/rgysi/misc# ceph pg ls scrubbing
> PG    OBJECTS  DEGRADED  MISPLACED  UNFOUND  BYTES        OMAP_BYTES*
>  OMAP_KEYS*  LOG   STATE                        SINCE  VERSION
>         REPORTED        UP         ACTING     SCRUB_STAMP
>                      DEEP_SCRUB_STAMP                 LAST_S
> CRUB_DURATION  SCRUB_SCHEDULING
> 25.1    37774         0          0        0  62869823142            0
>           0  2402  active+clean+scrubbing+deep     7s  105365'1178098
>  105365:8066292  [0,2,8]p0  [0,2,8]p0  2023-10-18T05:17:48.631392+0000
>  2023-10-08T11:30:58.883164+0000
>            3  deep scrubbing for 1s
>
>
> Best Regards,
>
> Reto
>
> Am Mi., 18. Okt. 2023 um 16:24 Uhr schrieb Jayjeet Chakraborty <
> jayjeetc@xxxxxxxx>:
>
>> Hi all,
>>
>> Just checking if someone had a chance to go through the scrub trigger
>> issue
>> above. Thanks.
>>
>> Best Regards,
>> *Jayjeet Chakraborty*
>> Ph.D. Student
>> Department of Computer Science and Engineering
>> University of California, Santa Cruz
>> *Email: jayjeetc@xxxxxxxx <jayjeetc@xxxxxxxx>*
>>
>>
>> On Mon, Oct 16, 2023 at 9:01 PM Jayjeet Chakraborty <jayjeetc@xxxxxxxx>
>> wrote:
>>
>> > Hi all,
>> >
>> > I am trying to trigger deep scrubbing in Ceph reef (18.2.0) on demand
>> on a
>> > set of files that I randomly write to CephFS. I have tried both invoking
>> > deep-scrub on CephFS using ceph tell and just deep scrubbing a
>> > particular PG. Unfortunately, none of that seems to be working for me.
>> I am
>> > monitoring the ceph status output, it never shows any scrubbing
>> > information. Can anyone please help me out on this ? In a nutshell, I
>> need
>> > Ceph to scrub for me anytime I want. I am using Ceph with default
>> configs
>> > for scrubbing. Thanks all.
>> >
>> > Best Regards,
>> > *Jayjeet Chakraborty*
>> > Ph.D. Student
>> > Department of Computer Science and Engineering
>> > University of California, Santa Cruz
>> > *Email: jayjeetc@xxxxxxxx <jayjeetc@xxxxxxxx>*
>> >
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux