why is there heavy read traffic during object delete?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I setup a cephfs file system with a cache tier over an erasure coded tier as an experiment:

  ceph osd erasure-code-profile set raid6 k=4 m=2 
  ceph osd pool create cephfs-metadata 512 512 
  ceph osd pool set cephfs-metadata size 3
  ceph osd pool create cache-data 2048 2048
  ceph osd pool create cephfs-data 256 256 erasure raid6 default_erasure
  ceph osd tier add cephfs-data cache-data
  ceph osd tier cache-mode cache-data writeback
  ceph osd tier set-overlay cephfs-data cache-data
  ceph osd pool set cache-data hit_set_type bloom
  ceph osd pool set cache-data target_max_bytes 1099511627776 

The file system was created from the cephfs-metadata and cephfs-data pools

After adding a lot of data to this and waiting for the pools to idle down and stabilize I removing the file system content with rm. I am seeing very strange behavior, the file system remove was quick, and then it started removing the data from the pools. However it appears to be reading the data from the erasure coded pool and creating empty content in the cache pool.

At its peak capacity the system looked like this:

    NAME                ID     CATEGORY     USED       %USED     MAX AVAIL     OBJECTS     DIRTY     READ      WRITE 
    cache-data          21     -              791G      5.25         6755G      256302      140k     22969     2138k 
    cephfs-data         22     -             4156G     27.54         4503G     1064086     1039k     51271     1046k 


2 hours later it looked like this:

    NAME                ID     CATEGORY     USED       %USED     MAX AVAIL     OBJECTS     DIRTY     READ      WRITE 
    cache-data          21     -              326G      2.17         7576G      702142      559k     22969     2689k 
    cephfs-data         22     -             3964G     26.27         5051G     1014842      991k      476k     1143k 

The object count in the erasure coded pool has gone down a little, the count in the cache pool has gone up a lot, there has been a lot of read activity in the erasure coded pool and write activity into both pools. The used count in the cache pool is also going down. It looks like the cache pool is gaining 9 objects for each one removed from the erasure code pool. Looking at the actual files being created by the OSD for this, they are empty.

What is going on here? It looks like this will take a day or so to complete at this rate of progress.

The ceph version here is the master branch from a couple of days ago.

Thanks

  Steve Lord



----------------------------------------------------------------------
The information contained in this transmission may be confidential. Any disclosure, copying, or further distribution of confidential information is not permitted unless such privilege is explicitly granted in writing by Quantum. Quantum reserves the right to have electronic communications, including email and attachments, sent across its networks filtered through anti virus and spam software programs and retain such messages in order to comply with applicable data security and retention requirements. Quantum is not responsible for the proper and complete transmission of the substance of this communication or for any delay in its receipt.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux