How to identify the index pool real usage?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

Today we had a big issue with slow ops on the nvme drives which holding the index pool.

Why the nvme shows full if on ceph is barely utilized? Which one I should belive?

When I check the ceph osd df it shows 10% usage of the osds (1x 2TB nvme drive has 4x osds on it):

ID   CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     META      AVAIL    %USE   VAR   PGS  STATUS
195   nvme  0.43660   1.00000  447 GiB   47 GiB  161 MiB   46 GiB   656 MiB  400 GiB  10.47  0.21   64      up
252   nvme  0.43660   1.00000  447 GiB   46 GiB  161 MiB   45 GiB   845 MiB  401 GiB  10.35  0.21   64      up
253   nvme  0.43660   1.00000  447 GiB   46 GiB  229 MiB   45 GiB   662 MiB  401 GiB  10.26  0.21   66      up
254   nvme  0.43660   1.00000  447 GiB   46 GiB  161 MiB   44 GiB   1.3 GiB  401 GiB  10.26  0.21   65      up
255   nvme  0.43660   1.00000  447 GiB   47 GiB  161 MiB   46 GiB   1.2 GiB  400 GiB  10.58  0.21   64      up
288   nvme  0.43660   1.00000  447 GiB   46 GiB  161 MiB   44 GiB   1.2 GiB  401 GiB  10.25  0.21   64      up
289   nvme  0.43660   1.00000  447 GiB   46 GiB  161 MiB   45 GiB   641 MiB  401 GiB  10.33  0.21   64      up
290   nvme  0.43660   1.00000  447 GiB   45 GiB  229 MiB   44 GiB   668 MiB  402 GiB  10.14  0.21   65      up

However in nvme list it says full:
Node             SN                   Model                                    Namespace Usage                      Format           FW Rev
---------------- -------------------- ---------------------------------------- --------- -------------------------- ---------------- --------
/dev/nvme0n1     90D0A00XTXTR         KCD6XLUL1T92                             1           1.92  TB /   1.92  TB    512   B +  0 B   GPK6
/dev/nvme1n1     60P0A003TXTR         KCD6XLUL1T92                             1           1.92  TB /   1.92  TB    512   B +  0 B   GPK6

With some other node the test was like:

  *   if none of the disk full, no slow ops.
  *   If 1x disk full and the other not, has slow ops but not too much
  *   if none of the disk full, no slow ops.

The full disks are very highly utilized during recovery and they are holding back the operations from the other nvmes.

What's the reason that even if the pgs are the same in the cluster +/-1 regarding space they are not equally utilized.

Thank you



________________________________
This message is confidential and is for the sole use of the intended recipient(s). It may also be privileged or otherwise protected by copyright or other legal rules. If you have received it by mistake please let us know by reply email and delete it from your system. It is prohibited to copy this message or disclose its content to anyone. Any confidentiality or privilege is not waived or lost by any mistaken delivery or unauthorized disclosure of the message. All messages sent to and from Agoda may be monitored to ensure compliance with company policies, to protect the company's interests and to remove potential malware. Electronic messages may be intercepted, amended, lost or deleted, or contain viruses.
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux