Re: Frequest LARGE_OMAP_OBJECTS in cephfs metadata pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Patrick, 

File system created around 4 months back. Using ceph version 14.2.3 version. 

[root@knode25 /]# ceph fs dump
dumped fsmap epoch 577
e577
enable_multiple, ever_enabled_multiple: 0,0
compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2}
legacy client fscid: 1

Filesystem 'cephfs01' (1)
fs_name	cephfs01
epoch	577
flags	32
created	2019-10-18 23:59:29.610249
modified	2020-02-22 03:13:09.425905
tableserver	0
root	0
session_timeout	60
session_autoclose	300
max_file_size	1099511627776
min_compat_client	-1 (unspecified)
last_failure	0
last_failure_osd_epoch	1608
compat	compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2}
max_mds	1
in	0
up	{0=2981519}
failed
damaged
stopped
data_pools	[2]
metadata_pool	1
inline_data	disabled
balancer
standby_count_wanted	1
2981519:	[v2:10.131.16.30:6808/3209191719,v1:10.131.16.30:6809/3209191719] 'cephfs01-b' mds.0.572 up:active seq 22141
2998684:	[v2:10.131.16.89:6832/54557615,v1:10.131.16.89:6833/54557615] 'cephfs01-a' mds.0.0 up:standby-replay seq 2


[root@knode25 /]# ceph fs status
cephfs01 - 290 clients
========
+------+----------------+------------+---------------+-------+-------+
| Rank |     State      |    MDS     |    Activity   |  dns  |  inos |
+------+----------------+------------+---------------+-------+-------+
|  0   |     active     | cephfs01-b | Reqs:  333 /s | 2738k | 2735k |
| 0-s  | standby-replay | cephfs01-a | Evts:  795 /s | 1368k | 1363k |
+------+----------------+------------+---------------+-------+-------+
+-------------------+----------+-------+-------+
|        Pool       |   type   |  used | avail |
+-------------------+----------+-------+-------+
| cephfs01-metadata | metadata | 2193M | 78.1T |
|   cephfs01-data0  |   data   |  753G | 78.1T |
+-------------------+----------+-------+-------+
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux