Hi,
I think I need to remove pools cephfs.cephfs.meta and
cephfs.cephfs.data using
cephosdpooldelete{pool-name}[{pool-name}--yes-i-really-really-mean-it]
by the way, as far as I know, deleting pools not allowed by
default, I have to allow it first.
|ceph tell mon.\* injectargs '--mon-allow-pool-delete=true' Thank you! |
yes, looks like deleting those pools would be safe. The injectargs
command should work, although I do it like this:
$ ceph config set mon mon_allow_pool_delete true
And then set it back to false when I'm done and other people have
access to the cluster to minimize the risk. ;-)
Zitat von Dmitry Melekhov <dm@xxxxxxxxxx>:
07.11.2023 19:02, Eugen Block пишет:
Sorry, I didn't really read your message until the end. Apparently,
you only have one cephfs, so the other two pools can probably be
deleted.
Just to get the full picture, can you share these outputs?
ceph orch ls mds
ceph orch ps --daemon_type mds
yes, sure
ceph orch ls mds
NAME PORTS RUNNING REFRESHED AGE PLACEMENT
mds.cephfs 3/3 2m ago 3M count:3
ceph orch ps --daemon_type mds
NAME HOST PORTS STATUS REFRESHED AGE
MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID
mds.cephfs.vmx1.fhqrcn vmx1 running (6h) 75s ago 3M
26.9M - 17.2.7 5dff4047e947 9f58642a2e00
mds.cephfs.vmx2.ynfnqw vmx2 running (6h) 2m ago 3M
26.7M - 17.2.7 5dff4047e947 2e1514e544ed
mds.cephfs.vmx3.ytsueq vmx3 running (6h) 76s ago 3M
23.6M - 17.2.7 5dff4047e947 c9133ecb9d02
You could check the pool stats for those two pools with 'ceph osd
pool stats', if they don't contain any data you probably won't see
any activity.
ceph osd pool stats
pool .mgr id 1
nothing is going on
pool rbdpool id 2
client io 6.3 MiB/s wr, 0 op/s rd, 527 op/s wr
pool cephfs_data id 3
nothing is going on
pool cephfs_metadata id 4
nothing is going on
pool cephfs.cephfs.meta id 5
nothing is going on
pool cephfs.cephfs.data id 6
nothing is going on
currently even real cephs contains almost nothing, there are only
libvirt vm's configs there...
but, cephfs contains only 2 pools:
ceph fs ls
name: cephfs, metadata pool: cephfs_metadata, data pools: [cephfs_data ]
and not cephfs.cephfs.meta and cephfs.cephfs.data ,which have no
application type.
To remove the (correct) service from cephadm you can either remove
it with the dashboard in the "Services" tab or with CLI: 'ceph orch
rm <MDS_SERVICE>'.
There is only mds service for running cephfs and pools
Depending on the output I requested this may not be necessary. But
if it is, this would remove the daemons and the service definition,
then you can go ahead and delete those pools, again either via
dashboard or CLI.
I think I need to remove pools cephfs.cephfs.meta and
cephfs.cephfs.data using
cephosdpooldelete{pool-name}[{pool-name}--yes-i-really-really-mean-it]
by the way, as far as I know, deleting pools not allowed by
default, I have to allow it first.
is this correct way to allow it?
|ceph tell mon.\* injectargs '--mon-allow-pool-delete=true' Thank you! |
Zitat von Dmitry Melekhov <dm@xxxxxxxxxx>:
07.11.2023 16:34, Eugen Block пишет:
Hi,
can you check for existing metadata?
ceph osd pool ls detail --format=json-pretty|grep -A 2
application_metadata
Yes, it is here:
ceph osd pool ls detail --format=json-pretty|grep -A 2 application_metadata
"application_metadata": {
"mgr": {}
}
--
"application_metadata": {
"rbd": {}
}
--
"application_metadata": {
"cephfs": {
"data": "cephfs"
--
"application_metadata": {
"cephfs": {
"metadata": "cephfs"
--
"application_metadata": {}
},
{
--
"application_metadata": {}
}
]
So, yes, application is not set on 2 pools.
Not sure if it's the same issue as described here:
https://lists.ceph.io/hyperkitty/list/ceph-users@xxxxxxx/thread/2NL2Q57HTSGDDBLARLRCVRVX2PE6FKDA/
There the fix was to set the the application metadata like this:
ceph osd pool application set <your metadata pool name> cephfs
metadata <your ceph fs filesystem name>
Yes, I guess it is another way to fix this- set application on pools.
But I don't sure I need to do this, because looks like I don't use
these 2 pools.
I think it is better to delete them if I not need them and if it is safe.
Thank you!
Zitat von Dmitry Melekhov <dm@xxxxxxxxxx>:
Hello!
I'm very new to ceph ,s orry I'm asking extremely basic questions.
I just upgraded 17.2.6 to 17.2.7 and got warning:
2 pool(s) do not have an application enabled
These pools are
5 cephfs.cephfs.meta
6 cephfs.cephfs.data
I don't remember why and how I created them, I just followed
some instruction...
And don't remember their state before upgrade :-(
And I see in dashboard 0 bytes is used in both pools.
But I have two other pools
3 cephfs_data
4 cephfs_metadata
which are in use by cephfs:
ceph fs ls
name: cephfs, metadata pool: cephfs_metadata, data pools: [cephfs_data ]
and really have data in them.
Could you tell me, can I just remove these two pools without
application, if everything works , i.e. cephfs is mounted and
accessible?
Thank you!
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx