Hi Mark, I have noticed exactly the same thing on Nautilus where host didn't work but chassis did work. I posted to this mailing list a few weeks ago. It's very strange that the host filter is not working. I also could not find any errors logged for this, so it looks like it's just ignoring the setting. I don't know if this is still a problem in Octopus or Pacific as I have not had a chance to upgrade our dev cluster yet. My workaround is that I've got a chassis level in my crush above host and the settings which differ are the same across a chassis. This may not suit your situation though. Rich On Fri, 26 Nov 2021 at 15:23, Mark Kirkwood <markkirkwood@xxxxxxxxxxxxxxxx> wrote: > > HI all, > > I'm looking at doing a Luminous to Nautilus upgrade. I'd like to > assimilate the config into the mon db. However we do have hosts with > differing [osd] config sections in their current ceph.conf files. I was > looking at using the crush type host:xxx to set these differently if > required. > > However my test case is not applying the mask to the particular host, e.g: > > markir@ceph2:~$ sudo ceph config set osd/host:ceph2 osd_memory_target > 1073741824 > > markir@ceph2:~$ sudo ceph config dump > WHO MASK LEVEL OPTION > VALUE RO > global advanced > auth_client_required cephx * > global advanced > auth_cluster_required cephx * > global advanced > auth_service_required cephx * > global advanced > cluster_network 192.168.124.0/24 * > global advanced osd_pool_default_size 2 > global advanced > public_network 192.168.123.0/24 * > mon advanced mon_warn_on_insecure_global_id_reclaim false > mon advanced > mon_warn_on_insecure_global_id_reclaim_allowed false > osd host:ceph2 basic osd_memory_target > 1073741824 > > > markir@ceph2:~$ sudo ceph config get osd.1 > WHO MASK LEVEL OPTION VALUE RO > global advanced auth_client_required cephx * > global advanced auth_cluster_required cephx * > global advanced auth_service_required cephx * > global advanced cluster_network 192.168.124.0/24 * > osd host:ceph2 basic osd_memory_target 1073741824 > global advanced osd_pool_default_size 2 > global advanced public_network 192.168.123.0/24 * > > markir@ceph2:~$ sudo ceph config show osd.1 > NAME VALUE SOURCE OVERRIDES IGNORES > auth_client_required cephx mon > auth_cluster_required cephx mon > auth_service_required cephx mon > cluster_network 192.168.124.0/24 mon > daemonize false override > keyring $osd_data/keyring default > leveldb_log default > mon_host 192.168.123.20 file > mon_initial_members ceph0 file > osd_pool_default_size 2 mon > public_network 192.168.123.0/24 mon > rbd_default_features 61 default > setgroup ceph cmdline > setuser ceph cmdline > > If I use a different mask, e.g: osd/class:hdd or even osd/root:default > then the setting *is* applied. I'm scratching my head about this - no > errors in either the mon or osd log to indicate why it is not being applied. > > The is ceph 14.2.22. And osd.1 is really on host ceph2: > > markir@ceph2:~$ sudo ceph osd tree > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > -1 0.11719 root default > -3 0.02930 host ceph1 > 0 hdd 0.02930 osd.0 up 1.00000 1.00000 > -5 0.02930 host ceph2 > 1 hdd 0.02930 osd.1 up 1.00000 1.00000 > -7 0.02930 host ceph3 > 2 hdd 0.02930 osd.2 up 1.00000 1.00000 > -9 0.02930 host ceph4 > 3 hdd 0.02930 osd.3 up 1.00000 1.00000 > > regards > > Mark > > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx