On Sun, Jun 23, 2019 at 4:27 PM Alex Litvak <alexander.v.litvak@xxxxxxxxx> wrote: > > Hello everyone, > > I encounter this in nautilus client and not with mimic. Removing admin socket entry from config on client makes no difference > > Error: > > rbd ls -p one > 2019-06-23 12:58:29.344 7ff2710b0700 -1 set_mon_vals failed to set admin_socket = /var/run/ceph/$name.$pid.asok: Configuration option 'admin_socket' may not be modified at runtime > 2019-06-23 12:58:29.348 7ff2708af700 -1 set_mon_vals failed to set admin_socket = /var/run/ceph/$name.$pid.asok: Configuration option 'admin_socket' may not be modified at runtime > > I have no issues running other ceph clients (no messages on the screen with ceph -s or ceph iostat from the same box.) > I connected to a few other client nodes and as root I can do the same string > rbd ls -p one > > > On all the nodes with user libvirt I have seen the admin_socket messages > > oneadmin@virt3n1-la:~$ rbd ls -p one --id libvirt > 2019-06-23 13:16:41.626 7f9ea0ff9700 -1 set_mon_vals failed to set admin_socket = /var/run/ceph/$name.$pid.asok: Configuration option 'admin_socket' may not be modified at runtime > 2019-06-23 13:16:41.626 7f9e8bfff700 -1 set_mon_vals failed to set admin_socket = /var/run/ceph/$name.$pid.asok: Configuration option 'admin_socket' may not be modified at runtime > > I can execute all rbd operations on the cluster from client otherwise. Commenting client in config file makes no difference > > This is an optimiised config distributed across the clients it is almost the same as on servers (no libvirt on servers) > > [client] > admin_socket = /var/run/ceph/$name.$pid.asok > > [client.libvirt] > admin socket = /var/run/ceph/$cluster-$type.$id.$pid.$cctid.asok # must be writable by QEMU and allowed by SELinux or AppArmor > log file = /var/log/ceph/qemu-guest-$pid.log # must be writable by QEMU and allowed by SELinux or AppArmor > > # Please do not change this file directly since it is managed by Ansible and will be overwritten > [global] > cluster network = 10.0.42.0/23 > fsid = 3947ba2d-1b01-4909-8e3a-f9714f427483 > log file = /dev/null > mon cluster log file = /dev/null > mon host = [v2:10.0.40.121:3300,v1:10.0.40.121:6789],[v2:10.0.40.122:3300,v1:10.0.40.122:6789],[v2:10.0.40.123:3300,v1:10.0.40.123:6789] > perf = True > public network = 10.0.40.0/23 > rocksdb_perf = True > > > Here is config from mon > > NAME VALUE SOURCE OVERRIDES IGNORES > cluster_network 10.0.42.0/23 file (mon[10.0.42.0/23]) > daemonize false override > debug_asok 0/0 mon > debug_auth 0/0 mon > debug_bdev 0/0 mon > debug_bluefs 0/0 mon > debug_bluestore 0/0 mon > debug_buffer 0/0 mon > debug_civetweb 0/0 mon > debug_client 0/0 mon > debug_compressor 0/0 mon > debug_context 0/0 mon > debug_crush 0/0 mon > debug_crypto 0/0 mon > debug_dpdk 0/0 mon > debug_eventtrace 0/0 mon > debug_filer 0/0 mon > debug_filestore 0/0 mon > debug_finisher 0/0 mon > debug_fuse 0/0 mon > debug_heartbeatmap 0/0 mon > debug_javaclient 0/0 mon > debug_journal 0/0 mon > debug_journaler 0/0 mon > debug_kinetic 0/0 mon > debug_kstore 0/0 mon > debug_leveldb 0/0 mon > debug_lockdep 0/0 mon > debug_mds 0/0 mon > debug_mds_balancer 0/0 mon > debug_mds_locker 0/0 mon > debug_mds_log 0/0 mon > debug_mds_log_expire 0/0 mon > debug_mds_migrator 0/0 mon > debug_memdb 0/0 mon > debug_mgr 0/0 mon > debug_mgrc 0/0 mon > debug_mon 0/0 mon > debug_monc 0/00 mon > debug_ms 0/0 mon > debug_none 0/0 mon > debug_objclass 0/0 mon > debug_objectcacher 0/0 mon > debug_objecter 0/0 mon > debug_optracker 0/0 mon > debug_osd 0/0 mon > debug_paxos 0/0 mon > debug_perfcounter 0/0 mon > debug_rados 0/0 mon > debug_rbd 0/0 mon > debug_rbd_mirror 0/0 mon > debug_rbd_replay 0/0 mon > debug_refs 0/0 mon > err_to_stderr true override > keyring $mon_data/keyring default > leveldb_block_size 65536 default > leveldb_cache_size 536870912 default > leveldb_compression false default > leveldb_log default > leveldb_write_buffer_size 33554432 default > log_file override file[/dev/null],mon[/dev/null] > log_stderr_prefix debug cmdline > log_to_stderr true override > log_to_syslog false override > mon_allow_pool_delete true mon > mon_cluster_log_file /dev/null file (mon[/dev/null]) > mon_cluster_log_to_stderr true cmdline > mon_data /var/lib/ceph/mon/ceph-storage2n2-la cmdline > mon_host [v2:10.0.40.121:3300,v1:10.0.40.121:6789],[v2:10.0.40.122:3300,v1:10.0.40.122:6789],[v2:10.0.40.123:3300,v1:10.0.40.123:6789] file > mon_initial_members storage2n1-la,storage2n2-la,storage2n3-la file > mon_osd_down_out_interval 300 mon > osd_pool_default_crush_rule -1 file (mon[-1]) > osd_scrub_begin_hour 19 mon > osd_scrub_end_hour 4 mon > osd_scrub_load_threshold 0.010000 mon > osd_scrub_sleep 0.100000 mon > perf true file (mon[true]) > public_addr v2:10.0.40.122:0/0 cmdline > public_network 10.0.40.0/23 file (mon[10.0.40.0/23]) > rbd_default_features 61 default > rocksdb_perf true file (mon[true]) > setgroup ceph cmdline > setuser ceph cmdline What's the mon config for the "client.admin" user? "ceph config get client.admin" > > I am not sure why I am getting this messages and why are they inconsistent across the nodes. For example I am not getting those when I execute rbd in containers running ceph daemons on server > cluster nodes. Any clue would be appreciated. > > > > > > > > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Jason _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com