radosgw in Nautilus: message "client_io->complete_request() returned Broken pipe"

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi @ll,

I have a Nautilus Ceph cluster UP with radosgw
in a zonegroup. I'm using the web frontend Beast
(the default in Nautilus). All seems to work fine
but in the log of radosgw I have this message:

    Apr 17 14:02:56 rgw-m-1 ceph-m-rgw.rgw-m-1.rgw0[888]: 2019-04-17 14:02:56.410 7fe659803700  0 ERROR: client_io->complete_request() returned Broken pipe

approximately every ~2-3 minutes (it's an average,
it's random, it's not every 2 minutes exactly).
I think the code which generates this message is
here:

    https://github.com/ceph/ceph/blob/master/src/rgw/rgw_process.cc#L283-L287

but I'm completely unqualified to understand the code.
What is the meaning this error message? Should I worry
about this message?

François (flaf)


PS: just in case, here my conf:

----------------------------------------
~$ cat /etc/ceph/ceph-m.conf
[client.rgw.rgw-m-1.rgw0]
host = rgw-m-1
keyring = /var/lib/ceph/radosgw/ceph-m-rgw.rgw-m-1.rgw0/keyring
log file = /var/log/ceph/ceph-m-rgw-rgw-m-1.rgw0.log
rgw frontends = beast endpoint=192.168.222.1:80
rgw thread pool size = 512

[client.rgw.rgw-m-2.rgw0]
host = rgw-m-2
keyring = /var/lib/ceph/radosgw/ceph-m-rgw.rgw-m-2.rgw0/keyring
log file = /var/log/ceph/ceph-m-rgw-rgw-m-2.rgw0.log
rgw frontends = beast endpoint=192.168.222.2:80
rgw thread pool size = 512

# Please do not change this file directly since it is managed by Ansible and will be overwritten
[global]
cluster network = 10.90.90.0/24
debug_rgw = 0/5
fsid = bb27079f-f116-4440-8a64-9ed430dc17be
log file = /dev/null
mon cluster log file = /dev/null
mon host = [v2:192.168.221.31:3300,v1:192.168.221.31:6789],[v2:192.168.221.32:3300,v1:192.168.221.32:6789],[v2:192.168.221.33:3300,v1:192.168.221.33:6789]
mon_osd_down_out_subtree_limit = host
mon_osd_min_down_reporters = 4
osd_crush_chooseleaf_type = 1
osd_crush_update_on_start = true
osd_pool_default_min_size = 2
osd_pool_default_pg_num = 8
osd_pool_default_ppg_num = 8
osd_pool_default_size = 3
public network = 192.168.221.0/25
rgw_enable_ops_log = true
rgw_log_http_headers = http_x_forwarded_for
rgw_ops_log_socket_path = /var/run/ceph/rgw-opslog.asok
rgw_realm = denmark
rgw_zone = zone-m
rgw_zonegroup = copenhagen
----------------------------------------

Installation via ceph-ansible with a docker deployment version stable 4.0.
ceph_docker_image: v4.0.0-stable-4.0-nautilus-centos-7-x86_64
ceph version 14.2.0 (3a54b2b6d167d4a2a19e003a705696d4fe619afc) nautilus (stable)
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux