On Wed, Sep 28, 2016 at 10:32 AM, Iban Cabrillo <cabrillo@xxxxxxxxxxxxxx> wrote: > Dear Admins, > During last day I have been trying to deploy a new radosgw, following > jewel guide, ceph cluster is healthy (3 mon and 2 osd servers ) > root@cephrgw ceph]# ceph -v > ceph version 10.2.3 (ecc23778eb545d8dd55e2e4735b53cc93f92e65b) > [root@cephrgw ceph]# rpm -qa | grep ceph > ceph-common-10.2.3-0.el7.x86_64 > libcephfs1-10.2.3-0.el7.x86_64 > ceph-deploy-1.5.36-0.noarch > ceph-release-1-1.el7.noarch > ceph-base-10.2.3-0.el7.x86_64 > ceph-radosgw-10.2.3-0.el7.x86_64 > python-cephfs-10.2.3-0.el7.x86_64 > ceph-selinux-10.2.3-0.el7.x86_64 > > Cityweb is running on default port: > [root@cephrgw ceph]# systemctl status ceph-radosgw@rgw.cephrgw.service > ● ceph-radosgw@rgw.cephrgw.service - Ceph rados gateway > Loaded: loaded (/usr/lib/systemd/system/ceph-radosgw@.service; enabled; > vendor preset: disabled) > Active: active (running) since mié 2016-09-28 10:20:34 CEST; 2s ago > Main PID: 29311 (radosgw) > CGroup: > /system.slice/system-ceph\x2dradosgw.slice/ceph-radosgw@rgw.cephrgw.service > └─29311 /usr/bin/radosgw -f --cluster ceph --name > client.rgw.cephrgw --setuser ceph --setgroup ceph > > sep 28 10:20:34 cephrgw.ifca.es systemd[1]: Started Ceph rados gateway. > sep 28 10:20:34 cephrgw.ifca.es systemd[1]: Starting Ceph rados gateway... > > And this pools were created on ceph storage: > .rgw.root 2 4 0 0 > 0 252 195 4 5 > default.rgw.control 0 8 0 0 > 0 0 0 0 0 > default.rgw.data.root 0 0 0 0 > 0 0 0 0 0 > default.rgw.gc 0 32 0 0 > 0 2112 2080 1408 0 > default.rgw.log 0 127 0 0 > 0 47625 47498 31750 0 > default.rgw.users.uid 0 0 0 0 > 0 0 0 0 0 > > But seems that zones are not well defined. > > radosgw-admin zone get --zone-id=default > 2016-09-28 10:24:07.142478 7fd810b219c0 0 failed reading obj info from > .rgw.root:zone_info.default: (2) No such file or directory > The zone id is not default this is the name try: radosgw-admin zone get --rgw-zone default > > [root@cephrgw ~]# radosgw-admin zone get > 2016-09-28 10:25:41.740162 7f18072799c0 1 -- :/0 messenger.start > 2016-09-28 10:25:41.741262 7f18072799c0 1 -- :/945549824 --> > 10.10.3.3:6789/0 -- auth(proto 0 30 bytes epoch 0) v1 -- ?+0 0x7f18085c9850 > con 0x7f18085c85a0 > 2016-09-28 10:25:41.742048 7f180726f700 1 -- 10.10.3.4:0/945549824 learned > my addr 10.10.3.4:0/945549824 > 2016-09-28 10:25:41.743168 7f17eae03700 1 -- 10.10.3.4:0/945549824 <== > mon.2 10.10.3.3:6789/0 1 ==== mon_map magic: 0 v1 ==== 495+0+0 (2693174994 0 > 0) 0x7f17d4000b90 con 0x7f18085c85a0 > 2016-09-28 10:25:41.743380 7f17eae03700 1 -- 10.10.3.4:0/945549824 <== > mon.2 10.10.3.3:6789/0 2 ==== auth_reply(proto 2 0 (0) Success) v1 ==== > 33+0+0 (3801669063 0 0) 0x7f17d4001010 con 0x7f18085c85a0 > 2016-09-28 10:25:41.743696 7f17eae03700 1 -- 10.10.3.4:0/945549824 --> > 10.10.3.3:6789/0 -- auth(proto 2 32 bytes epoch 0) v1 -- ?+0 0x7f17e0001730 > con 0x7f18085c85a0 > 2016-09-28 10:25:41.744541 7f17eae03700 1 -- 10.10.3.4:0/945549824 <== > mon.2 10.10.3.3:6789/0 3 ==== auth_reply(proto 2 0 (0) Success) v1 ==== > 206+0+0 (1705741500 0 0) 0x7f17d4001010 con 0x7f18085c85a0 > 2016-09-28 10:25:41.744765 7f17eae03700 1 -- 10.10.3.4:0/945549824 --> > 10.10.3.3:6789/0 -- auth(proto 2 165 bytes epoch 0) v1 -- ?+0 0x7f17e0001bf0 > con 0x7f18085c85a0 > 2016-09-28 10:25:41.745619 7f17eae03700 1 -- 10.10.3.4:0/945549824 <== > mon.2 10.10.3.3:6789/0 4 ==== auth_reply(proto 2 0 (0) Success) v1 ==== > 393+0+0 (482591267 0 0) 0x7f17d40008c0 con 0x7f18085c85a0 > 2016-09-28 10:25:41.745783 7f17eae03700 1 -- 10.10.3.4:0/945549824 --> > 10.10.3.3:6789/0 -- mon_subscribe({monmap=0+}) v2 -- ?+0 0x7f18085cd560 con > 0x7f18085c85a0 > 2016-09-28 10:25:41.745967 7f18072799c0 1 -- 10.10.3.4:0/945549824 --> > 10.10.3.3:6789/0 -- mon_subscribe({osdmap=0}) v2 -- ?+0 0x7f18085c9850 con > 0x7f18085c85a0 > 2016-09-28 10:25:41.746521 7f17eae03700 1 -- 10.10.3.4:0/945549824 <== > mon.2 10.10.3.3:6789/0 5 ==== mon_map magic: 0 v1 ==== 495+0+0 (2693174994 0 > 0) 0x7f17d40012b0 con 0x7f18085c85a0 > 2016-09-28 10:25:41.746669 7f17daffd700 2 > RGWDataChangesLog::ChangesRenewThread: start > 2016-09-28 10:25:41.746882 7f18072799c0 20 get_system_obj_state: > rctx=0x7ffe0afd57e0 obj=.rgw.root:default.realm state=0x7f18085cf4e8 > s->prefetch_data=0 > 2016-09-28 10:25:41.746962 7f17eae03700 1 -- 10.10.3.4:0/945549824 <== > mon.2 10.10.3.3:6789/0 6 ==== osd_map(5792..5792 src has 5225..5792) v3 ==== > 13145+0+0 (1223904398 0 0) 0x7f17d40008c0 con 0x7f18085c85a0 > 2016-09-28 10:25:41.747661 7f18072799c0 1 -- 10.10.3.4:0/945549824 --> > 10.10.3.12:6810/8166 -- osd_op(client.2974205.0:1 26.85fca992 default.realm > [getxattrs,stat] snapc 0=[] ack+read+known_if_redirected e5792) v7 -- ?+0 > 0x7f18085d3410 con 0x7f18085d1eb0 > 2016-09-28 10:25:41.749762 7f17e83fc700 1 -- 10.10.3.4:0/945549824 <== > osd.10 10.10.3.12:6810/8166 1 ==== osd_op_reply(1 default.realm > [getxattrs,stat] v0'0 uv0 ack = -2 ((2) No such file or directory)) v7 ==== > 175+0+0 (2651345336 0 0) 0x7f17cc000a50 con 0x7f18085d1eb0 > unable to initialize zone: (2) No such file or directory > 2016-09-28 10:25:41.749933 7f18072799c0 10 could not read realm id: (2) No > such file or directory > 2016-09-28 10:25:41.750290 7f18072799c0 1 -- 10.10.3.4:0/945549824 > mark_down 0x7f18085d1eb0 -- 0x7f18085d0ba0 > 2016-09-28 10:25:41.750375 7f18072799c0 1 -- 10.10.3.4:0/945549824 > mark_down 0x7f18085c85a0 -- 0x7f18085c72e0 > 2016-09-28 10:25:41.750575 7f18072799c0 1 -- 10.10.3.4:0/945549824 > mark_down_all > 2016-09-28 10:25:41.750888 7f18072799c0 1 -- 10.10.3.4:0/945549824 shutdown > complete. > > > > this is my conf: > > [global] > fsid = 6f5a65a7-316c-4825-afcb-428608941dd1 > mon_initial_members = cephadm, cephmon02, cephmon03 > mon_host = 10.10.3.1,10.10.3.2,10.10.3.3 > auth_cluster_required = cephx > auth_service_required = cephx > auth_client_required = cephx > filestore_xattr_use_omap = true > osd_pool_default_size = 2 > public_network = 10.10.0.0/16 > cluster_network = 192.168.254.0/27 > debug ms = 1 > debug rgw = 20 > > [osd] > osd_journal_size = 20000 > > [client.cinder] > keyring = /etc/ceph/ceph.client.cinder.keyring > > [client.rgw.cephrgw] > host = cephrwg > keyring = /etc/ceph/ceph.client.rgw.cephrgw.keyring > log_file = /var/log/ceph/client.rgw.cephrgw.log > > Auth list: > client.bootstrap-rgw > key: AQALRwFWPBPDIxAABXXXXcwUsGPn0fMhhP19wg== > caps: [mon] allow profile bootstrap-rgw > > client.rgw.cephrgw > key: AQAfsWNXUIv9MBAAXXXXzGk4strXE0UbWW4yzg== > caps: [mon] allow rw > caps: [osd] allow rwx > > [root@cephrgw ~]# cat /etc/ceph/ceph.client.rgw.cephrgw.keyring > [client.rgw.cephrgw] > key = AQAfsWNXUIv9MBAAXXXXzGk4strXE0UbWW4yzg== > > Any Idea? > > regards, I > > > -- > ############################################################################ > Iban Cabrillo Bartolome > Instituto de Fisica de Cantabria (IFCA) > Santander, Spain > Tel: +34942200969 > PGP PUBLIC KEY: > http://pgp.mit.edu/pks/lookup?op=get&search=0xD9DF0B3D6C8C08AC > ############################################################################ > Bertrand Russell: > "El problema con el mundo es que los estúpidos están seguros de todo y los > inteligentes están llenos de dudas" > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com