Troubles seting up radosgw

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Dear Admins,
   During last day I have been trying to deploy a new radosgw, following jewel guide, ceph cluster is healthy (3 mon and 2 osd servers )
   root@cephrgw ceph]# ceph -v
ceph version 10.2.3 (ecc23778eb545d8dd55e2e4735b53cc93f92e65b)
   [root@cephrgw ceph]# rpm -qa | grep ceph
ceph-common-10.2.3-0.el7.x86_64
libcephfs1-10.2.3-0.el7.x86_64
ceph-deploy-1.5.36-0.noarch
ceph-release-1-1.el7.noarch
ceph-base-10.2.3-0.el7.x86_64
ceph-radosgw-10.2.3-0.el7.x86_64
python-cephfs-10.2.3-0.el7.x86_64
ceph-selinux-10.2.3-0.el7.x86_64

Cityweb is running on default port:
[root@cephrgw ceph]# systemctl status ceph-radosgw@rgw.cephrgw.service
● ceph-radosgw@rgw.cephrgw.service - Ceph rados gateway
   Loaded: loaded (/usr/lib/systemd/system/ceph-radosgw@.service; enabled; vendor preset: disabled)
   Active: active (running) since mié 2016-09-28 10:20:34 CEST; 2s ago
 Main PID: 29311 (radosgw)
   CGroup: /system.slice/system-ceph\x2dradosgw.slice/ceph-radosgw@rgw.cephrgw.service
           └─29311 /usr/bin/radosgw -f --cluster ceph --name client.rgw.cephrgw --setuser ceph --setgroup ceph

sep 28 10:20:34 cephrgw.ifca.es systemd[1]: Started Ceph rados gateway.
sep 28 10:20:34 cephrgw.ifca.es systemd[1]: Starting Ceph rados gateway...

And this pools were created on ceph storage:
.rgw.root                  2            4            0            0            0          252          195            4            5
default.rgw.control            0            8            0            0            0            0            0            0            0
default.rgw.data.root            0            0            0            0            0            0            0            0            0
default.rgw.gc             0           32            0            0            0         2112         2080         1408            0
default.rgw.log            0          127            0            0            0        47625        47498        31750            0
default.rgw.users.uid            0            0            0            0            0            0            0            0            0

But seems that zones are not well defined.

radosgw-admin zone get --zone-id=default
2016-09-28 10:24:07.142478 7fd810b219c0  0 failed reading obj info from .rgw.root:zone_info.default: (2) No such file or directory


[root@cephrgw ~]# radosgw-admin zone get
2016-09-28 10:25:41.740162 7f18072799c0  1 -- :/0 messenger.start
2016-09-28 10:25:41.741262 7f18072799c0  1 -- :/945549824 --> 10.10.3.3:6789/0 -- auth(proto 0 30 bytes epoch 0) v1 -- ?+0 0x7f18085c9850 con 0x7f18085c85a0
2016-09-28 10:25:41.742048 7f180726f700  1 -- 10.10.3.4:0/945549824 learned my addr 10.10.3.4:0/945549824
2016-09-28 10:25:41.743168 7f17eae03700  1 -- 10.10.3.4:0/945549824 <== mon.2 10.10.3.3:6789/0 1 ==== mon_map magic: 0 v1 ==== 495+0+0 (2693174994 0 0) 0x7f17d4000b90 con 0x7f18085c85a0
2016-09-28 10:25:41.743380 7f17eae03700  1 -- 10.10.3.4:0/945549824 <== mon.2 10.10.3.3:6789/0 2 ==== auth_reply(proto 2 0 (0) Success) v1 ==== 33+0+0 (3801669063 0 0) 0x7f17d4001010 con 0x7f18085c85a0
2016-09-28 10:25:41.743696 7f17eae03700  1 -- 10.10.3.4:0/945549824 --> 10.10.3.3:6789/0 -- auth(proto 2 32 bytes epoch 0) v1 -- ?+0 0x7f17e0001730 con 0x7f18085c85a0
2016-09-28 10:25:41.744541 7f17eae03700  1 -- 10.10.3.4:0/945549824 <== mon.2 10.10.3.3:6789/0 3 ==== auth_reply(proto 2 0 (0) Success) v1 ==== 206+0+0 (1705741500 0 0) 0x7f17d4001010 con 0x7f18085c85a0
2016-09-28 10:25:41.744765 7f17eae03700  1 -- 10.10.3.4:0/945549824 --> 10.10.3.3:6789/0 -- auth(proto 2 165 bytes epoch 0) v1 -- ?+0 0x7f17e0001bf0 con 0x7f18085c85a0
2016-09-28 10:25:41.745619 7f17eae03700  1 -- 10.10.3.4:0/945549824 <== mon.2 10.10.3.3:6789/0 4 ==== auth_reply(proto 2 0 (0) Success) v1 ==== 393+0+0 (482591267 0 0) 0x7f17d40008c0 con 0x7f18085c85a0
2016-09-28 10:25:41.745783 7f17eae03700  1 -- 10.10.3.4:0/945549824 --> 10.10.3.3:6789/0 -- mon_subscribe({monmap=0+}) v2 -- ?+0 0x7f18085cd560 con 0x7f18085c85a0
2016-09-28 10:25:41.745967 7f18072799c0  1 -- 10.10.3.4:0/945549824 --> 10.10.3.3:6789/0 -- mon_subscribe({osdmap=0}) v2 -- ?+0 0x7f18085c9850 con 0x7f18085c85a0
2016-09-28 10:25:41.746521 7f17eae03700  1 -- 10.10.3.4:0/945549824 <== mon.2 10.10.3.3:6789/0 5 ==== mon_map magic: 0 v1 ==== 495+0+0 (2693174994 0 0) 0x7f17d40012b0 con 0x7f18085c85a0
2016-09-28 10:25:41.746669 7f17daffd700  2 RGWDataChangesLog::ChangesRenewThread: start
2016-09-28 10:25:41.746882 7f18072799c0 20 get_system_obj_state: rctx=0x7ffe0afd57e0 obj=.rgw.root:default.realm state=0x7f18085cf4e8 s->prefetch_data=0
2016-09-28 10:25:41.746962 7f17eae03700  1 -- 10.10.3.4:0/945549824 <== mon.2 10.10.3.3:6789/0 6 ==== osd_map(5792..5792 src has 5225..5792) v3 ==== 13145+0+0 (1223904398 0 0) 0x7f17d40008c0 con 0x7f18085c85a0
2016-09-28 10:25:41.747661 7f18072799c0  1 -- 10.10.3.4:0/945549824 --> 10.10.3.12:6810/8166 -- osd_op(client.2974205.0:1 26.85fca992 default.realm [getxattrs,stat] snapc 0=[] ack+read+known_if_redirected e5792) v7 -- ?+0 0x7f18085d3410 con 0x7f18085d1eb0
2016-09-28 10:25:41.749762 7f17e83fc700  1 -- 10.10.3.4:0/945549824 <== osd.10 10.10.3.12:6810/8166 1 ==== osd_op_reply(1 default.realm [getxattrs,stat] v0'0 uv0 ack = -2 ((2) No such file or directory)) v7 ==== 175+0+0 (2651345336 0 0) 0x7f17cc000a50 con 0x7f18085d1eb0
unable to initialize zone: (2) No such file or directory
2016-09-28 10:25:41.749933 7f18072799c0 10 could not read realm id: (2) No such file or directory
2016-09-28 10:25:41.750290 7f18072799c0  1 -- 10.10.3.4:0/945549824 mark_down 0x7f18085d1eb0 -- 0x7f18085d0ba0
2016-09-28 10:25:41.750375 7f18072799c0  1 -- 10.10.3.4:0/945549824 mark_down 0x7f18085c85a0 -- 0x7f18085c72e0
2016-09-28 10:25:41.750575 7f18072799c0  1 -- 10.10.3.4:0/945549824 mark_down_all
2016-09-28 10:25:41.750888 7f18072799c0  1 -- 10.10.3.4:0/945549824 shutdown complete.



this is my conf:

[global]
fsid = 6f5a65a7-316c-4825-afcb-428608941dd1
mon_initial_members = cephadm, cephmon02, cephmon03
mon_host = 10.10.3.1,10.10.3.2,10.10.3.3
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
filestore_xattr_use_omap = true
osd_pool_default_size = 2
public_network = 10.10.0.0/16
cluster_network = 192.168.254.0/27
debug ms = 1
debug rgw = 20

[osd]
osd_journal_size = 20000

[client.cinder]
keyring = /etc/ceph/ceph.client.cinder.keyring

[client.rgw.cephrgw]
host = cephrwg
keyring = /etc/ceph/ceph.client.rgw.cephrgw.keyring
log_file = /var/log/ceph/client.rgw.cephrgw.log

Auth list:
client.bootstrap-rgw
key: AQALRwFWPBPDIxAABXXXXcwUsGPn0fMhhP19wg==
caps: [mon] allow profile bootstrap-rgw

client.rgw.cephrgw
key: AQAfsWNXUIv9MBAAXXXXzGk4strXE0UbWW4yzg==
caps: [mon] allow rw
caps: [osd] allow rwx

[root@cephrgw ~]# cat /etc/ceph/ceph.client.rgw.cephrgw.keyring
[client.rgw.cephrgw]
key = AQAfsWNXUIv9MBAAXXXXzGk4strXE0UbWW4yzg==

Any Idea?

regards, I


--
############################################################################
Iban Cabrillo Bartolome
Instituto de Fisica de Cantabria (IFCA)
Santander, Spain
Tel: +34942200969
PGP PUBLIC KEY: http://pgp.mit.edu/pks/lookup?op=get&search=0xD9DF0B3D6C8C08AC
############################################################################
Bertrand Russell:
"El problema con el mundo es que los estúpidos están seguros de todo y los inteligentes están llenos de dudas"
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux