Re: Troubles seting up radosgw

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Sep 28, 2016 at 10:32 AM, Iban Cabrillo <cabrillo@xxxxxxxxxxxxxx> wrote:
> Dear Admins,
>    During last day I have been trying to deploy a new radosgw, following
> jewel guide, ceph cluster is healthy (3 mon and 2 osd servers )
>    root@cephrgw ceph]# ceph -v
> ceph version 10.2.3 (ecc23778eb545d8dd55e2e4735b53cc93f92e65b)
>    [root@cephrgw ceph]# rpm -qa | grep ceph
> ceph-common-10.2.3-0.el7.x86_64
> libcephfs1-10.2.3-0.el7.x86_64
> ceph-deploy-1.5.36-0.noarch
> ceph-release-1-1.el7.noarch
> ceph-base-10.2.3-0.el7.x86_64
> ceph-radosgw-10.2.3-0.el7.x86_64
> python-cephfs-10.2.3-0.el7.x86_64
> ceph-selinux-10.2.3-0.el7.x86_64
>
> Cityweb is running on default port:
> [root@cephrgw ceph]# systemctl status ceph-radosgw@rgw.cephrgw.service
> ● ceph-radosgw@rgw.cephrgw.service - Ceph rados gateway
>    Loaded: loaded (/usr/lib/systemd/system/ceph-radosgw@.service; enabled;
> vendor preset: disabled)
>    Active: active (running) since mié 2016-09-28 10:20:34 CEST; 2s ago
>  Main PID: 29311 (radosgw)
>    CGroup:
> /system.slice/system-ceph\x2dradosgw.slice/ceph-radosgw@rgw.cephrgw.service
>            └─29311 /usr/bin/radosgw -f --cluster ceph --name
> client.rgw.cephrgw --setuser ceph --setgroup ceph
>
> sep 28 10:20:34 cephrgw.ifca.es systemd[1]: Started Ceph rados gateway.
> sep 28 10:20:34 cephrgw.ifca.es systemd[1]: Starting Ceph rados gateway...
>
> And this pools were created on ceph storage:
> .rgw.root                  2            4            0            0
> 0          252          195            4            5
> default.rgw.control            0            8            0            0
> 0            0            0            0            0
> default.rgw.data.root            0            0            0            0
> 0            0            0            0            0
> default.rgw.gc             0           32            0            0
> 0         2112         2080         1408            0
> default.rgw.log            0          127            0            0
> 0        47625        47498        31750            0
> default.rgw.users.uid            0            0            0            0
> 0            0            0            0            0
>
> But seems that zones are not well defined.
>
> radosgw-admin zone get --zone-id=default
> 2016-09-28 10:24:07.142478 7fd810b219c0  0 failed reading obj info from
> .rgw.root:zone_info.default: (2) No such file or directory
>
The zone id is not default this is the name try:
 radosgw-admin zone get --rgw-zone default

>
> [root@cephrgw ~]# radosgw-admin zone get
> 2016-09-28 10:25:41.740162 7f18072799c0  1 -- :/0 messenger.start
> 2016-09-28 10:25:41.741262 7f18072799c0  1 -- :/945549824 -->
> 10.10.3.3:6789/0 -- auth(proto 0 30 bytes epoch 0) v1 -- ?+0 0x7f18085c9850
> con 0x7f18085c85a0
> 2016-09-28 10:25:41.742048 7f180726f700  1 -- 10.10.3.4:0/945549824 learned
> my addr 10.10.3.4:0/945549824
> 2016-09-28 10:25:41.743168 7f17eae03700  1 -- 10.10.3.4:0/945549824 <==
> mon.2 10.10.3.3:6789/0 1 ==== mon_map magic: 0 v1 ==== 495+0+0 (2693174994 0
> 0) 0x7f17d4000b90 con 0x7f18085c85a0
> 2016-09-28 10:25:41.743380 7f17eae03700  1 -- 10.10.3.4:0/945549824 <==
> mon.2 10.10.3.3:6789/0 2 ==== auth_reply(proto 2 0 (0) Success) v1 ====
> 33+0+0 (3801669063 0 0) 0x7f17d4001010 con 0x7f18085c85a0
> 2016-09-28 10:25:41.743696 7f17eae03700  1 -- 10.10.3.4:0/945549824 -->
> 10.10.3.3:6789/0 -- auth(proto 2 32 bytes epoch 0) v1 -- ?+0 0x7f17e0001730
> con 0x7f18085c85a0
> 2016-09-28 10:25:41.744541 7f17eae03700  1 -- 10.10.3.4:0/945549824 <==
> mon.2 10.10.3.3:6789/0 3 ==== auth_reply(proto 2 0 (0) Success) v1 ====
> 206+0+0 (1705741500 0 0) 0x7f17d4001010 con 0x7f18085c85a0
> 2016-09-28 10:25:41.744765 7f17eae03700  1 -- 10.10.3.4:0/945549824 -->
> 10.10.3.3:6789/0 -- auth(proto 2 165 bytes epoch 0) v1 -- ?+0 0x7f17e0001bf0
> con 0x7f18085c85a0
> 2016-09-28 10:25:41.745619 7f17eae03700  1 -- 10.10.3.4:0/945549824 <==
> mon.2 10.10.3.3:6789/0 4 ==== auth_reply(proto 2 0 (0) Success) v1 ====
> 393+0+0 (482591267 0 0) 0x7f17d40008c0 con 0x7f18085c85a0
> 2016-09-28 10:25:41.745783 7f17eae03700  1 -- 10.10.3.4:0/945549824 -->
> 10.10.3.3:6789/0 -- mon_subscribe({monmap=0+}) v2 -- ?+0 0x7f18085cd560 con
> 0x7f18085c85a0
> 2016-09-28 10:25:41.745967 7f18072799c0  1 -- 10.10.3.4:0/945549824 -->
> 10.10.3.3:6789/0 -- mon_subscribe({osdmap=0}) v2 -- ?+0 0x7f18085c9850 con
> 0x7f18085c85a0
> 2016-09-28 10:25:41.746521 7f17eae03700  1 -- 10.10.3.4:0/945549824 <==
> mon.2 10.10.3.3:6789/0 5 ==== mon_map magic: 0 v1 ==== 495+0+0 (2693174994 0
> 0) 0x7f17d40012b0 con 0x7f18085c85a0
> 2016-09-28 10:25:41.746669 7f17daffd700  2
> RGWDataChangesLog::ChangesRenewThread: start
> 2016-09-28 10:25:41.746882 7f18072799c0 20 get_system_obj_state:
> rctx=0x7ffe0afd57e0 obj=.rgw.root:default.realm state=0x7f18085cf4e8
> s->prefetch_data=0
> 2016-09-28 10:25:41.746962 7f17eae03700  1 -- 10.10.3.4:0/945549824 <==
> mon.2 10.10.3.3:6789/0 6 ==== osd_map(5792..5792 src has 5225..5792) v3 ====
> 13145+0+0 (1223904398 0 0) 0x7f17d40008c0 con 0x7f18085c85a0
> 2016-09-28 10:25:41.747661 7f18072799c0  1 -- 10.10.3.4:0/945549824 -->
> 10.10.3.12:6810/8166 -- osd_op(client.2974205.0:1 26.85fca992 default.realm
> [getxattrs,stat] snapc 0=[] ack+read+known_if_redirected e5792) v7 -- ?+0
> 0x7f18085d3410 con 0x7f18085d1eb0
> 2016-09-28 10:25:41.749762 7f17e83fc700  1 -- 10.10.3.4:0/945549824 <==
> osd.10 10.10.3.12:6810/8166 1 ==== osd_op_reply(1 default.realm
> [getxattrs,stat] v0'0 uv0 ack = -2 ((2) No such file or directory)) v7 ====
> 175+0+0 (2651345336 0 0) 0x7f17cc000a50 con 0x7f18085d1eb0
> unable to initialize zone: (2) No such file or directory
> 2016-09-28 10:25:41.749933 7f18072799c0 10 could not read realm id: (2) No
> such file or directory
> 2016-09-28 10:25:41.750290 7f18072799c0  1 -- 10.10.3.4:0/945549824
> mark_down 0x7f18085d1eb0 -- 0x7f18085d0ba0
> 2016-09-28 10:25:41.750375 7f18072799c0  1 -- 10.10.3.4:0/945549824
> mark_down 0x7f18085c85a0 -- 0x7f18085c72e0
> 2016-09-28 10:25:41.750575 7f18072799c0  1 -- 10.10.3.4:0/945549824
> mark_down_all
> 2016-09-28 10:25:41.750888 7f18072799c0  1 -- 10.10.3.4:0/945549824 shutdown
> complete.
>
>
>
> this is my conf:
>
> [global]
> fsid = 6f5a65a7-316c-4825-afcb-428608941dd1
> mon_initial_members = cephadm, cephmon02, cephmon03
> mon_host = 10.10.3.1,10.10.3.2,10.10.3.3
> auth_cluster_required = cephx
> auth_service_required = cephx
> auth_client_required = cephx
> filestore_xattr_use_omap = true
> osd_pool_default_size = 2
> public_network = 10.10.0.0/16
> cluster_network = 192.168.254.0/27
> debug ms = 1
> debug rgw = 20
>
> [osd]
> osd_journal_size = 20000
>
> [client.cinder]
> keyring = /etc/ceph/ceph.client.cinder.keyring
>
> [client.rgw.cephrgw]
> host = cephrwg
> keyring = /etc/ceph/ceph.client.rgw.cephrgw.keyring
> log_file = /var/log/ceph/client.rgw.cephrgw.log
>
> Auth list:
> client.bootstrap-rgw
> key: AQALRwFWPBPDIxAABXXXXcwUsGPn0fMhhP19wg==
> caps: [mon] allow profile bootstrap-rgw
>
> client.rgw.cephrgw
> key: AQAfsWNXUIv9MBAAXXXXzGk4strXE0UbWW4yzg==
> caps: [mon] allow rw
> caps: [osd] allow rwx
>
> [root@cephrgw ~]# cat /etc/ceph/ceph.client.rgw.cephrgw.keyring
> [client.rgw.cephrgw]
> key = AQAfsWNXUIv9MBAAXXXXzGk4strXE0UbWW4yzg==
>
> Any Idea?
>
> regards, I
>
>
> --
> ############################################################################
> Iban Cabrillo Bartolome
> Instituto de Fisica de Cantabria (IFCA)
> Santander, Spain
> Tel: +34942200969
> PGP PUBLIC KEY:
> http://pgp.mit.edu/pks/lookup?op=get&search=0xD9DF0B3D6C8C08AC
> ############################################################################
> Bertrand Russell:
> "El problema con el mundo es que los estúpidos están seguros de todo y los
> inteligentes están llenos de dudas"
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux