Monitors are bound to a particular IP address. If you tell the client to connect to "52.28.87.xxx:6789" and the monitor responds saying "no, I'm 172.31.15.xxx:6789"...I don't think anything is expected to work. You'll need to pick one network or the other. -Greg On Fri, Jun 19, 2015 at 11:47 PM, Christoph Schäfer <schaefer@xxxxxxxxxxx> wrote: > Hi all, > > I currently have a test cluster running on aws. The vms are connected to each other using the internal ip addresses provided by amazon. > > I can mount a cephfs by using the internal ip address from the mon0 node with this command: > sudo mount -t ceph 172.31.15.xxx:6789:/ /client -o name=admin,secretfile=/root/secret > > However, if I try to mount on a client using the external ip of the mon0 node, the mount fails. Even if I try to mount on the mon0 node itself. > sudo mount -t ceph 52.28.87.xxx:6789:/ /client -o name=admin,secretfile=/root/secret > > The mds log seems to point to some sort of keyring/auth issue, which I can’t wrap my mind around. > ----- ceph-mds -i a -d ----- > starting mds.a at :/0 > 2015-06-19 22:42:38.265025 7fcb8c4b6800 0 ceph version 0.94.2 (5fb85614ca8f354284c713a2f9c610860720bbf3), process ceph-mds, pid 2518 > 2015-06-19 22:42:38.266442 7fcb8c4b6800 -1 asok(0x40ca000) AdminSocketConfigObs::init: failed: AdminSocket::bind_and_listen: failed to bind the UNIX domain socket to '/var/run/ceph/ceph-mds.a.asok': (13) Permission denied > 2015-06-19 22:42:38.266780 7fcb8c4b6800 -1 monclient(hunting): ERROR: missing keyring, cannot use cephx for authentication > 2015-06-19 22:42:38.266895 7fcb8c4b6800 -1 mds.-1.0 log_to_monitors {default=true} > 2015-06-19 22:42:38.267534 7fcb8c4b6800 -1 mds.-1.0 ERROR: failed to authenticate: (95) Operation not supported > 2015-06-19 22:42:38.267584 7fcb8c4b6800 1 mds.-1.0 suicide. wanted down:dne, now up:boot > --------------------------- > > I have a vague feeling it might be related to the „public network“ entry in the ceph.conf. But how do I have to change the config to make it work? > > I double checked firewall rules and ports, everything finde with the fw. > > I’m really puzzled here, any pointers are welcome… thanks in advance! > > Best > Chris > > ----- ceph.conf ----- > [global] > fsid = ef013617-4582-4ab1-bd65-8b7bca638e44 > mon_initial_members = mon0 > mon_host = 172.31.15.xxx > auth_cluster_required = cephx > auth_service_required = cephx > auth_client_required = cephx > > public network = 172.31.0.0/16 > mon_clock_drift_allowed = 1 > mon_pg_warn_max_per_osd = 0 > > [osd] > filestore xattr use omap = true > osd pool default size = 3 > osd pool default min size = 3 > osd crush chooseleaf type = 1 > ---------------------- > > ----- ceph -s ----- > ceph@mon0:~$ ceph -s > cluster ef013617-4582-4ab1-bd65-8b7bca638e44 > health HEALTH_OK > monmap e3: 3 mons at {mon0=172.31.15.xx:6789/0,mon1=172.31.4.xxx:6789/0,mon2=172.31.15.xxx:6789/0} > election epoch 36, quorum 0,1,2 mon1,mon0,mon2 > mdsmap e33: 1/1/1 up {0=osd2=up:active}, 2 up:standby > osdmap e75: 3 osds: 3 up, 3 in > pgmap v404: 384 pgs, 3 pools, 2000 MB data, 520 objects > 4122 MB used, 2079 GB / 2083 GB avail > 384 active+clean > ---------------------- > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com