Re: Monitor configuration issue

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 01/01/2014 09:29 PM, Matt Rabbitt wrote:
I only have four because I want to remove the original one I used to
create the cluster.  I tried what you suggested and rebooted all my
nodes but I'm still having the same problem.  I'm running Emperor on
Ubuntu 12.04 on all my nodes by the way.  Here is what I'm seeing as I
run ceph -w and reboot my original monitor.

      osdmap e124: 12 osds: 12 up, 12 in
       pgmap v26271: 528 pgs, 3 pools, 6979 MB data, 1883 objects
             20485 MB used, 44670 GB / 44690 GB avail
                  528 active+clean


2014-01-01 16:21:30.807305 mon.0 [INF] pgmap v26271: 528 pgs: 528
active+clean; 6979 MB data, 20485 MB us
                     ed, 44670 GB / 44690 GB avail
2014-01-01 16:22:06.098971 7f272d539700  0 monclient: hunting for new mon
2014-01-01 16:23:04.823206 7fe84c1bb700  0 -- :/1019476 >>
10.0.10.11:6789/0 <http://10.0.10.11:6789/0> pipe(0x7fe840009090 sd=3
                                                    :0 s=1 pgs=0 cs=0
l=1 c=0x7fe8400092f0).fault
2014-01-01 16:23:07.821642 7fe8443f9700  0 -- :/1019476 >>
10.0.10.11:6789/0 <http://10.0.10.11:6789/0> pipe(0x7fe840004140 sd=3
                                                    :0 s=1 pgs=0 cs=0
l=1 c=0x7fe8400043a0).fault

^this fault error continues until the monitor comes back online.

My best guess here is that your ceph.conf on the admin node (or wherever you're running that 'ceph -w' command) only contains mon.0's host/ip.

Those fault messages are shown whenever something (a client, a daemon) tries to connect to the monitor and fails. What *should* hapeen if you have the remaining monitors up and your ceph.conf well configured is that (in this case) the client will try the another monitor in the list. So you'd see that fault message when trying to connect to your down monitor, and it would eventually succeed as soon as it is able to connect to another monitor.

If you want to check if this is actually what is happening, try running the command with '--debug-monc 10'. It should tell you which monitor it is attempting to connect to. Also, check your client's ceph.conf.

Let us know how it goes.

  -Joao




On Wed, Jan 1, 2014 at 4:04 PM, Wolfgang Hennerbichler <wogri@xxxxxxxxx
<mailto:wogri@xxxxxxxxx>> wrote:

    Matt,

    first of all: four monitors is a bad idea. use an odd number for
    mons, e. g. three. your other problem is your configuration file.
    the mon_initial members and mon_host directives should include all
    monitor daemons. see my cluster:

    mon_initial_members = node01,node02,node03
    mon_host = 10.32.0.181,10.32.0.182,10.32.0.183

    hth
    wogri
    --
    http://www.wogri.at

    On 01 Jan 2014, at 21:55, Matt Rabbitt <mlrabbitt@xxxxxxxxx
    <mailto:mlrabbitt@xxxxxxxxx>> wrote:

     > I created a cluster, four monitors, and 12 OSDs using the
    ceph-deploy tool.  I initially created this cluster with one
    monitor, then added a "public network" statement in ceph.conf so
    that I could use ceph-deploy to add the other monitors.  When I run
    ceph -w now everything checks out and all monitors and OSDs show up
    and I can read and write data to my pool.  The problem is when I
    shut down the monitor that I initially used to configure the
    cluster, nothing works anymore.  If I run ceph -w all I get is fault
    errors about that first monitor being down, and I can't read or
    write data even though the other three monitors are still up.  What
    did I do wrong here?  I've been looking over the documentation and I
    see all kinds of info about having a mon addr attribute in my config
    or a public ip in the [mon] section but my config doesn't have
    anything like that in it.  Here is my complete config:
     >
     > [global]
     > fsid = a0ab5715-f9e6-4d71-8da6-0ad976ac350c
     > mon_initial_members = storage1
     > mon_host = 10.0.10.11
     > auth_supported = cephx
     > osd_journal_size = 6144
     > filestore_xattr_use_omap = true
     > public network = 10.0.10.0/24 <http://10.0.10.0/24>
     > _______________________________________________
     > ceph-users mailing list
     > ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
     > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



--
Joao Eduardo Luis
Software Engineer | http://inktank.com | http://ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux