Re: ceph monitors stuck in a loop after install with ceph-deploy

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, 24 Jul 2013, S?bastien RICCIO wrote:
> 
> Hi! While trying to install ceph using ceph-deploy the monitors nodes are
> stuck waiting on this process:
> /usr/bin/python /usr/sbin/ceph-create-keys -i a (or b or c)
> 
> I tried to run mannually the command and it loops on this:
> connect to /var/run/ceph/ceph-mon.a.asok failed with (2) No such file or
> directory
> INFO:ceph-create-keys:ceph-mon admin socket not ready yet.
> But the existing sock on the nodes are /var/run/ceph/ceph-mon.ceph01.asok
> 
> Is that a bug in ceph-deploy or maybe my config file is wrong ?

It's the config file.  You no longer need to (or should) enumerate the 
daemons in the config file; the sysvinit/upstart scripts find them in 
/var/lib/ceph/{osd,mon,mds}/*.  See below:

> Version:  ceph -v
> ceph version 0.61.6 (59ddece17e36fef69ecf40e239aeffad33c9db35)
> 
> Note that using "ceph" command line utility on the nodes is working. So it
> looks that it know the good paths...
> 
> Config file:
> 
> [global]
> fsid = a1394dff-94da-4ef4-a123-55d85e839ffb
> mon_initial_members = ceph01, ceph02, ceph03
> mon_host = 10.111.80.1,10.111.80.2,10.111.80.3
> auth_supported = cephx
> osd_journal_size = 10000
> filestore_xattr_use_omap = true
> auth_cluster_required = none
> auth_service_required = none
> auth_client_required = none
> 
> [client]
> rbd_cache = true
> rbd_cache_size = 536870912
> rbd_cache_max_dirty = 134217728
> rbd_cache_target_dirty = 33554432
> rbd_cache_max_dirty_age = 5
> 
> [osd]
> osd_data = /var/lib/ceph/osd/ceph-$id
> osd_journal = /var/lib/ceph/osd/ceph-$id/journal
> osd_journal_size = 10000
> osd_mkfs_type = xfs
> osd_mkfs_options_xfs = "-f -i size=2048"
> osd_mount_options_xfs = "rw,noatime,inode64,logbsize=256k"
> keyring = /var/lib/ceph/osd/ceph-$id/keyring.osd.$id
> osd_op_threads = 24
> osd_disk_threads = 24
> osd_recovery_max_active = 1
> journal_dio = true
> journal_aio = true
> filestore_max_sync_interval = 100
> filestore_min_sync_interval = 50
> filestore_queue_max_ops = 2000
> filestore_queue_max_bytes = 536870912
> filestore_queue_committing_max_ops = 2000
> filestore_queue_committing_max_bytes = 536870912
> osd_max_backfills = 1

Just drop everything from here...

> 
> [osd.0]
> host = ceph01
> 
> [osd.1]
> host = ceph01
> 
> [osd.2]
> host = ceph02
> 
> [osd.3]
> host = ceph02
> 
> [osd.4]
> host = ceph03
> 
> [osd.5]
> host = ceph03
> 
> [mon.a]
> host = ceph01
> 
> [mon.b]
> host = ceph02
> 
> [mon.c]
> host = ceph03

...to here!

sage



> 
> Cheers,
> S?bastien
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> 
> 
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux