Hi guys,
I use ceph-deploy to manage my cluster, but I get failed while creating the OSD, the process seems to hang up at creating first osd. By the way, SELinux is disabled, and my ceph-disk is patched according to the page:http://www.spinics.net/lists/ceph-users/msg03258.html
can you guys give me some advise?
(1) the output of ceph-deploy is:
Invoked (1.3.1): /usr/bin/ceph-deploy osd create ceph0:sdb:sda ceph0:sdd:sda ceph0:sde:sda ceph0:sdf:sda ceph0:sdg:sda ceph0:sdh:sda ceph1:sdb:sda ceph1:sdd:sda ceph1:sde:sda ceph1:sdf:sda ceph1:sdg:sda ceph1:sdh:sda ceph2:sdb:sda ceph2:sdd:sda ceph2:sde:sda ceph2:sdf:sda ceph2:sdg:sda ceph2:sdh:sda
[ceph_deploy.osd][DEBUG ] Preparing cluster ceph disks ceph0:/dev/sdb:/dev/sda ceph0:/dev/sdd:/dev/sda ceph0:/dev/sde:/dev/sda ceph0:/dev/sdf:/dev/sda ceph0:/dev/sdg:/dev/sda ceph0:/dev/sdh:/dev/sda ceph1:/dev/sdb:/dev/sda ceph1:/dev/sdd:/dev/sda ceph1:/dev/sde:/dev/sda ceph1:/dev/sdf:/dev/sda ceph1:/dev/sdg:/dev/sda ceph1:/dev/sdh:/dev/sda ceph2:/dev/sdb:/dev/sda ceph2:/dev/sdd:/dev/sda ceph2:/dev/sde:/dev/sda ceph2:/dev/sdf:/dev/sda ceph2:/dev/sdg:/dev/sda ceph2:/dev/sdh:/dev/sda
[ceph0][DEBUG ] connected to host: ceph0
[ceph0][DEBUG ] detect platform information from remote host
[ceph0][DEBUG ] detect machine type
[ceph_deploy.osd][INFO ] Distro info: CentOS 6.4 Final
[ceph_deploy.osd][DEBUG ] Deploying osd to ceph0
[ceph0][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf
[ceph0][INFO ] Running command: udevadm trigger --subsystem-match=block --action="">
[ceph_deploy.osd][DEBUG ] Preparing host ceph0 disk /dev/sdb journal /dev/sda activate True
[ceph0][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdb /dev/sda
[ceph0][ERROR ] WARNING:ceph-disk:OSD will not be hot-swappable if journal is not the same device as the osd data
[ceph0][ERROR ] Warning: WARNING: the kernel failed to re-read the partition table on /dev/sda (Device or resource busy). As a result, it may not reflect all of your changes until after reboot.
[ceph0][ERROR ] BLKPG: Device or resource busy
[ceph0][ERROR ] error adding partition 1
[ceph0][DEBUG ] The operation has completed successfully.
[ceph0][DEBUG ] The operation has completed successfully.
[ceph0][DEBUG ] meta-data="" isize=2048 agcount=4, agsize=61047597 blks
[ceph0][DEBUG ] = sectsz=512 attr=2, projid32bit=0
[ceph0][DEBUG ] data = bsize=4096 blocks=244190385, imaxpct=25
[ceph0][DEBUG ] = sunit=0 swidth=0 blks
[ceph0][DEBUG ] naming =version 2 bsize=4096 ascii-ci=0
[ceph0][DEBUG ] log =internal log bsize=4096 blocks=119233, version=2
[ceph0][DEBUG ] = sectsz=512 sunit=0 blks, lazy-count=1
[ceph0][DEBUG ] realtime =none extsz=4096 blocks=0, rtextents=0
[ceph0][DEBUG ] The operation has completed successfully.
[ceph0][INFO ] Running command: udevadm trigger --subsystem-match=block --action="">
[ceph_deploy.osd][DEBUG ] Host ceph0 is now ready for osd use.
[ceph0][DEBUG ] connected to host: ceph0
[ceph0][DEBUG ] detect platform information from remote host
[ceph0][DEBUG ] detect machine type
[ceph_deploy.osd][INFO ] Distro info: CentOS 6.4 Final
[ceph_deploy.osd][DEBUG ] Preparing host ceph0 disk /dev/sdd journal /dev/sda activate True
[ceph0][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdd /dev/sda
[ceph0][ERROR ] WARNING:ceph-disk:OSD will not be hot-swappable if journal is not the same device as the osd data
2 the mount system for that osd shows:
[root@host ~]# mount -l
/dev/sdc1 on / type ext4 (rw)
proc on /proc type proc (rw)
sysfs on /sys type sysfs (rw)
devpts on /dev/pts type devpts (rw,gid=5,mode=620)
tmpfs on /dev/shm type tmpfs (rw)
none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
/dev/sdb1 on /var/lib/ceph/tmp/mnt.6D02EM type xfs (rw,noatime)
3 my testbed information is:
os: centos 6.4 Final
ceph: dumpling 67.4
three hosts: ceph0 ceph1 ceph2
each host have 3 disk sharing one ssd disk as journal
4 my ceph config is as this:
osd journal size = 9500
;osd mkfs type = xfs
;auth supported = none
auth_cluster_required = none
auth_service_required = none
auth_client_required = none
public_network = 172.18.11.0/24
cluster_network = 10.10.11.0/24
osd pool default size = 3
ms nocrc = true
osd op threads = 4
filestore op threads = 0
mon sync fs threshold = 0
osd pool default pg num = 100
osd pool default pgp num = 100
5 the output of ceph0 running command: pe -ef|grep ceph
[root@ceph0 ~]# ps -ef|grep ceph
root 13922 1 0 05:59 ? 00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 1 sdb1 sdb
root 14059 13922 0 05:59 ? 00:00:00 python /usr/sbin/ceph-disk -v activate /dev/sdb1
root 14090 1 0 05:59 ? 00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 1 sda1 sda
root 14107 14090 0 05:59 ? 00:00:00 python /usr/sbin/ceph-disk -v activate-journal /dev/sda1
root 14220 14059 0 05:59 ? 00:00:00 /usr/bin/ceph-osd --cluster ceph --mkfs --mkkey -i 1 --monmap /var/lib/ceph/tmp/mnt.6D02EM/activate.monmap --osd-data /var/lib/ceph/tmp/mnt.6D02EM --osd-journal /var/lib/ceph/tmp/mnt.6D02EM/journal --osd-uuid 68381525-92ae-4e35-bf7f-aa34e31f2fae --keyring /var/lib/ceph/tmp/mnt.6D02EM/keyring
root 14280 1 0 06:02 ? 00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 1 sdb1 sdb
root 14305 14280 0 06:02 ? 00:00:00 python /usr/sbin/ceph-disk -v activate /dev/sdb1
root 14337 1 0 06:02 ? 00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 1 sda1 sda
root 14344 1 0 06:02 ? 00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 2 sda2 sda
root 14366 14337 0 06:02 ? 00:00:00 python /usr/sbin/ceph-disk -v activate-journal /dev/sda1
root 14371 14344 0 06:02 ? 00:00:00 python /usr/sbin/ceph-disk -v activate-journal /dev/sda2
root 14409 1 0 06:05 ? 00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 1 sdb1 sdb
root 14430 14409 0 06:05 ? 00:00:00 python /usr/sbin/ceph-disk -v activate /dev/sdb1
root 14444 1 0 06:05 ? 00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 1 sda1 sda
root 14457 1 0 06:05 ? 00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 2 sda2 sda
root 14468 14444 0 06:05 ? 00:00:00 python /usr/sbin/ceph-disk -v activate-journal /dev/sda1
root 14483 14457 0 06:05 ? 00:00:00 python /usr/sbin/ceph-disk -v activate-journal /dev/sda2
root 14601 13254 0 06:18 pts/2 00:00:00 grep ceph
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
- Follow-Ups:
- Re: ceph-deploy: osd creating hung with one ssd disk as shared journal
- From: Gruher, Joseph R
- Re: ceph-deploy: osd creating hung with one ssd disk as shared journal
- Prev by Date: how to use rados_exec
- Next by Date: HDD bad sector, pg inconsistent, no object remapping
- Previous by thread: how to use rados_exec
- Next by thread: Re: ceph-deploy: osd creating hung with one ssd disk as shared journal
- Index(es):