Re: ceph-deploy: osd creating hung with one ssd disk as shared journal

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I didn’t think you could specify the journal in this manner (just pointing multiple OSDs on the same host all to journal /dev/sda).  Don’t you either need to partition the SSD and point each SSD to a separate partition, or format and mount the SSD and each OSD will use a unique file on the mount?  I’ve always created a separate partition on the SSD for each journal.

 

Preparing cluster ceph disks ceph0:/dev/sdb:/dev/sda ceph0:/dev/sdd:/dev/sda ceph0:/dev/sde:/dev/sda ceph0:/dev/sdf:/dev/sda ceph0:/dev/sdg:/dev/sda ceph0:/dev/sdh:/dev/sda

 

From: ceph-users-bounces@xxxxxxxxxxxxxx [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of Tim Zhang
Sent: Tuesday, November 12, 2013 2:20 AM
To: ceph-users@xxxxxxxxxxxxxx
Subject: ceph-deploy: osd creating hung with one ssd disk as shared journal

 

Hi guys,

I use ceph-deploy to manage my cluster, but I get failed while creating the OSD, the process seems to hang up at creating first osd. By the way, SELinux is disabled, and my ceph-disk is patched according to the page:http://www.spinics.net/lists/ceph-users/msg03258.html

can you guys give me some advise?

(1) the output of ceph-deploy is:

Invoked (1.3.1): /usr/bin/ceph-deploy osd create ceph0:sdb:sda ceph0:sdd:sda ceph0:sde:sda ceph0:sdf:sda ceph0:sdg:sda ceph0:sdh:sda ceph1:sdb:sda ceph1:sdd:sda ceph1:sde:sda ceph1:sdf:sda ceph1:sdg:sda ceph1:sdh:sda ceph2:sdb:sda ceph2:sdd:sda ceph2:sde:sda ceph2:sdf:sda ceph2:sdg:sda ceph2:sdh:sda

[ceph_deploy.osd][DEBUG ] Preparing cluster ceph disks ceph0:/dev/sdb:/dev/sda ceph0:/dev/sdd:/dev/sda ceph0:/dev/sde:/dev/sda ceph0:/dev/sdf:/dev/sda ceph0:/dev/sdg:/dev/sda ceph0:/dev/sdh:/dev/sda ceph1:/dev/sdb:/dev/sda ceph1:/dev/sdd:/dev/sda ceph1:/dev/sde:/dev/sda ceph1:/dev/sdf:/dev/sda ceph1:/dev/sdg:/dev/sda ceph1:/dev/sdh:/dev/sda ceph2:/dev/sdb:/dev/sda ceph2:/dev/sdd:/dev/sda ceph2:/dev/sde:/dev/sda ceph2:/dev/sdf:/dev/sda ceph2:/dev/sdg:/dev/sda ceph2:/dev/sdh:/dev/sda

[ceph0][DEBUG ] connected to host: ceph0 

[ceph0][DEBUG ] detect platform information from remote host

[ceph0][DEBUG ] detect machine type

[ceph_deploy.osd][INFO  ] Distro info: CentOS 6.4 Final

[ceph_deploy.osd][DEBUG ] Deploying osd to ceph0

[ceph0][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf

[ceph0][INFO  ] Running command: udevadm trigger --subsystem-match=block --action="">

[ceph_deploy.osd][DEBUG ] Preparing host ceph0 disk /dev/sdb journal /dev/sda activate True

[ceph0][INFO  ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdb /dev/sda

[ceph0][ERROR ] WARNING:ceph-disk:OSD will not be hot-swappable if journal is not the same device as the osd data

[ceph0][ERROR ] Warning: WARNING: the kernel failed to re-read the partition table on /dev/sda (Device or resource busy).  As a result, it may not reflect all of your changes until after reboot.

[ceph0][ERROR ] BLKPG: Device or resource busy

[ceph0][ERROR ] error adding partition 1

[ceph0][DEBUG ] The operation has completed successfully.

[ceph0][DEBUG ] The operation has completed successfully.

[ceph0][DEBUG ] meta-data=""              isize=2048   agcount=4, agsize=61047597 blks

[ceph0][DEBUG ]          =                       sectsz=512   attr=2, projid32bit=0

[ceph0][DEBUG ] data     =                       bsize=4096   blocks=244190385, imaxpct=25

[ceph0][DEBUG ]          =                       sunit=0      swidth=0 blks

[ceph0][DEBUG ] naming   =version 2              bsize=4096   ascii-ci=0

[ceph0][DEBUG ] log      =internal log           bsize=4096   blocks=119233, version=2

[ceph0][DEBUG ]          =                       sectsz=512   sunit=0 blks, lazy-count=1

[ceph0][DEBUG ] realtime =none                   extsz=4096   blocks=0, rtextents=0

[ceph0][DEBUG ] The operation has completed successfully.

[ceph0][INFO  ] Running command: udevadm trigger --subsystem-match=block --action="">

[ceph_deploy.osd][DEBUG ] Host ceph0 is now ready for osd use.

[ceph0][DEBUG ] connected to host: ceph0 

[ceph0][DEBUG ] detect platform information from remote host

[ceph0][DEBUG ] detect machine type

[ceph_deploy.osd][INFO  ] Distro info: CentOS 6.4 Final

[ceph_deploy.osd][DEBUG ] Preparing host ceph0 disk /dev/sdd journal /dev/sda activate True

[ceph0][INFO  ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdd /dev/sda

[ceph0][ERROR ] WARNING:ceph-disk:OSD will not be hot-swappable if journal is not the same device as the osd data

 

2 the mount system for that osd shows:

[root@host ~]# mount -l

/dev/sdc1 on / type ext4 (rw)

proc on /proc type proc (rw)

sysfs on /sys type sysfs (rw)

devpts on /dev/pts type devpts (rw,gid=5,mode=620)

tmpfs on /dev/shm type tmpfs (rw)

none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)

/dev/sdb1 on /var/lib/ceph/tmp/mnt.6D02EM type xfs (rw,noatime)

 

3 my testbed information is:

os: centos 6.4 Final

ceph: dumpling 67.4

three hosts: ceph0 ceph1 ceph2

each host have 3 disk sharing one ssd disk as journal

 

4 my ceph config is as this:

osd journal size = 9500

;osd mkfs type = xfs

;auth supported = none

auth_cluster_required = none

auth_service_required = none

auth_client_required = none

public_network = 172.18.11.0/24

cluster_network = 10.10.11.0/24

osd pool default size = 3

ms nocrc = true

osd op threads = 4

filestore op threads = 0

mon sync fs threshold = 0

osd pool default pg num = 100

osd pool default pgp num = 100

 

5 the output of ceph0 running command: pe -ef|grep ceph

[root@ceph0 ~]# ps -ef|grep ceph

root     13922     1  0 05:59 ?        00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 1 sdb1 sdb

root     14059 13922  0 05:59 ?        00:00:00 python /usr/sbin/ceph-disk -v activate /dev/sdb1

root     14090     1  0 05:59 ?        00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 1 sda1 sda

root     14107 14090  0 05:59 ?        00:00:00 python /usr/sbin/ceph-disk -v activate-journal /dev/sda1

root     14220 14059  0 05:59 ?        00:00:00 /usr/bin/ceph-osd --cluster ceph --mkfs --mkkey -i 1 --monmap /var/lib/ceph/tmp/mnt.6D02EM/activate.monmap --osd-data /var/lib/ceph/tmp/mnt.6D02EM --osd-journal /var/lib/ceph/tmp/mnt.6D02EM/journal --osd-uuid 68381525-92ae-4e35-bf7f-aa34e31f2fae --keyring /var/lib/ceph/tmp/mnt.6D02EM/keyring

root     14280     1  0 06:02 ?        00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 1 sdb1 sdb

root     14305 14280  0 06:02 ?        00:00:00 python /usr/sbin/ceph-disk -v activate /dev/sdb1

root     14337     1  0 06:02 ?        00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 1 sda1 sda

root     14344     1  0 06:02 ?        00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 2 sda2 sda

root     14366 14337  0 06:02 ?        00:00:00 python /usr/sbin/ceph-disk -v activate-journal /dev/sda1

root     14371 14344  0 06:02 ?        00:00:00 python /usr/sbin/ceph-disk -v activate-journal /dev/sda2

root     14409     1  0 06:05 ?        00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 1 sdb1 sdb

root     14430 14409  0 06:05 ?        00:00:00 python /usr/sbin/ceph-disk -v activate /dev/sdb1

root     14444     1  0 06:05 ?        00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 1 sda1 sda

root     14457     1  0 06:05 ?        00:00:00 /bin/sh /usr/sbin/ceph-disk-udev 2 sda2 sda

root     14468 14444  0 06:05 ?        00:00:00 python /usr/sbin/ceph-disk -v activate-journal /dev/sda1

root     14483 14457  0 06:05 ?        00:00:00 python /usr/sbin/ceph-disk -v activate-journal /dev/sda2

root     14601 13254  0 06:18 pts/2    00:00:00 grep ceph

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux