Let me just clarify... the prepare process created all 10 partitions in sdg the thing is that only 2 (sdg1, sdg2) would be present in /dev. The partx bit is just a hack as I am not familiar with the entire sequence.
Initially I was deploying this test cluster in 5 nodes, each with 10 spinners, 1 OS spinner, 1 ssd for journal. *All* nodes would only bring up the first 2 osds.
From the start the partitions for journals are there:
~]# parted /dev/sdg
GNU Parted 2.1
Using /dev/sdg
Welcome to GNU Parted! Type 'help' to view a list of commands.
(parted) p
Model: ATA Samsung SSD 840 (scsi)
Disk /dev/sdg: 512GB
Sector size (logical/physical): 512B/512B
Partition Table: gpt
Number Start End Size File system Name Flags
1 1049kB 4295MB 4294MB ceph journal
2 4296MB 8590MB 4294MB ceph journal
3 8591MB 12.9GB 4294MB ceph journal
4 12.9GB 17.2GB 4294MB ceph journal
5 17.2GB 21.5GB 4294MB ceph journal
6 21.5GB 25.8GB 4294MB ceph journal
7 25.8GB 30.1GB 4294MB ceph journal
8 30.1GB 34.4GB 4294MB ceph journal
9 34.4GB 38.7GB 4294MB ceph journal
10 38.7GB 42.9GB 4294MB ceph journal
After partx all the entries show up under /dev and I have been able to install the cluster successfully.
The only weirdness happened with only one node. Not everything was entirely active+clean. That got resolved after I added the 2nd node.
At the moment with 3 nodes:
2013-08-08 17:38:38.328991 mon.0 [INF] pgmap v412: 192 pgs: 192 active+clean; 9518 bytes data, 1153 MB used, 83793 GB / 83794 GB avail
Thanks,
On Thu, Aug 8, 2013 at 8:17 AM, Sage Weil <sage@xxxxxxxxxxx> wrote:
Very strange!On Wed, 7 Aug 2013, Tren Blackburn wrote:
> On Tue, Aug 6, 2013 at 11:14 AM, Joao Pedras <jppedras@xxxxxxxxx> wrote:
> Greetings all.
> I am installing a test cluster using one ssd (/dev/sdg) to hold the
> journals. Ceph's version is 0.61.7 and I am using ceph-deploy obtained
> from ceph's git yesterday. This is on RHEL6.4, fresh install.
>
> When preparing the first 2 drives, sda and sdb, all goes well and the
> journals get created in sdg1 and sdg2:
>
> $> ceph-deploy osd prepare ceph00:sda:sdg ceph00:sdb:sdg
> [ceph_deploy.osd][DEBUG ] Preparing cluster ceph disks
> ceph00:/dev/sda:/dev/sdg ceph00:/dev/sdb:/dev/sdg
> [ceph_deploy.osd][DEBUG ] Deploying osd to ceph00
> [ceph_deploy.osd][DEBUG ] Host ceph00 is now ready for osd use.
> [ceph_deploy.osd][DEBUG ] Preparing host ceph00 disk /dev/sda journal
> /dev/sdg activate False
> [ceph_deploy.osd][DEBUG ] Preparing host ceph00 disk /dev/sdb journal
> /dev/sdg activate False
>
> When preparing sdc or any disk after the first 2 I get the following
> in that osd's log but no errors on ceph-deploy:
>
> # tail -f /var/log/ceph/ceph-osd.2.log
> 2013-08-06 10:51:36.655053 7f5ba701a780 0 ceph version 0.61.7
> (8f010aff684e820ecc837c25ac77c7a05d7191ff), process ceph-osd, pid
> 11596
> 2013-08-06 10:51:36.658671 7f5ba701a780 1
> filestore(/var/lib/ceph/tmp/mnt.i2NK47) mkfs in
> /var/lib/ceph/tmp/mnt.i2NK47
> 2013-08-06 10:51:36.658697 7f5ba701a780 1
> filestore(/var/lib/ceph/tmp/mnt.i2NK47) mkfs fsid is already set to
> 5d1beb09-1f80-421d-a88c-57789e2fc33e
> 2013-08-06 10:51:36.813783 7f5ba701a780 1
> filestore(/var/lib/ceph/tmp/mnt.i2NK47) leveldb db exists/created
> 2013-08-06 10:51:36.813964 7f5ba701a780 -1 journal FileJournal::_open:
> disabling aio for non-block journal. Use journal_force_aio to force
> use of aio anyway
> 2013-08-06 10:51:36.813999 7f5ba701a780 1 journal _open
> /var/lib/ceph/tmp/mnt.i2NK47/journal fd 10: 0 bytes, block size 4096
> bytes, directio = 1, aio = 0
> 2013-08-06 10:51:36.814035 7f5ba701a780 -1 journal check: ondisk fsid
> 00000000-0000-0000-0000-000000000000 doesn't match expected
> 5d1beb09-1f80-421d-a88c-57789e2fc33e, invalid (someone else's?)
> journal
> 2013-08-06 10:51:36.814093 7f5ba701a780 -1
> filestore(/var/lib/ceph/tmp/mnt.i2NK47) mkjournal error creating
> journal on /var/lib/ceph/tmp/mnt.i2NK47/journal: (22) Invalid argument
> 2013-08-06 10:51:36.814125 7f5ba701a780 -1 OSD::mkfs: FileStore::mkfs
> failed with error -22
> 2013-08-06 10:51:36.814185 7f5ba701a780 -1 ** ERROR: error creating
> empty object store in /var/lib/ceph/tmp/mnt.i2NK47: (22) Invalid
> argument
>
> I have cleaned the disks with dd, zapped them and so forth but this
> always occurs. If doing sdc/sdd first, for example, then sda or
> whatever follows fails with similar errors.
>
> Does anyone have any insight on this issue?
What does the partition table look like at this point? Does the joural
nsymlink in the osd data directory point to the right partition/device on
the failing osd?
sage
Joao Pedras
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com