On Fri, Aug 16, 2013 at 8:32 AM, Pavel Timoschenkov <Pavel@xxxxxxxxxxxxxxxxxxxxxxxxxxx> wrote: > <<<I suspect that there are left over partitions in /dev/sdaa that are > causing this to <<<fail, I *think* that we could pass the `-t` flag with the > filesystem and prevent this. > > Hi. Any changes ( > > Can you create a build that passes the -t flag with mount? > I tried going through these steps again and could not get any other ideas except to pass in that flag for mounting. Would you be willing to try a patch? (http://fpaste.org/33099/37691580/) You would need to apply it to the `ceph-disk` executable. > > > > > > > From: Pavel Timoschenkov > Sent: Thursday, August 15, 2013 3:43 PM > To: 'Alfredo Deza' > Cc: Samuel Just; ceph-users@xxxxxxxx > Subject: RE: ceph-deploy and journal on separate disk > > > > The separate commands (e.g. `ceph-disk -v prepare /dev/sda1`) works because > then the journal is on the same device as the OSD data, so the execution is > different to get them to a working state. > > I suspect that there are left over partitions in /dev/sdaa that are causing > this to fail, I *think* that we could pass the `-t` flag with the filesystem > and prevent this. > > Just to be sure, could you list all the partitions on /dev/sdaa (if > /dev/sdaa is the whole device)? > > Something like: > > sudo parted /dev/sdaa print > > Or if you prefer any other way that could tell use what are all the > partitions in that device. > > > > > > After > > ceph-deploy disk zap ceph001:sdaa ceph001:sda1 > > > > root@ceph001:~# parted /dev/sdaa print > > Model: ATA ST3000DM001-1CH1 (scsi) > > Disk /dev/sdaa: 3001GB > > Sector size (logical/physical): 512B/4096B > > Partition Table: gpt > > > > Number Start End Size File system Name Flags > > > > root@ceph001:~# parted /dev/sda1 print > > Model: Unknown (unknown) > > Disk /dev/sda1: 10.7GB > > Sector size (logical/physical): 512B/512B > > Partition Table: gpt > > So that is after running `disk zap`. What does it say after using > ceph-deploy and failing? > > > > Number Start End Size File system Name Flags > > > > After ceph-disk -v prepare /dev/sdaa /dev/sda1: > > > > root@ceph001:~# parted /dev/sdaa print > > Model: ATA ST3000DM001-1CH1 (scsi) > > Disk /dev/sdaa: 3001GB > > Sector size (logical/physical): 512B/4096B > > Partition Table: gpt > > > > Number Start End Size File system Name Flags > > 1 1049kB 3001GB 3001GB xfs ceph data > > > > And > > > > root@ceph001:~# parted /dev/sda1 print > > Model: Unknown (unknown) > > Disk /dev/sda1: 10.7GB > > Sector size (logical/physical): 512B/512B > > Partition Table: gpt > > > > Number Start End Size File system Name Flags > > > > With the same errors: > > > > root@ceph001:~# ceph-disk -v prepare /dev/sdaa /dev/sda1 > > DEBUG:ceph-disk:Journal /dev/sda1 is a partition > > WARNING:ceph-disk:OSD will not be hot-swappable if journal is not the same > device as the osd data > > DEBUG:ceph-disk:Creating osd partition on /dev/sdaa > > Information: Moved requested sector from 34 to 2048 in > > order to align on 2048-sector boundaries. > > The operation has completed successfully. > > DEBUG:ceph-disk:Creating xfs fs on /dev/sdaa1 > > meta-data=/dev/sdaa1 isize=2048 agcount=32, agsize=22892700 > blks > > = sectsz=512 attr=2, projid32bit=0 > > data = bsize=4096 blocks=732566385, imaxpct=5 > > = sunit=0 swidth=0 blks > > naming =version 2 bsize=4096 ascii-ci=0 > > log =internal log bsize=4096 blocks=357698, version=2 > > = sectsz=512 sunit=0 blks, lazy-count=1 > > realtime =none extsz=4096 blocks=0, rtextents=0 > > DEBUG:ceph-disk:Mounting /dev/sdaa1 on /var/lib/ceph/tmp/mnt.UkJbwx with > options noatime > > mount: /dev/sdaa1: more filesystems detected. This should not happen, > > use -t <type> to explicitly specify the filesystem type or > > use wipefs(8) to clean up the device. > > > > mount: you must specify the filesystem type > > ceph-disk: Mounting filesystem failed: Command '['mount', '-o', 'noatime', > '--', '/dev/sdaa1', '/var/lib/ceph/tmp/mnt.UkJbwx']' returned non-zero exit > status 32 > > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com