Hi. With this patch - is all ok. Thanks for help! -----Original Message----- From: Alfredo Deza [mailto:alfredo.deza@xxxxxxxxxxx] Sent: Wednesday, August 21, 2013 7:16 PM To: Pavel Timoschenkov Cc: ceph-users@xxxxxxxx Subject: Re: ceph-deploy and journal on separate disk On Wed, Aug 21, 2013 at 9:33 AM, Pavel Timoschenkov <Pavel@xxxxxxxxxxxxxxxxxxxxxxxxxxx> wrote: > Hi. Thanks for patch. But after patched ceph src and install it, I have not ceph-disk or ceph-deploy command. > I did the following steps: > git clone --recursive https://github.com/ceph/ceph.git patch -p0 < > <patch name> ./autogen.sh ./configure make make install What am I > doing wrong? Oh I meant to patch it directly, there was no need to rebuild/make/install again because the file is a plain Python file (no compilation needed). Can you try that instead? > > -----Original Message----- > From: Alfredo Deza [mailto:alfredo.deza@xxxxxxxxxxx] > Sent: Monday, August 19, 2013 3:38 PM > To: Pavel Timoschenkov > Cc: ceph-users@xxxxxxxx > Subject: Re: ceph-deploy and journal on separate disk > > On Fri, Aug 16, 2013 at 8:32 AM, Pavel Timoschenkov <Pavel@xxxxxxxxxxxxxxxxxxxxxxxxxxx> wrote: >> <<<I suspect that there are left over partitions in /dev/sdaa that >> are causing this to <<<fail, I *think* that we could pass the `-t` >> flag with the filesystem and prevent this. >> >> Hi. Any changes ( >> >> Can you create a build that passes the -t flag with mount? >> > > I tried going through these steps again and could not get any other ideas except to pass in that flag for mounting. Would you be willing to try a patch? > (http://fpaste.org/33099/37691580/) > > You would need to apply it to the `ceph-disk` executable. > > >> >> >> >> >> >> >> From: Pavel Timoschenkov >> Sent: Thursday, August 15, 2013 3:43 PM >> To: 'Alfredo Deza' >> Cc: Samuel Just; ceph-users@xxxxxxxx >> Subject: RE: ceph-deploy and journal on separate disk >> >> >> >> The separate commands (e.g. `ceph-disk -v prepare /dev/sda1`) works >> because then the journal is on the same device as the OSD data, so >> the execution is different to get them to a working state. >> >> I suspect that there are left over partitions in /dev/sdaa that are >> causing this to fail, I *think* that we could pass the `-t` flag with >> the filesystem and prevent this. >> >> Just to be sure, could you list all the partitions on /dev/sdaa (if >> /dev/sdaa is the whole device)? >> >> Something like: >> >> sudo parted /dev/sdaa print >> >> Or if you prefer any other way that could tell use what are all the >> partitions in that device. >> >> >> >> >> >> After >> >> ceph-deploy disk zap ceph001:sdaa ceph001:sda1 >> >> >> >> root@ceph001:~# parted /dev/sdaa print >> >> Model: ATA ST3000DM001-1CH1 (scsi) >> >> Disk /dev/sdaa: 3001GB >> >> Sector size (logical/physical): 512B/4096B >> >> Partition Table: gpt >> >> >> >> Number Start End Size File system Name Flags >> >> >> >> root@ceph001:~# parted /dev/sda1 print >> >> Model: Unknown (unknown) >> >> Disk /dev/sda1: 10.7GB >> >> Sector size (logical/physical): 512B/512B >> >> Partition Table: gpt >> >> So that is after running `disk zap`. What does it say after using >> ceph-deploy and failing? >> >> >> >> Number Start End Size File system Name Flags >> >> >> >> After ceph-disk -v prepare /dev/sdaa /dev/sda1: >> >> >> >> root@ceph001:~# parted /dev/sdaa print >> >> Model: ATA ST3000DM001-1CH1 (scsi) >> >> Disk /dev/sdaa: 3001GB >> >> Sector size (logical/physical): 512B/4096B >> >> Partition Table: gpt >> >> >> >> Number Start End Size File system Name Flags >> >> 1 1049kB 3001GB 3001GB xfs ceph data >> >> >> >> And >> >> >> >> root@ceph001:~# parted /dev/sda1 print >> >> Model: Unknown (unknown) >> >> Disk /dev/sda1: 10.7GB >> >> Sector size (logical/physical): 512B/512B >> >> Partition Table: gpt >> >> >> >> Number Start End Size File system Name Flags >> >> >> >> With the same errors: >> >> >> >> root@ceph001:~# ceph-disk -v prepare /dev/sdaa /dev/sda1 >> >> DEBUG:ceph-disk:Journal /dev/sda1 is a partition >> >> WARNING:ceph-disk:OSD will not be hot-swappable if journal is not the >> same device as the osd data >> >> DEBUG:ceph-disk:Creating osd partition on /dev/sdaa >> >> Information: Moved requested sector from 34 to 2048 in >> >> order to align on 2048-sector boundaries. >> >> The operation has completed successfully. >> >> DEBUG:ceph-disk:Creating xfs fs on /dev/sdaa1 >> >> meta-data=/dev/sdaa1 isize=2048 agcount=32, agsize=22892700 >> blks >> >> = sectsz=512 attr=2, projid32bit=0 >> >> data = bsize=4096 blocks=732566385, imaxpct=5 >> >> = sunit=0 swidth=0 blks >> >> naming =version 2 bsize=4096 ascii-ci=0 >> >> log =internal log bsize=4096 blocks=357698, version=2 >> >> = sectsz=512 sunit=0 blks, lazy-count=1 >> >> realtime =none extsz=4096 blocks=0, rtextents=0 >> >> DEBUG:ceph-disk:Mounting /dev/sdaa1 on /var/lib/ceph/tmp/mnt.UkJbwx >> with options noatime >> >> mount: /dev/sdaa1: more filesystems detected. This should not happen, >> >> use -t <type> to explicitly specify the filesystem type or >> >> use wipefs(8) to clean up the device. >> >> >> >> mount: you must specify the filesystem type >> >> ceph-disk: Mounting filesystem failed: Command '['mount', '-o', >> 'noatime', '--', '/dev/sdaa1', '/var/lib/ceph/tmp/mnt.UkJbwx']' >> returned non-zero exit status 32 >> >> _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com