Re: ceph mkfs failed

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

i think it's not related with my local FS. i build a ext4 on a ramdisk
and used it as OSD.
when i run the iozone or fio on the mounted client point, it  shows
the same info as before:

2013-02-08 11:45:06.803915 7f28ec7c4700  0 -- 165.91.215.237:6801/7101
>> 165.91.215.237:0/1990103183 pipe(0x2ded240 sd=803 :6801 pgs=0 cs=0
l=0).accept peer addr is really 165.91.215.237:0/1990103183 (socket is
165.91.215.237:60553/0)
2013-02-08 11:45:06.879009 7f28f7add700 -1 *** Caught signal
(Segmentation fault) **
 in thread 7f28f7add700

the ceph -s shows, also the same as using my own local FS:

  health HEALTH_WARN 384 pgs degraded; 384 pgs stuck unclean; recovery
21/42 degraded (50.000%)
   monmap e1: 1 mons at {0=165.91.215.237:6789/0}, election epoch 2, quorum 0 0
   osdmap e3: 1 osds: 1 up, 1 in
    pgmap v7: 384 pgs: 384 active+degraded; 21003 bytes data, 276 MB
used, 3484 MB / 3961 MB avail; 21/42 degraded (50.000%)
   mdsmap e4: 1/1/1 up {0=0=up:active}

dmesg shows:

[  656.799209] libceph: client4099 fsid da0fe76d-8506-4bf8-8b49-172fd8bc6d1f
[  656.800657] libceph: mon0 165.91.215.237:6789 session established
[  683.789954] libceph: osd0 165.91.215.237:6801 socket closed (con state OPEN)
[  683.790007] libceph: osd0 165.91.215.237:6801 socket error on write
[  684.909095] libceph: osd0 165.91.215.237:6801 socket error on write
[  685.903425] libceph: osd0 165.91.215.237:6801 socket error on write
[  687.903937] libceph: osd0 165.91.215.237:6801 socket error on write
[  691.897037] libceph: osd0 165.91.215.237:6801 socket error on write
[  699.899197] libceph: osd0 165.91.215.237:6801 socket error on write
[  715.903415] libceph: osd0 165.91.215.237:6801 socket error on write
[  747.912122] libceph: osd0 165.91.215.237:6801 socket error on write
[  811.929323] libceph: osd0 165.91.215.237:6801 socket error on write

Thanks,
Sheng


On Fri, Feb 8, 2013 at 11:07 AM, sheng qiu <herbert1984106@xxxxxxxxx> wrote:
> Hi Sage,
>
> it's a memory based fs similar to pramfs.
>
> Thanks,
> Sheng
>
> On Fri, Feb 8, 2013 at 11:02 AM, Sage Weil <sage@xxxxxxxxxxx> wrote:
>> Hi Sheng-
>>
>> On Fri, 8 Feb 2013, sheng qiu wrote:
>>> least pass through the init-ceph script). i made a minor change on
>>> ceph code, i changed the link_object() in LFNIndex.cc, basically i
>>> changed the hard link call ::link() to symlink(), as my local fs does
>>> not support hard link (the directory entry stores together with the
>>> related inodes).
>>
>> Unrelated question: which local fs are you using?
>>
>> sage
>
>
>
> --
> Sheng Qiu
> Texas A & M University
> Room 332B Wisenbaker
> email: herbert1984106@xxxxxxxxx
> College Station, TX 77843-3259



-- 
Sheng Qiu
Texas A & M University
Room 332B Wisenbaker
email: herbert1984106@xxxxxxxxx
College Station, TX 77843-3259
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux