Re: ceph mkfs failed

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



ok, this is tested using ext3/ext4 on a normal SSD as OSD.

ceph -s shows:
health HEALTH_WARN 384 pgs degraded; 384 pgs stuck unclean; recovery
22/44 degraded (50.000%)
   monmap e1: 1 mons at {0=165.91.215.237:6789/0}, election epoch 2, quorum 0 0
   osdmap e3: 1 osds: 1 up, 1 in
    pgmap v10: 384 pgs: 384 active+degraded; 26716 bytes data, 1184 MB
used, 55857 MB / 60093 MB avail; 22/44 degraded (50.000%)
   mdsmap e4: 1/1/1 up {0=0=up:active}

dmesg shows:
[  212.758376] libceph: client4106 fsid f60af615-67cb-4245-91cb-22752821f3e6
[  212.759869] libceph: mon0 165.91.215.237:6789 session established
[  338.292461] libceph: osd0 165.91.215.237:6801 socket closed (con state OPEN)
[  338.292483] libceph: osd0 165.91.215.237:6801 socket error on write
[  339.161231] libceph: osd0 165.91.215.237:6801 socket error on write
[  340.159003] libceph: osd0 165.91.215.237:6801 socket error on write
[  342.158514] libceph: osd0 165.91.215.237:6801 socket error on write
[  346.149549] libceph: osd0 165.91.215.237:6801 socket error on write

osd.0.log shows:
2013-02-08 14:52:51.649726 7f82780f6700  0 -- 165.91.215.237:6801/7135
>> 165.91.215.237:0/3238315774 pipe(0x2d61240 sd=803 :6801 pgs=0 cs=0
l=0).accept peer addr is really 165.91.215.237:0/3238315774 (socket is
165.91.215.237:57270/0)
2013-02-08 14:53:26.103770 7f8283c10700 -1 *** Caught signal
(Segmentation fault) **
 in thread 7f8283c10700

 ceph version 0.56.1 (e4a541624df62ef353e754391cbbb707f54b16f7)
 1: ./ceph-osd() [0x78648a]
 2: (()+0x10060) [0x7f828cb0e060]
 3: (fwrite()+0x34) [0x7f828aea3ec4]
 4: (leveldb::log::Writer::EmitPhysicalRecord(leveldb::log::RecordType,
char const*, unsigned long)+0x11f) [0x76d93f]
 5: (leveldb::log::Writer::AddRecord(leveldb::Slice const&)+0x74) [0x76dae4]
 6: (leveldb::DBImpl::Write(leveldb::WriteOptions const&,
leveldb::WriteBatch*)+0x160) [0x763050]
 7: (LevelDBStore::submit_transaction(std::tr1::shared_ptr<KeyValueDB::TransactionImpl>)+0x2a)
[0x74ec1a]
 8: (DBObjectMap::remove_xattrs(hobject_t const&,
std::set<std::string, std::less<std::string>,
std::allocator<std::string> > const&, SequencerPosition const*)+0x16a)
[0x746fca]
 9: (FileStore::_setattrs(coll_t, hobject_t const&,
std::map<std::string, ceph::buffer::ptr, std::less<std::string>,
std::allocator<std::pair<std::string const, ceph::buffer::ptr> > >&,
SequencerPosition const&)+0xe7f) [0x719aff]
 10: (FileStore::_do_transaction(ObjectStore::Transaction&, unsigned
long, int)+0x3cba) [0x71e7da]
 11: (FileStore::do_transactions(std::list<ObjectStore::Transaction*,
std::allocator<ObjectStore::Transaction*> >&, unsigned long)+0x4c)
[0x72152c]
 12: (FileStore::_do_op(FileStore::OpSequencer*)+0x1b1) [0x6f1331]
 13: (ThreadPool::worker(ThreadPool::WorkThread*)+0x4bc) [0x827dec]
 14: (ThreadPool::WorkThread::entry()+0x10) [0x829cb0]
 15: (()+0x7efc) [0x7f828cb05efc]
 16: (clone()+0x6d) [0x7f828af1cf8d]
 NOTE: a copy of the executable, or `objdump -rdS <executable>` is
needed to interpret this.


any suggestions?

Thanks,
Sheng

On Fri, Feb 8, 2013 at 11:53 AM, sheng qiu <herbert1984106@xxxxxxxxx> wrote:
> Hi,
>
> i think it's not related with my local FS. i build a ext4 on a ramdisk
> and used it as OSD.
> when i run the iozone or fio on the mounted client point, it  shows
> the same info as before:
>
> 2013-02-08 11:45:06.803915 7f28ec7c4700  0 -- 165.91.215.237:6801/7101
>>> 165.91.215.237:0/1990103183 pipe(0x2ded240 sd=803 :6801 pgs=0 cs=0
> l=0).accept peer addr is really 165.91.215.237:0/1990103183 (socket is
> 165.91.215.237:60553/0)
> 2013-02-08 11:45:06.879009 7f28f7add700 -1 *** Caught signal
> (Segmentation fault) **
>  in thread 7f28f7add700
>
> the ceph -s shows, also the same as using my own local FS:
>
>   health HEALTH_WARN 384 pgs degraded; 384 pgs stuck unclean; recovery
> 21/42 degraded (50.000%)
>    monmap e1: 1 mons at {0=165.91.215.237:6789/0}, election epoch 2, quorum 0 0
>    osdmap e3: 1 osds: 1 up, 1 in
>     pgmap v7: 384 pgs: 384 active+degraded; 21003 bytes data, 276 MB
> used, 3484 MB / 3961 MB avail; 21/42 degraded (50.000%)
>    mdsmap e4: 1/1/1 up {0=0=up:active}
>
> dmesg shows:
>
> [  656.799209] libceph: client4099 fsid da0fe76d-8506-4bf8-8b49-172fd8bc6d1f
> [  656.800657] libceph: mon0 165.91.215.237:6789 session established
> [  683.789954] libceph: osd0 165.91.215.237:6801 socket closed (con state OPEN)
> [  683.790007] libceph: osd0 165.91.215.237:6801 socket error on write
> [  684.909095] libceph: osd0 165.91.215.237:6801 socket error on write
> [  685.903425] libceph: osd0 165.91.215.237:6801 socket error on write
> [  687.903937] libceph: osd0 165.91.215.237:6801 socket error on write
> [  691.897037] libceph: osd0 165.91.215.237:6801 socket error on write
> [  699.899197] libceph: osd0 165.91.215.237:6801 socket error on write
> [  715.903415] libceph: osd0 165.91.215.237:6801 socket error on write
> [  747.912122] libceph: osd0 165.91.215.237:6801 socket error on write
> [  811.929323] libceph: osd0 165.91.215.237:6801 socket error on write
>
> Thanks,
> Sheng
>
>
> On Fri, Feb 8, 2013 at 11:07 AM, sheng qiu <herbert1984106@xxxxxxxxx> wrote:
>> Hi Sage,
>>
>> it's a memory based fs similar to pramfs.
>>
>> Thanks,
>> Sheng
>>
>> On Fri, Feb 8, 2013 at 11:02 AM, Sage Weil <sage@xxxxxxxxxxx> wrote:
>>> Hi Sheng-
>>>
>>> On Fri, 8 Feb 2013, sheng qiu wrote:
>>>> least pass through the init-ceph script). i made a minor change on
>>>> ceph code, i changed the link_object() in LFNIndex.cc, basically i
>>>> changed the hard link call ::link() to symlink(), as my local fs does
>>>> not support hard link (the directory entry stores together with the
>>>> related inodes).
>>>
>>> Unrelated question: which local fs are you using?
>>>
>>> sage
>>
>>
>>
>> --
>> Sheng Qiu
>> Texas A & M University
>> Room 332B Wisenbaker
>> email: herbert1984106@xxxxxxxxx
>> College Station, TX 77843-3259
>
>
>
> --
> Sheng Qiu
> Texas A & M University
> Room 332B Wisenbaker
> email: herbert1984106@xxxxxxxxx
> College Station, TX 77843-3259



-- 
Sheng Qiu
Texas A & M University
Room 332B Wisenbaker
email: herbert1984106@xxxxxxxxx
College Station, TX 77843-3259
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux