ok, this is tested using ext3/ext4 on a normal SSD as OSD. ceph -s shows: health HEALTH_WARN 384 pgs degraded; 384 pgs stuck unclean; recovery 22/44 degraded (50.000%) monmap e1: 1 mons at {0=165.91.215.237:6789/0}, election epoch 2, quorum 0 0 osdmap e3: 1 osds: 1 up, 1 in pgmap v10: 384 pgs: 384 active+degraded; 26716 bytes data, 1184 MB used, 55857 MB / 60093 MB avail; 22/44 degraded (50.000%) mdsmap e4: 1/1/1 up {0=0=up:active} dmesg shows: [ 212.758376] libceph: client4106 fsid f60af615-67cb-4245-91cb-22752821f3e6 [ 212.759869] libceph: mon0 165.91.215.237:6789 session established [ 338.292461] libceph: osd0 165.91.215.237:6801 socket closed (con state OPEN) [ 338.292483] libceph: osd0 165.91.215.237:6801 socket error on write [ 339.161231] libceph: osd0 165.91.215.237:6801 socket error on write [ 340.159003] libceph: osd0 165.91.215.237:6801 socket error on write [ 342.158514] libceph: osd0 165.91.215.237:6801 socket error on write [ 346.149549] libceph: osd0 165.91.215.237:6801 socket error on write osd.0.log shows: 2013-02-08 14:52:51.649726 7f82780f6700 0 -- 165.91.215.237:6801/7135 >> 165.91.215.237:0/3238315774 pipe(0x2d61240 sd=803 :6801 pgs=0 cs=0 l=0).accept peer addr is really 165.91.215.237:0/3238315774 (socket is 165.91.215.237:57270/0) 2013-02-08 14:53:26.103770 7f8283c10700 -1 *** Caught signal (Segmentation fault) ** in thread 7f8283c10700 ceph version 0.56.1 (e4a541624df62ef353e754391cbbb707f54b16f7) 1: ./ceph-osd() [0x78648a] 2: (()+0x10060) [0x7f828cb0e060] 3: (fwrite()+0x34) [0x7f828aea3ec4] 4: (leveldb::log::Writer::EmitPhysicalRecord(leveldb::log::RecordType, char const*, unsigned long)+0x11f) [0x76d93f] 5: (leveldb::log::Writer::AddRecord(leveldb::Slice const&)+0x74) [0x76dae4] 6: (leveldb::DBImpl::Write(leveldb::WriteOptions const&, leveldb::WriteBatch*)+0x160) [0x763050] 7: (LevelDBStore::submit_transaction(std::tr1::shared_ptr<KeyValueDB::TransactionImpl>)+0x2a) [0x74ec1a] 8: (DBObjectMap::remove_xattrs(hobject_t const&, std::set<std::string, std::less<std::string>, std::allocator<std::string> > const&, SequencerPosition const*)+0x16a) [0x746fca] 9: (FileStore::_setattrs(coll_t, hobject_t const&, std::map<std::string, ceph::buffer::ptr, std::less<std::string>, std::allocator<std::pair<std::string const, ceph::buffer::ptr> > >&, SequencerPosition const&)+0xe7f) [0x719aff] 10: (FileStore::_do_transaction(ObjectStore::Transaction&, unsigned long, int)+0x3cba) [0x71e7da] 11: (FileStore::do_transactions(std::list<ObjectStore::Transaction*, std::allocator<ObjectStore::Transaction*> >&, unsigned long)+0x4c) [0x72152c] 12: (FileStore::_do_op(FileStore::OpSequencer*)+0x1b1) [0x6f1331] 13: (ThreadPool::worker(ThreadPool::WorkThread*)+0x4bc) [0x827dec] 14: (ThreadPool::WorkThread::entry()+0x10) [0x829cb0] 15: (()+0x7efc) [0x7f828cb05efc] 16: (clone()+0x6d) [0x7f828af1cf8d] NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed to interpret this. any suggestions? Thanks, Sheng On Fri, Feb 8, 2013 at 11:53 AM, sheng qiu <herbert1984106@xxxxxxxxx> wrote: > Hi, > > i think it's not related with my local FS. i build a ext4 on a ramdisk > and used it as OSD. > when i run the iozone or fio on the mounted client point, it shows > the same info as before: > > 2013-02-08 11:45:06.803915 7f28ec7c4700 0 -- 165.91.215.237:6801/7101 >>> 165.91.215.237:0/1990103183 pipe(0x2ded240 sd=803 :6801 pgs=0 cs=0 > l=0).accept peer addr is really 165.91.215.237:0/1990103183 (socket is > 165.91.215.237:60553/0) > 2013-02-08 11:45:06.879009 7f28f7add700 -1 *** Caught signal > (Segmentation fault) ** > in thread 7f28f7add700 > > the ceph -s shows, also the same as using my own local FS: > > health HEALTH_WARN 384 pgs degraded; 384 pgs stuck unclean; recovery > 21/42 degraded (50.000%) > monmap e1: 1 mons at {0=165.91.215.237:6789/0}, election epoch 2, quorum 0 0 > osdmap e3: 1 osds: 1 up, 1 in > pgmap v7: 384 pgs: 384 active+degraded; 21003 bytes data, 276 MB > used, 3484 MB / 3961 MB avail; 21/42 degraded (50.000%) > mdsmap e4: 1/1/1 up {0=0=up:active} > > dmesg shows: > > [ 656.799209] libceph: client4099 fsid da0fe76d-8506-4bf8-8b49-172fd8bc6d1f > [ 656.800657] libceph: mon0 165.91.215.237:6789 session established > [ 683.789954] libceph: osd0 165.91.215.237:6801 socket closed (con state OPEN) > [ 683.790007] libceph: osd0 165.91.215.237:6801 socket error on write > [ 684.909095] libceph: osd0 165.91.215.237:6801 socket error on write > [ 685.903425] libceph: osd0 165.91.215.237:6801 socket error on write > [ 687.903937] libceph: osd0 165.91.215.237:6801 socket error on write > [ 691.897037] libceph: osd0 165.91.215.237:6801 socket error on write > [ 699.899197] libceph: osd0 165.91.215.237:6801 socket error on write > [ 715.903415] libceph: osd0 165.91.215.237:6801 socket error on write > [ 747.912122] libceph: osd0 165.91.215.237:6801 socket error on write > [ 811.929323] libceph: osd0 165.91.215.237:6801 socket error on write > > Thanks, > Sheng > > > On Fri, Feb 8, 2013 at 11:07 AM, sheng qiu <herbert1984106@xxxxxxxxx> wrote: >> Hi Sage, >> >> it's a memory based fs similar to pramfs. >> >> Thanks, >> Sheng >> >> On Fri, Feb 8, 2013 at 11:02 AM, Sage Weil <sage@xxxxxxxxxxx> wrote: >>> Hi Sheng- >>> >>> On Fri, 8 Feb 2013, sheng qiu wrote: >>>> least pass through the init-ceph script). i made a minor change on >>>> ceph code, i changed the link_object() in LFNIndex.cc, basically i >>>> changed the hard link call ::link() to symlink(), as my local fs does >>>> not support hard link (the directory entry stores together with the >>>> related inodes). >>> >>> Unrelated question: which local fs are you using? >>> >>> sage >> >> >> >> -- >> Sheng Qiu >> Texas A & M University >> Room 332B Wisenbaker >> email: herbert1984106@xxxxxxxxx >> College Station, TX 77843-3259 > > > > -- > Sheng Qiu > Texas A & M University > Room 332B Wisenbaker > email: herbert1984106@xxxxxxxxx > College Station, TX 77843-3259 -- Sheng Qiu Texas A & M University Room 332B Wisenbaker email: herbert1984106@xxxxxxxxx College Station, TX 77843-3259 -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html