64 bit big endian > -----Original Message----- > From: Sage Weil [mailto:sage@xxxxxxxxxxxx] > Sent: Friday, October 24, 2014 5:47 PM > To: Prashanth Nednoor > Cc: ceph-devel@xxxxxxxxxxxxxxx; Philip Kufeldt > Subject: RE: Having issues trying to get the OSD up on a MIPS64!!! > > Hi Prashanth, > > On Fri, 24 Oct 2014, Prashanth Nednoor wrote: > > Hi Sage, > > > > Thank you for the prompt response. > > Is there anything in /dev/disk/by-partuuid/ or is it missing entirely? > > Nothing , it was Missing Entirely. > > GOOD NEWS: I worked around this issue, if I set my journal path in the > /etc/ceph.conf. > > > > My udev version is udevd --version 164 > > Hmm, that should be new enough, but it seems like it isn't setting up the > links. What distro is it? On most systems it's /lib/udev/rules.d/60-persistent- > storage.rules that does it. Maybe see if running partprobe /dev/sda or run > 'udevadm monitor' and do 'udevadm trigger /dev/sda' in another terminal to > see what happens. > > Or, work around it like you did. :) > > > I still see the segfaults, I have attached details. > > I put the osd debug logs(osd-output.txt) and the > leveldb_bt(leveldb_bt.txt). > > Looks like we have an issue in leveldb.... > > Yeah, that looks like a problem with leveldb. What distro is this? What > version leveldb? > > I don't actually know anything about MIPS.. what's teh wordsize and > endianess? > > sage > > > > > > HERE IS THE BACK TRACE: I have attached the gdb before running it. > > #0 0x77f68ee0 in leveldb::SkipList<char const*, > > leveldb::MemTable::KeyComparator>::FindGreaterOrEqual(char const* > > const&, leveldb::SkipList<char const*, > > leveldb::MemTable::KeyComparator>::Node**) const () from > > /usr/local/lib/libleveldb.so.1 > > #1 0x77f69054 in leveldb::SkipList<char const*, > > leveldb::MemTable::KeyComparator>::Insert(char const* const&) () from > > /usr/local/lib/libleveldb.so.1 > > #2 0x77f68618 in leveldb::MemTable::Add(unsigned long long, > leveldb::ValueType, leveldb::Slice const&, leveldb::Slice const&) () > > from /usr/local/lib/libleveldb.so.1 > > #3 0x77f7e434 in leveldb::(anonymous > namespace)::MemTableInserter::Put(leveldb::Slice const&, leveldb::Slice > const&) () > > from /usr/local/lib/libleveldb.so.1 > > #4 0x77f7e93c in > > leveldb::WriteBatch::Iterate(leveldb::WriteBatch::Handler*) const () > > from /usr/local/lib/libleveldb.so.1 > > #5 0x77f7eb8c in > > leveldb::WriteBatchInternal::InsertInto(leveldb::WriteBatch const*, > > leveldb::MemTable*) () from /usr/local/lib/libleveldb.so.1 > > #6 0x77f59360 in leveldb::DBImpl::Write(leveldb::WriteOptions const&, > > leveldb::WriteBatch*) () from /usr/local/lib/libleveldb.so.1 > > #7 0x00a5dda0 in LevelDBStore::submit_transaction_sync > > (this=0x1f77d10, t=<value optimized out>) at os/LevelDBStore.cc:146 > > #8 0x00b0d344 in DBObjectMap::sync (this=0x1f7af28, oid=0x0, > > spos=0x72cfe3b8) at os/DBObjectMap.cc:1126 > > #9 0x009b10b8 in FileStore::_set_replay_guard (this=0x1f72450, fd=17, > > spos=..., hoid=0x0, in_progress=false) at os/FileStore.cc:2070 > > #10 0x009b1c0c in FileStore::_set_replay_guard (this=0x1f72450, > cid=DWARF-2 expression error: DW_OP_reg operations must be used either > alone or in conjuction with DW_OP_piece. > > ) at os/FileStore.cc:2047 > > #11 0x009b2138 in FileStore::_create_collection (this=0x1f72450, c=DWARF- > 2 expression error: DW_OP_reg operations must be used either alone or in > conjuction with DW_OP_piece. > > ) at os/FileStore.cc:4753 > > #12 0x009e42a8 in FileStore::_do_transaction (this=0x1f72450, t=..., > > op_seq=<value optimized out>, trans_num=0, handle=0x72cfec3c) at > > os/FileStore.cc:2413 > > #13 0x009eb47c in FileStore::_do_transactions (this=0x1f72450, > > tls=..., op_seq=2, handle=0x72cfec3c) at os/FileStore.cc:1952 > > #14 0x009eb858 in FileStore::_do_op (this=0x1f72450, osr=0x1f801b8, > > handle=...) at os/FileStore.cc:1761 > > #15 0x00c8f0bc in ThreadPool::worker (this=0x1f72cf0, wt=0x1f7ea90) at > > common/WorkQueue.cc:128 > > #16 0x00c91b94 in ThreadPool::WorkThread::entry() () > > #17 0x77f1c0a8 in start_thread () from /lib/libpthread.so.0 > > #18 0x777c1738 in ?? () from /lib/libc.so.6 > > > > Do I need to set any variable to set the cache size etcetc in ceph.conf. > > I only have osd_leveldb_cache_size=5242880 for now. > > > > > > Thanks > > Prashanth > > > > > > > > > > > > > > > > -----Original Message----- > > From: Sage Weil [mailto:sage@xxxxxxxxxxxx] > > Sent: Thursday, October 23, 2014 5:54 PM > > To: Prashanth Nednoor > > Cc: ceph-devel@xxxxxxxxxxxxxxx > > Subject: Re: Having issues trying to get the OSD up on a MIPS64!!! > > > > Hi Prashanth, > > > > On Thu, 23 Oct 2014, Prashanth Nednoor wrote: > > > Hello Everyone, > > > > > > We are using ceph-0.86, good news is we were able to compile and > > > load all the libraries and binaries needed to configure a CEPH-OSD > > > on MIPS > > > 64 platform. The CEPH monitor is also able to detect the OSD, but > > > not up yet, as the osd activate failed. > > > Since we don?t have the required CEPH deploy utility for MIPS64, we > > > are following the manual procedure to create and activate an OSD. > > > We have disabled authentication between the clients and the OSD?s > > > for now. > > > > > > Has any body tried CEPH on a MIPS64? > > > /dev/sda is a 2TB local hard drive. > > > > > > This is how my partition looks after ceph-disk-prepare > > > /home/prashan/ceph-0.86/src# parted GNU Parted 2.3 Using /dev/sda > > > Welcome to GNU Parted! Type 'help' to view a list of commands. > > > (parted) p > > > Model: ATA TOSHIBA MQ01ABB2 (scsi) > > > Disk /dev/sda: 2000GB > > > Sector size (logical/physical): 512B/4096B Partition Table: gpt > > > > > > Number Start End Size File system Name Flags > > > 2 1049kB 5369MB 5368MB ceph journal > > > 1 5370MB 2000GB 1995GB xfs ceph data > > > > > > > > > > > > The following are the steps to create an OSD > > > 1) ceph-disk zap /dev/sda > > > 2) ceph-disk-prepare --cluster f615496c-b40a-4905-bbcd- > > > 2d3e181ff21a --fs-type xfs /dev/sda > > > 3) mount /dev/sda1 /var/lib/ceph/osd/ceph-0/ > > > 4) ceph-osd -i 0 ?mkfs is giving an error , > > > filestore(/var/lib/ceph/osd/ceph-0) could not find > > > 23c2fcde/osd_superblock/0//-1 in index: (2) No such file. > > > After this it segfaults. We have analyzed this further with the help > > > of strace and root caused this as objectmap file reading issue. > > > open("/var/lib/ceph/osd/ceph-0/current/omap/000005.log", O_RDONLY) > = > > > 11, the first time it reads 32k, the read succeeds with 63 bytes and > > > it tries to read again with 27k and the read returns 0 bytes and the > > > CEPH osd segfaults. > > > > Can you generate a full log with --debug-osd 20 --debug-filestore 20 -- > debug-jouranl 20 passed to ceph-osd --mkfs and post that somewhere? It > should tell us where things are going wrong. In particular, we want to see if > that file/object is being written properly. It will also have a backtrace > showing exactly where it crashed. > > > > > Please note that ceph-disk prepare creates a journal in a path which > > > is not valid(dev/disk/by-partuuid/cbd4a5d1-012f-4863-b492- > 080ad2a505cb). > > > So after step3 above I remove this journal below and manually create > > > a journal file before doing step4 above. > > > > > > > > > ls -l /var/lib/ceph/osd/ceph-0/ > > > total 16 > > > -rw-r--r-- 1 root root 37 Oct 22 21:40 ceph_fsid > > > -rw-r--r-- 1 root root 37 Oct 22 21:40 fsid lrwxrwxrwx 1 root root > > > 58 Oct 22 21:40 journal -> /dev/disk/by- > > > partuuid/cbd4a5d1-012f-4863-b492-080ad2a505cb > > > > Is there anything in /dev/disk/by-partuuid/ or is it missing entirely? > > Maybe you have an old udev. What distro is this? > > > > sage > > > > > -rw-r--r-- 1 root root 37 Oct 22 21:40 journal_uuid > > > -rw-r--r-- 1 root root 21 Oct 22 21:40 magic > > > > > > Any pointers to move ahead will be greatly appreciated?? > > > > > > thanks > > > Prashanth > > > > > > > > > > > > -- > > > To unsubscribe from this list: send the line "unsubscribe ceph-devel" > > > in the body of a message to majordomo@xxxxxxxxxxxxxxx More > majordomo > > > info at http://vger.kernel.org/majordomo-info.html > > > > > > > > -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html