Re: CephFS and the next jewel release v10.2.3

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



You've got two jobs never getting locks, 2 dead on Java linker issues
(which probably never worked; they aren't any kind of release
blocker), and one OSD crash. The crash is

2016-08-25 09:21:45.616945 7fa22623a700 -1 os/filestore/FileStore.cc:
In function 'void
FileStore::_do_transaction(ObjectStore::Transaction&, uint64_t, int,
ThreadPool::TPHandle*)' thread 7fa22623a700 time 2016-08-25
09:21:45.612229
os/filestore/FileStore.cc: 2912: FAILED assert(0 == "unexpected error")

That most often is an fs/disk issue of some kind, but the kernel logs
are sadly empty so I can't guarantee it.


On Thu, Aug 25, 2016 at 2:13 PM, Loic Dachary <loic@xxxxxxxxxxx> wrote:
> Hi,
>
> The run completed at http://pulpito.ceph.com/loic-2016-08-25_06:39:08-fs-jewel-backports-distro-basic-smithi/ with no valgrind error but a few other failures. Note that it is based on the jewel-backports branch which contains a few more backports in addition to https://github.com/ceph/ceph/pull/10847 and I'm not sure what to blame yet. I'll look into it tomorrow unless you beat me to it :-)
>
> Cheers
>
> On 24/08/2016 23:03, Gregory Farnum wrote:
>> On Wed, Aug 24, 2016 at 2:02 PM, Loic Dachary <loic@xxxxxxxxxxx> wrote:
>>>
>>>
>>> On 24/08/2016 22:46, Gregory Farnum wrote:
>>>> On Wed, Aug 24, 2016 at 11:11 AM, Gregory Farnum <gfarnum@xxxxxxxxxx> wrote:
>>>>> On Tue, Aug 23, 2016 at 1:59 AM, Loic Dachary <loic@xxxxxxxxxxx> wrote:
>>>>>> Hi John,
>>>>>>
>>>>>> The next jewel release as found at https://github.com/ceph/ceph/tree/jewel passed the fs suite (http://tracker.ceph.com/issues/16344#note-30 except for three valgrind failures). Do you think the jewel branch is ready for QE to start their own round of testing ?
>>>>>>
>>>>>> Cheers
>>>>>>
>>>>>> P.S. CC'ing Greg because John is on vacation, in case a decision is to be made before he returns.
>>>>>
>>>>> Ugh, I went to check these valgrind failures and they're a failed
>>>>> mutex assert on shutdown. Let me dig into what's happening. :(
>>>>> -Greg
>>>>
>>>> Hmm, we seem to have some badly-colliding backports. If you look at
>>>> commit ac449472df2978123192d5f22247d8b8bc5efb28 and MDSRank.cc line
>>>> 242, you'll see it does mds_lock.Unlock() twice in quick succession.
>>>> :(
>>>>
>>>> One of them came in from 7c2eab19, and the other from 1d3a8168. I
>>>> think maybe the patches got reordered in their backport order or
>>>> something.
>>>>
>>>> Patrick, can you give https://github.com/ceph/ceph/pull/10847 a quick
>>>> review? Loic, can you then run the failed tests against a backport
>>>> branch which includes that patch?
>>>
>>> https://github.com/ceph/ceph/pull/10847 has been pushed as part of the jewel-backports integration branch right now and I'll schedule a fs run as soon as it finishes building. Thanks for the quick fix !
>>
>> Sure. Assuming that works out everything else looked good to me. :)
>> --
>> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>
> --
> Loïc Dachary, Artisan Logiciel Libre
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux