Re: Ceph-osd Daemon Receives Segmentation Fault on Trusty After Upgrading to 0.94.10 Release

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> i read from an OpenStack/Ceph tuning document and when you set this parameter to true you could get better performance for block level storage

Only if you copy sparse rbd images a lot, however in practice sparse
rbd images are rare.
On the other hand fiemap is somewhat counter-intuitive and tricky to
use correctly,
hence quite a number of fiemap related bugs have lurked into the code
(not only in ceph).
That's why ceph no longer uses fiemap by default.

> Is it completely safe to directly change it from true to false and restart Ceph deamons in order.

Disabling fiemap and restarting ceph-osd's should be pretty safe
(usual precautions still apply:
it's wise to set noout flag before restarting OSDs, and restart them
one by one giving each OSD
enough time to perform peering and recovery).

Best regards,
      Alexey


On Tue, Mar 21, 2017 at 5:00 PM, Özhan Rüzgar Karaman
<oruzgarkaraman@xxxxxxxxx> wrote:
> Hi Alexey;
> 1 year ago, i read from an OpenStack/Ceph tuning document and when you set
> this parameter to true you could get better performance for block level
> storage.
>
> Is it completely safe to directly change it from true to false and restart
> Ceph deamons in order.
>
> Thanks for all your support.
>
> Özhan
>
>
> On Tue, Mar 21, 2017 at 3:27 PM, Alexey Sheplyakov
> <asheplyakov@xxxxxxxxxxxx> wrote:
>>
>> Hi,
>>
>> This looks like a bug [1]. You can work it around by disabling the
>> fiemap feature, like this:
>>
>> [osd]
>> filestore fiemap = false
>>
>> Fiemap should have been disabled by default, perhaps you've explicitly
>> enabled it?
>>
>> [1] http://tracker.ceph.com/issues/19323
>>
>> Best regards,
>>       Alexey
>>
>> On Tue, Mar 21, 2017 at 12:21 PM, Özhan Rüzgar Karaman
>> <oruzgarkaraman@xxxxxxxxx> wrote:
>> > Hi Wido;
>> > After 30 minutes osd id 3 crashed also with segmentation fault, i
>> > uploaded
>> > logs again to the same location as ceph.log.wido.20170321-3.tgz. So now
>> > all
>> > OSD deamons on that server is crashed.
>> >
>> > Thanks
>> > Özhan
>> >
>> > On Tue, Mar 21, 2017 at 10:57 AM, Özhan Rüzgar Karaman
>> > <oruzgarkaraman@xxxxxxxxx> wrote:
>> >>
>> >> Hi Wido;
>> >> At weekend i roll back all servers to 0.94.9-1 version and all worked
>> >> fine
>> >> with old release.
>> >>
>> >> Today i upgraded all monitor servers and 1 osd server to 0.94.10-1
>> >> version. All OSD servers has 2 osds. I update the ceph.conf on the osd
>> >> server removed debug lines and restart osd daemons.
>> >>
>> >> This time osd id 3 started and operated successfully but osd id 2
>> >> failed
>> >> again with same segmentation fault.
>> >>
>> >> I have uploaded new logs as to the same destination as
>> >> ceph.log.wido.20170321-2.tgz and its link is below again.
>> >>
>> >>
>> >>
>> >> https://drive.google.com/drive/folders/0B_hD9LJqrkd7NmtJOW5YUnh6UE0?usp=sharing
>> >>
>> >> Thanks for all your help.
>> >>
>> >> Özhan
>> >>
>> >>
>> >> On Sun, Mar 19, 2017 at 8:47 PM, Wido den Hollander <wido@xxxxxxxx>
>> >> wrote:
>> >>>
>> >>>
>> >>> > Op 17 maart 2017 om 8:39 schreef Özhan Rüzgar Karaman
>> >>> > <oruzgarkaraman@xxxxxxxxx>:
>> >>> >
>> >>> >
>> >>> > Hi;
>> >>> > Yesterday i started to upgrade my Ceph environment from 0.94.9 to
>> >>> > 0.94.10.
>> >>> > All monitor servers upgraded successfully but i experience problems
>> >>> > on
>> >>> > starting upgraded OSD daemons.
>> >>> >
>> >>> > When i try to start an Ceph OSD Daemon(/usr/bin/ceph-osd) receives
>> >>> > Segmentation Fault and it kills after 2-3 minutes. To clarify the
>> >>> > issue
>> >>> > i
>> >>> > have role backed Ceph packages on that OSD Server  back to 0.94.9
>> >>> > and
>> >>> > problematic servers could rejoin to the 0.94.10 cluster.
>> >>> >
>> >>> > My environment is standard 14.04.5 Ubuntu Trusty server with 4.4.x
>> >>> > kernel
>> >>> > and i am using standard packages from
>> >>> > http://eu.ceph.com/debian-hammer
>> >>> > nothing special on my environment.
>> >>> >
>> >>> > I have uploaded the Ceph OSD Logs to the link below.
>> >>> >
>> >>> >
>> >>> >
>> >>> > https://drive.google.com/drive/folders/0B_hD9LJqrkd7NmtJOW5YUnh6UE0?usp=sharing
>> >>> >
>> >>> > And my ceph.conf is below
>> >>> >
>> >>> > [global]
>> >>> > fsid = a3742d34-9b51-4a36-bf56-4defb62b2b8e
>> >>> > mon_initial_members = mont1, mont2, mont3
>> >>> > mon_host = 172.16.51.101,172.16.51.102,172.16.51.103
>> >>> > auth_cluster_required = cephx
>> >>> > auth_service_required = cephx
>> >>> > auth_client_required = cephx
>> >>> > filestore_xattr_use_omap = true
>> >>> > public_network = 172.16.51.0/24
>> >>> > cluster_network = 172.16.51.0/24
>> >>> > debug_ms = 0/0
>> >>> > debug_auth = 0/0
>> >>> >
>> >>> > [mon]
>> >>> > mon_allow_pool_delete = false
>> >>> > mon_osd_down_out_interval = 300
>> >>> > osd_pool_default_flag_nodelete = true
>> >>> >
>> >>> > [osd]
>> >>> > filestore_max_sync_interval = 15
>> >>> > filestore_fiemap = true
>> >>> > osd_max_backfills = 1
>> >>> > osd_backfill_scan_min = 16
>> >>> > osd_backfill_scan_max = 128
>> >>> > osd_max_scrubs = 1
>> >>> > osd_scrub_sleep = 1
>> >>> > osd_scrub_chunk_min = 2
>> >>> > osd_scrub_chunk_max = 16
>> >>> > debug_osd = 0/0
>> >>> > debug_filestore = 0/0
>> >>> > debug_rbd = 0/0
>> >>> > debug_rados = 0/0
>> >>> > debug_journal = 0/0
>> >>> > debug_journaler = 0/0
>> >>>
>> >>> Can you try without all the debug_* lines and see what the log then
>> >>> yields?
>> >>>
>> >>> It's crashing on something which isn't logged now.
>> >>>
>> >>> Wido
>> >>>
>> >>> >
>> >>> > Thanks for all help.
>> >>> >
>> >>> > Özhan
>> >>> > _______________________________________________
>> >>> > ceph-users mailing list
>> >>> > ceph-users@xxxxxxxxxxxxxx
>> >>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >>
>> >>
>> >
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@xxxxxxxxxxxxxx
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux