Performance really drops from 700MB/s to 10MB/s

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I use nmon on each OSD server, this is a really good tool to find out 
what is going on regarding CPU, Mem, Disks and Networking



German Anders

















> --- Original message ---
> Asunto: Re: [ceph-users] Performance really drops from 700MB/s to 
> 10MB/s
> De: Craig Lewis <clewis at centraldesktop.com>
> Para: Mariusz Gronczewski <mariusz.gronczewski at efigence.com>
> Cc: German Anders <ganders at despegar.com>, Ceph Users 
> <ceph-users at lists.ceph.com>
> Fecha: Thursday, 14/08/2014 15:42
>
> I find graphs really help here.  One screen that has all the disk I/O
> and latency for all OSDs makes it easy to pin point the bottleneck.
>
> If you don't have that, I'd go low tech: Watch the blinky lights. It's
> really easy to see which disk is the hotspot.
>
>
>
> On Thu, Aug 14, 2014 at 6:56 AM, Mariusz Gronczewski
> <mariusz.gronczewski at efigence.com> wrote:
>>
>> Actual OSD (/var/log/ceph/ceph-osd.$id) logs would be more useful.
>>
>> Few ideas:
>>
>> * do 'ceph health detail' to get detail of which OSD is stalling
>> * 'ceph osd perf' to see latency of each osd
>> * 'ceph --admin-daemon /var/run/ceph/ceph-osd.$id.asok 
>> dump_historic_ops' shows "recent slow" ops
>>
>> I actually have very similiar problem, cluster goes full speed 
>> (sometimes even for hours) and suddenly everything stops for a minute 
>> or 5, no disk IO, no IO wait (so disks are fine), no IO errors in 
>> kernel log, and OSDs only complain that other OSD subop is slow (but 
>> on that OSD everything looks fine too)
>>
>> On Wed, 13 Aug 2014 16:04:30 -0400, German Anders
>> <ganders at despegar.com> wrote:
>>
>>>
>>> Also, even a "ls -ltr" could be done inside the /mnt of the RBD that
>>> it freeze the prompt. Any ideas? I've attach some syslogs from one of
>>> the OSD servers and also from the client. Both are running Ubuntu
>>> 14.04LTS with Kernel  3.15.8.
>>> The cluster is not usable at this point, since I can't run a "ls" on
>>> the rbd.
>>>
>>> Thanks in advance,
>>>
>>> Best regards,
>>>
>>>
>>> German Anders
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>>
>>>> --- Original message ---
>>>> Asunto: Re: [ceph-users] Performance really drops from 700MB/s to
>>>> 10MB/s
>>>> De: German Anders <ganders at despegar.com>
>>>> Para: Mark Nelson <mark.nelson at inktank.com>
>>>> Cc: <ceph-users at lists.ceph.com>
>>>> Fecha: Wednesday, 13/08/2014 11:09
>>>>
>>>>
>>>> Actually is very strange, since if i run the fio test on the client,
>>>> and also un parallel run a iostat on all the OSD servers, i don't see
>>>> any workload going on over the disks, I mean... nothing! 0.00....and
>>>> also the fio script on the client is reacting very rare too:
>>>>
>>>>
>>>> $ sudo fio --filename=/dev/rbd1 --direct=1 --rw=write --bs=4m
>>>> --size=10G --iodepth=16 --ioengine=libaio --runtime=60
>>>> --group_reporting --name=file99
>>>> file99: (g=0): rw=write, bs=4M-4M/4M-4M/4M-4M, ioengine=libaio,
>>>> iodepth=16
>>>> fio-2.1.3
>>>> Starting 1 process
>>>> Jobs: 1 (f=1): [W] [2.1% done] [0KB/0KB/0KB /s] [0/0/0 iops] [eta
>>>> 01h:26m:43s]
>>>>
>>>> It's seems like is doing nothing..
>>>>
>>>>
>>>>
>>>> German Anders
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>>
>>>>> --- Original message ---
>>>>> Asunto: Re: [ceph-users] Performance really drops from 700MB/s to
>>>>> 10MB/s
>>>>> De: Mark Nelson <mark.nelson at inktank.com>
>>>>> Para: <ceph-users at lists.ceph.com>
>>>>> Fecha: Wednesday, 13/08/2014 11:00
>>>>>
>>>>> On 08/13/2014 08:19 AM, German Anders wrote:
>>>>>>
>>>>>>
>>>>>> Hi to all,
>>>>>>
>>>>>>                                I'm having a particular behavior on a 
>>>>>> new Ceph cluster.
>>>>>> I've map
>>>>>> a RBD to a client and issue some performance tests with fio, at this
>>>>>> point everything goes just fine (also the results :) ), but then I try
>>>>>> to run another new test on a new RBD on the same client, and suddenly
>>>>>> the performance goes below 10MB/s and it took almost 10 minutes to
>>>>>> complete a 10G file test, if I issue a *ceph -w* I don't see anything
>>>>>> suspicious, any idea what can be happening here?
>>>>>
>>>>> When things are going fast, are your disks actually writing data out
>>>>> as
>>>>> fast as your client IO would indicate? (don't forgot to count
>>>>> replication!)  It may be that the great speed is just writing data
>>>>> into
>>>>> the tmpfs journals (if the test is only 10GB and spread across 36
>>>>> OSDs,
>>>>> it could finish pretty quickly writing to tmpfs!).  FWIW, tmpfs
>>>>> journals
>>>>> aren't very safe.  It's not something you want to use outside of
>>>>> testing
>>>>> except in unusual circumstances.
>>>>>
>>>>> In your tests, when things are bad: it's generally worth checking to
>>>>> see
>>>>> if any one disk/osd is backed up relative to the others.  There are a
>>>>> couple of ways to accomplish this.  the Ceph admin socket can tell you
>>>>> information about each OSD ie how many outstanding IOs and a history
>>>>> of
>>>>> slow ops.  You can also look at per-disk statistics with something
>>>>> like
>>>>> iostat or collectl.
>>>>>
>>>>> Hope this helps!
>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>                                The cluster is made of:
>>>>>>
>>>>>> 3 x MON Servers
>>>>>> 4 x OSD Servers (3TB SAS 6G disks for OSD daemons & tmpfs for Journal
>>>>>> ->
>>>>>> there's one tmpfs of 36GB that is share by 9 OSD daemons, on each
>>>>>> server)
>>>>>> 2 x Network SW (Cluster and Public)
>>>>>> 10GbE speed on both networks
>>>>>>
>>>>>>                                The ceph.conf file is the following:
>>>>>>
>>>>>> [global]
>>>>>> fsid = 56e56e4c-ea59-4157-8b98-acae109bebe1
>>>>>> mon_initial_members = cephmon01, cephmon02, cephmon03
>>>>>> mon_host = 10.97.10.1,10.97.10.2,10.97.10.3
>>>>>> auth_client_required = cephx
>>>>>> auth_cluster_required = cephx
>>>>>> auth_service_required = cephx
>>>>>> filestore_xattr_use_omap = true
>>>>>> public_network = 10.97.0.0/16
>>>>>> cluster_network = 192.168.10.0/24
>>>>>> osd_pool_default_size = 2
>>>>>> glance_api_version = 2
>>>>>>
>>>>>> [mon]
>>>>>> debug_optracker = 0
>>>>>>
>>>>>> [mon.cephmon01]
>>>>>> host = cephmon01
>>>>>> mon_addr = 10.97.10.1:6789
>>>>>>
>>>>>> [mon.cephmon02]
>>>>>> host = cephmon02
>>>>>> mon_addr = 10.97.10.2:6789
>>>>>>
>>>>>> [mon.cephmon03]
>>>>>> host = cephmon03
>>>>>> mon_addr = 10.97.10.3:6789
>>>>>>
>>>>>> [osd]
>>>>>> journal_dio = false
>>>>>> osd_journal_size = 4096
>>>>>> fstype = btrfs
>>>>>> debug_optracker = 0
>>>>>>
>>>>>> [osd.0]
>>>>>> host = cephosd01
>>>>>> devs = /dev/sdc1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.1]
>>>>>> host = cephosd01
>>>>>> devs = /dev/sdd1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.2]
>>>>>> host = cephosd01
>>>>>> devs = /dev/sdf1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.3]
>>>>>> host = cephosd01
>>>>>> devs = /dev/sdg1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.4]
>>>>>> host = cephosd01
>>>>>> devs = /dev/sdi1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.5]
>>>>>> host = cephosd01
>>>>>> devs = /dev/sdj1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.6]
>>>>>> host = cephosd01
>>>>>> devs = /dev/sdl1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.7]
>>>>>> host = cephosd01
>>>>>> devs = /dev/sdm1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.8]
>>>>>> host = cephosd01
>>>>>> devs = /dev/sdn1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.9]
>>>>>> host = cephosd02
>>>>>> devs = /dev/sdc1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.10]
>>>>>> host = cephosd02
>>>>>> devs = /dev/sdd1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.11]
>>>>>> host = cephosd02
>>>>>> devs = /dev/sdf1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.12]
>>>>>> host = cephosd02
>>>>>> devs = /dev/sdg1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.13]
>>>>>> host = cephosd02
>>>>>> devs = /dev/sdi1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.14]
>>>>>> host = cephosd02
>>>>>> devs = /dev/sdj1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.15]
>>>>>> host = cephosd02
>>>>>> devs = /dev/sdl1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.16]
>>>>>> host = cephosd02
>>>>>> devs = /dev/sdm1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.17]
>>>>>> host = cephosd02
>>>>>> devs = /dev/sdn1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.18]
>>>>>> host = cephosd03
>>>>>> devs = /dev/sdc1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.19]
>>>>>> host = cephosd03
>>>>>> devs = /dev/sdd1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.20]
>>>>>> host = cephosd03
>>>>>> devs = /dev/sdf1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.21]
>>>>>> host = cephosd03
>>>>>> devs = /dev/sdg1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.22]
>>>>>> host = cephosd03
>>>>>> devs = /dev/sdi1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.23]
>>>>>> host = cephosd03
>>>>>> devs = /dev/sdj1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.24]
>>>>>> host = cephosd03
>>>>>> devs = /dev/sdl1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.25]
>>>>>> host = cephosd03
>>>>>> devs = /dev/sdm1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.26]
>>>>>> host = cephosd03
>>>>>> devs = /dev/sdn1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.27]
>>>>>> host = cephosd04
>>>>>> devs = /dev/sdc1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.28]
>>>>>> host = cephosd04
>>>>>> devs = /dev/sdd1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.29]
>>>>>> host = cephosd04
>>>>>> devs = /dev/sdf1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.30]
>>>>>> host = cephosd04
>>>>>> devs = /dev/sdg1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.31]
>>>>>> host = cephosd04
>>>>>> devs = /dev/sdi1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.32]
>>>>>> host = cephosd04
>>>>>> devs = /dev/sdj1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.33]
>>>>>> host = cephosd04
>>>>>> devs = /dev/sdl1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.34]
>>>>>> host = cephosd04
>>>>>> devs = /dev/sdm1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [osd.35]
>>>>>> host = cephosd04
>>>>>> devs = /dev/sdn1
>>>>>> osd_journal = /mnt/ramdisk/$cluster-$id-journal
>>>>>>
>>>>>> [client.volumes]
>>>>>> keyring = /etc/ceph/ceph.client.volumes.keyring
>>>>>>
>>>>>>
>>>>>> Thanks in advance,
>>>>>>
>>>>>> Best regards,
>>>>>>
>>>>>> *German Anders
>>>>>> *
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> ceph-users mailing list
>>>>>> ceph-users at lists.ceph.com
>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> ceph-users at lists.ceph.com
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list
>>>> ceph-users at lists.ceph.com
>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>
>>
>>
>>
>> --
>> Mariusz Gronczewski, Administrator
>>
>> Efigence S. A.
>> ul. Wo?oska 9a, 02-583 Warszawa
>> T: [+48] 22 380 13 13
>> F: [+48] 22 380 13 14
>> E: mariusz.gronczewski at efigence.com
>> <mailto:mariusz.gronczewski at efigence.com>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users at lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140814/b7879822/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux