Re: two osd stack on peereng after start osd to recovery

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,
Something interesting, osd whith problems eats much more memory.
Standard is about 300m,
This osd eats even 30G.

Can i do any tests to help find where the problem is?

--
Regards
Dominik

2013/7/16 Dominik Mostowiec <dominikmostowiec@xxxxxxxxx>:
> Hi,
> I noticed that problem is more frequent at nigth where traffic is smaller.
> Mabye it is caused by scrubbing (multiple scrubbings on one osd) and
> to small "filestore op threads" or something other num threads
> settings in my config:
>         osd heartbeat grace = 15
>         filestore flush min = 0
>         filestore flusher = false
>         filestore fiemap = false
>         filestore op threads = 4
>         filestore queue max ops = 4096
>         filestore queue max bytes = 10485760
>         filestore queue committing max bytes = 10485760
>         osd op threads = 8
>         osd disk threads = 4
>         osd recovery threads = 1
>         osd recovery max active = 1
>         osd recovery op priority = 1
>         osd client op priority = 100
>         osd max backfills = 1
> --
> Regards
> Dominik
>
> 2013/7/4 Dominik Mostowiec <dominikmostowiec@xxxxxxxxx>:
>> I reported bug: http://tracker.ceph.com/issues/5504
>>
>> --
>> Regards
>> Dominik
>>
>> 2013/7/2 Dominik Mostowiec <dominikmostowiec@xxxxxxxxx>:
>>> Hi,
>>> Some osd.87 performance graphs:
>>> https://www.dropbox.com/s/o07wae2041hu06l/osd_87_performance.PNG
>>> After 11.05 I have restarted it.
>>>
>>> Mons .., maybe this is the problem.
>>>
>>> --
>>> Regards
>>> Dominik
>>>
>>> 2013/7/2 Andrey Korolyov <andrey@xxxxxxx>:
>>>> Hi Dominik,
>>>>
>>>> What`s about performance on the osd.87 at this moment, do you have any
>>>> related measurements?
>>>>
>>>> As for mine version of this issue, seems that quorum has some kind of
>>>> degradation over time - when I restarted mons, problem has gone and peering
>>>> time lowered by factor of ten or so.  Also seems that the problem has a
>>>> cumulative origin in the quorum - I did disk replacement over last week and
>>>> every time peering gets worse and worse.  I assume that it`s a time to put
>>>> more or less formalized problems to the bugtracker:
>>>> - such degradation over a time plus stuck placement groups,
>>>> - newer kind of problem related to the epochs too - restarting one mon
>>>> resulting to slight dataplacement change at the moment when _first rebooted_
>>>> monitor came up, not shown up with one hour delays between quorum restart.
>>>>
>>>>
>>>>
>>>>
>>>> On Tue, Jul 2, 2013 at 1:37 PM, Dominik Mostowiec
>>>> <dominikmostowiec@xxxxxxxxx> wrote:
>>>>>
>>>>> Hi,
>>>>> I got it.
>>>>>
>>>>> ceph health details
>>>>> HEALTH_WARN 3 pgs peering; 3 pgs stuck inactive; 5 pgs stuck unclean;
>>>>> recovery 64/38277874 degraded (0.000%)
>>>>> pg 5.df9 is stuck inactive for 138669.746512, current state peering,
>>>>> last acting [87,2,151]
>>>>> pg 5.a82 is stuck inactive for 138638.121867, current state peering,
>>>>> last acting [151,87,42]
>>>>> pg 5.80d is stuck inactive for 138621.069523, current state peering,
>>>>> last acting [151,47,87]
>>>>> pg 5.df9 is stuck unclean for 138669.746761, current state peering,
>>>>> last acting [87,2,151]
>>>>> pg 5.ae2 is stuck unclean for 139479.810499, current state active,
>>>>> last acting [87,151,28]
>>>>> pg 5.7b6 is stuck unclean for 139479.693271, current state active,
>>>>> last acting [87,105,2]
>>>>> pg 5.a82 is stuck unclean for 139479.713859, current state peering,
>>>>> last acting [151,87,42]
>>>>> pg 5.80d is stuck unclean for 139479.800820, current state peering,
>>>>> last acting [151,47,87]
>>>>> pg 5.df9 is peering, acting [87,2,151]
>>>>> pg 5.a82 is peering, acting [151,87,42]
>>>>> pg 5.80d is peering, acting [151,47,87]
>>>>> recovery 64/38277874 degraded (0.000%)
>>>>>
>>>>>
>>>>> osd pg query for 5.df9:
>>>>> { "state": "peering",
>>>>>   "up": [
>>>>>         87,
>>>>>         2,
>>>>>         151],
>>>>>   "acting": [
>>>>>         87,
>>>>>         2,
>>>>>         151],
>>>>>   "info": { "pgid": "5.df9",
>>>>>       "last_update": "119454'58844953",
>>>>>       "last_complete": "119454'58844953",
>>>>>       "log_tail": "119454'58843952",
>>>>>       "last_backfill": "MAX",
>>>>>       "purged_snaps": "[]",
>>>>>       "history": { "epoch_created": 365,
>>>>>           "last_epoch_started": 119456,
>>>>>           "last_epoch_clean": 119456,
>>>>>           "last_epoch_split": 117806,
>>>>>           "same_up_since": 119458,
>>>>>           "same_interval_since": 119458,
>>>>>           "same_primary_since": 119458,
>>>>>           "last_scrub": "119442'58732630",
>>>>>           "last_scrub_stamp": "2013-06-29 20:02:24.817352",
>>>>>           "last_deep_scrub": "119271'57224023",
>>>>>           "last_deep_scrub_stamp": "2013-06-23 02:04:49.654373",
>>>>>           "last_clean_scrub_stamp": "2013-06-29 20:02:24.817352"},
>>>>>       "stats": { "version": "119454'58844953",
>>>>>           "reported": "119458'42382189",
>>>>>           "state": "peering",
>>>>>           "last_fresh": "2013-06-30 20:35:29.489826",
>>>>>           "last_change": "2013-06-30 20:35:28.469854",
>>>>>           "last_active": "2013-06-30 20:33:24.126599",
>>>>>           "last_clean": "2013-06-30 20:33:24.126599",
>>>>>           "last_unstale": "2013-06-30 20:35:29.489826",
>>>>>           "mapping_epoch": 119455,
>>>>>           "log_start": "119454'58843952",
>>>>>           "ondisk_log_start": "119454'58843952",
>>>>>           "created": 365,
>>>>>           "last_epoch_clean": 365,
>>>>>           "parent": "0.0",
>>>>>           "parent_split_bits": 0,
>>>>>           "last_scrub": "119442'58732630",
>>>>>           "last_scrub_stamp": "2013-06-29 20:02:24.817352",
>>>>>           "last_deep_scrub": "119271'57224023",
>>>>>           "last_deep_scrub_stamp": "2013-06-23 02:04:49.654373",
>>>>>           "last_clean_scrub_stamp": "2013-06-29 20:02:24.817352",
>>>>>           "log_size": 135341,
>>>>>           "ondisk_log_size": 135341,
>>>>>           "stats_invalid": "0",
>>>>>           "stat_sum": { "num_bytes": 1010563373,
>>>>>               "num_objects": 3099,
>>>>>               "num_object_clones": 0,
>>>>>               "num_object_copies": 0,
>>>>>               "num_objects_missing_on_primary": 0,
>>>>>               "num_objects_degraded": 0,
>>>>>               "num_objects_unfound": 0,
>>>>>               "num_read": 302,
>>>>>               "num_read_kb": 0,
>>>>>               "num_write": 32264,
>>>>>               "num_write_kb": 798650,
>>>>>               "num_scrub_errors": 0,
>>>>>               "num_objects_recovered": 8235,
>>>>>               "num_bytes_recovered": 2085653757,
>>>>>               "num_keys_recovered": 249061471},
>>>>>           "stat_cat_sum": {},
>>>>>           "up": [
>>>>>                 87,
>>>>>                 2,
>>>>>                 151],
>>>>>           "acting": [
>>>>>                 87,
>>>>>                 2,
>>>>>                 151]},
>>>>>       "empty": 0,
>>>>>       "dne": 0,
>>>>>       "incomplete": 0,
>>>>>       "last_epoch_started": 119454},
>>>>>   "recovery_state": [
>>>>>         { "name": "Started\/Primary\/Peering\/GetLog",
>>>>>           "enter_time": "2013-06-30 20:35:28.545478",
>>>>>           "newest_update_osd": 2},
>>>>>         { "name": "Started\/Primary\/Peering",
>>>>>           "enter_time": "2013-06-30 20:35:28.469841",
>>>>>           "past_intervals": [
>>>>>                 { "first": 119453,
>>>>>                   "last": 119454,
>>>>>                   "maybe_went_rw": 1,
>>>>>                   "up": [
>>>>>                         87,
>>>>>                         2,
>>>>>                         151],
>>>>>                   "acting": [
>>>>>                         87,
>>>>>                         2,
>>>>>                         151]},
>>>>>                 { "first": 119455,
>>>>>                   "last": 119457,
>>>>>                   "maybe_went_rw": 1,
>>>>>                   "up": [
>>>>>                         2,
>>>>>                         151],
>>>>>                   "acting": [
>>>>>                         2,
>>>>>                         151]}],
>>>>>           "probing_osds": [
>>>>>                 2,
>>>>>                 87,
>>>>>                 151],
>>>>>           "down_osds_we_would_probe": [],
>>>>>           "peering_blocked_by": []},
>>>>>         { "name": "Started",
>>>>>           "enter_time": "2013-06-30 20:35:28.469765"}]}
>>>>>
>>>>>
>>>>> For other PGs: https://www.dropbox.com/s/q5iv8lwzecioy3d/pg_query.tar.tz
>>>>>
>>>>> --
>>>>> Regards
>>>>> Dominik
>>>>>
>>>>> 2013/6/30 Andrey Korolyov <andrey@xxxxxxx>:
>>>>> > That`s not a loop as it looks, sorry  - I had reproduced issue many
>>>>> > times and there is no such cpu-eating behavior in most cases, only
>>>>> > locked pgs are presented. Also I may celebrate returning of 'wrong
>>>>> > down mark' bug, at least for the 0.61.4 tag. For first one, I`ll send
>>>>> > a link with core as quick as I will be able to reproduce it on my test
>>>>> > env, and second one linked with 100% disk utilization, so I`m not sure
>>>>> > if this is right behavior or wrong.
>>>>> >
>>>>> > On Sat, Jun 29, 2013 at 1:28 AM, Sage Weil <sage@xxxxxxxxxxx> wrote:
>>>>> >> On Sat, 29 Jun 2013, Andrey Korolyov wrote:
>>>>> >>> There is almost same problem with the 0.61 cluster, at least with same
>>>>> >>> symptoms. Could be reproduced quite easily - remove an osd and then
>>>>> >>> mark it as out and with quite high probability one of neighbors will
>>>>> >>> be stuck at the end of peering process with couple of peering pgs with
>>>>> >>> primary copy on it. Such osd process seems to be stuck in some kind of
>>>>> >>> lock, eating exactly 100% of one core.
>>>>> >>
>>>>> >> Which version?
>>>>> >> Can you attach with gdb and get a backtrace to see what it is chewing
>>>>> >> on?
>>>>> >>
>>>>> >> Thanks!
>>>>> >> sage
>>>>> >>
>>>>> >>
>>>>> >>>
>>>>> >>> On Thu, Jun 13, 2013 at 8:42 PM, Gregory Farnum <greg@xxxxxxxxxxx>
>>>>> >>> wrote:
>>>>> >>> > On Thu, Jun 13, 2013 at 6:33 AM, S?awomir Skowron <szibis@xxxxxxxxx>
>>>>> >>> > wrote:
>>>>> >>> >> Hi, sorry for late response.
>>>>> >>> >>
>>>>> >>> >> https://docs.google.com/file/d/0B9xDdJXMieKEdHFRYnBfT3lCYm8/view
>>>>> >>> >>
>>>>> >>> >> Logs in attachment, and on google drive, from today.
>>>>> >>> >>
>>>>> >>> >> https://docs.google.com/file/d/0B9xDdJXMieKEQzVNVHJ1RXFXZlU/view
>>>>> >>> >>
>>>>> >>> >> We have such problem today. And new logs are on google drive with
>>>>> >>> >> today date.
>>>>> >>> >>
>>>>> >>> >> Strange is that problematic osd.71 have about 10-15%, more space
>>>>> >>> >> used
>>>>> >>> >> then other osd in cluster.
>>>>> >>> >>
>>>>> >>> >> Today in one hour osd.71 fails 3 times in mon log, and after third
>>>>> >>> >> recovery has been stuck, and many 500 errors appears in http layer
>>>>> >>> >> on
>>>>> >>> >> top of rgw. When it's stuck, restarting osd71, osd.23, and osd.108,
>>>>> >>> >> all from stucked pg, helps, but i run even repair on this osd, just
>>>>> >>> >> in
>>>>> >>> >> case.
>>>>> >>> >>
>>>>> >>> >> I have some theory, that on this pg is rgw index of objects, or one
>>>>> >>> >> of
>>>>> >>> >> osd in this pg, have some problems with local filesystem or drive
>>>>> >>> >> bellow (raid controller reports nothing about that), but i do not
>>>>> >>> >> see
>>>>> >>> >> any problem in system.
>>>>> >>> >>
>>>>> >>> >> How can we find in which pg/osd index of objects in rgw bucket
>>>>> >>> >> exist ??
>>>>> >>> >
>>>>> >>> > You can find the location of any named object by grabbing the OSD
>>>>> >>> > map
>>>>> >>> > from the cluster and using the osdmaptool: "osdmaptool <mapfile>
>>>>> >>> > --test-map-object <objname> --pool <poolid>".
>>>>> >>> >
>>>>> >>> > You're not providing any context for your issue though, so we really
>>>>> >>> > can't help. What symptoms are you observing?
>>>>> >>> > -Greg
>>>>> >>> > Software Engineer #42 @ http://inktank.com | http://ceph.com
>>>>> >>> > _______________________________________________
>>>>> >>> > ceph-users mailing list
>>>>> >>> > ceph-users@xxxxxxxxxxxxxx
>>>>> >>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>> >>> _______________________________________________
>>>>> >>> ceph-users mailing list
>>>>> >>> ceph-users@xxxxxxxxxxxxxx
>>>>> >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>> >>>
>>>>> > _______________________________________________
>>>>> > ceph-users mailing list
>>>>> > ceph-users@xxxxxxxxxxxxxx
>>>>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Pozdrawiam
>>>>> Dominik
>>>>
>>>>
>>>
>>>
>>>
>>> --
>>> Pozdrawiam
>>> Dominik
>>
>>
>>
>> --
>> Pozdrawiam
>> Dominik
>
>
>
> --
> Pozdrawiam
> Dominik



-- 
Pozdrawiam
Dominik
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux