active+remapped after remove osd via ceph osd out

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,
After set chooseleaf_descend_once=0, and migration 20% PGs ceph is HEALTH_OK.
"chooseleaf_descend_once" optimal value is 1 :-(

--
Regards
Dominik


2014-08-21 15:59 GMT+02:00 Dominik Mostowiec <dominikmostowiec at gmail.com>:
> Hi,
> I have 2 PG in active+remapped state.
>
> ceph health detail
> HEALTH_WARN 2 pgs stuck unclean; recovery 24/348041229 degraded (0.000%)
> pg 3.1a07 is stuck unclean for 29239.046024, current state
> active+remapped, last acting [167,80,145]
> pg 3.154a is stuck unclean for 29239.039777, current state
> active+remapped, last acting [377,224,292]
> recovery 24/348041229 degraded (0.000%)
>
> This happend when i call "ceph osd reweight-by-utilization 102"
>
> What can be wrong ?
>
> ceph -v -> ceph version 0.67.10 (9d446bd416c52cd785ccf048ca67737ceafcdd7f)
>
> Tunables:
> ceph osd crush dump | tail -n 4
>   "tunables": { "choose_local_tries": 0,
>       "choose_local_fallback_tries": 0,
>       "choose_total_tries": 60,
>       "chooseleaf_descend_once": 1}}
>
> Cluster:
> 6 racks X 3 hosts X 22 OSDs. (396 osds: 396 up, 396 in)
>
>> crushtool -i ../crush2  --min-x 0 --num-rep 3  --max-x 10624 --test --show-bad-mappings
> is clean.
>
> When 'ceph osd reweight' for all osd is 1.0 is ok, but i have nearfull OSD's.
>
> There is no missing OSD's in crushmap
>> grep device /tmp/crush.txt | grep -v osd
> # devices
>
> ceph osd dump | grep -i pool
> pool 0 'data' rep size 3 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 28459 owner 0
> crash_replay_interval 45
> pool 1 'metadata' rep size 3 min_size 1 crush_ruleset 1 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 28460 owner 0
> pool 2 'rbd' rep size 3 min_size 1 crush_ruleset 2 object_hash
> rjenkins pg_num 64 pgp_num 64 last_change 28461 owner 0
> pool 3 '.rgw.buckets' rep size 3 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8192 pgp_num 8192 last_change 73711 owner
> 0
> pool 4 '.log' rep size 3 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 2048 pgp_num 2048 last_change 90517 owner 0
> pool 5 '.rgw' rep size 3 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 128 pgp_num 128 last_change 72467 owner 0
> pool 6 '.users.uid' rep size 3 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 28465 owner 0
> pool 7 '.users' rep size 3 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 28466 owner 0
> pool 8 '.usage' rep size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 28467 owner
> 18446744073709551615
> pool 9 '.intent-log' rep size 3 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 28468 owner
> 18446744073709551615
> pool 10 '.rgw.control' rep size 3 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 33485 owner
> 18446744073709551615
> pool 11 '.rgw.gc' rep size 3 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 33487 owner
> 18446744073709551615
> pool 12 '.rgw.root' rep size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 44540 owner 0
> pool 13 '' rep size 2 min_size 1 crush_ruleset 0 object_hash rjenkins
> pg_num 8 pgp_num 8 last_change 46912 owner 0
>
>> ceph pg 3.1a07 query
> { "state": "active+remapped",
>   "epoch": 181721,
>   "up": [
>         167,
>         80],
>   "acting": [
>         167,
>         80,
>         145],
>   "info": { "pgid": "3.1a07",
>       "last_update": "181719'94809",
>       "last_complete": "181719'94809",
>       "log_tail": "159997'91808",
>       "last_backfill": "MAX",
>       "purged_snaps": "[]",
>       "history": { "epoch_created": 4,
>           "last_epoch_started": 179611,
>           "last_epoch_clean": 179611,
>           "last_epoch_split": 11522,
>           "same_up_since": 179610,
>           "same_interval_since": 179610,
>           "same_primary_since": 179610,
>           "last_scrub": "160655'94695",
>           "last_scrub_stamp": "2014-08-19 04:16:20.308318",
>           "last_deep_scrub": "158290'91157",
>           "last_deep_scrub_stamp": "2014-08-12 05:15:25.557591",
>           "last_clean_scrub_stamp": "2014-08-19 04:16:20.308318"},
>       "stats": { "version": "181719'94809",
>           "reported_seq": "995830",
>           "reported_epoch": "181721",
>           "state": "active+remapped",
>           "last_fresh": "2014-08-21 14:53:14.050284",
>           "last_change": "2014-08-21 09:42:07.473356",
>           "last_active": "2014-08-21 14:53:14.050284",
>           "last_clean": "2014-08-21 07:38:51.366084",
>           "last_became_active": "2013-10-25 13:59:36.125019",
>           "last_unstale": "2014-08-21 14:53:14.050284",
>           "mapping_epoch": 179606,
>           "log_start": "159997'91808",
>           "ondisk_log_start": "159997'91808",
>           "created": 4,
>           "last_epoch_clean": 179611,
>           "parent": "0.0",
>           "parent_split_bits": 0,
>           "last_scrub": "160655'94695",
>           "last_scrub_stamp": "2014-08-19 04:16:20.308318",
>           "last_deep_scrub": "158290'91157",
>           "last_deep_scrub_stamp": "2014-08-12 05:15:25.557591",
>           "last_clean_scrub_stamp": "2014-08-19 04:16:20.308318",
>           "log_size": 3001,
>           "ondisk_log_size": 3001,
>           "stats_invalid": "0",
>           "stat_sum": { "num_bytes": 2880784014,
>               "num_objects": 12108,
>               "num_object_clones": 0,
>               "num_object_copies": 0,
>               "num_objects_missing_on_primary": 0,
>               "num_objects_degraded": 0,
>               "num_objects_unfound": 0,
>               "num_read": 645471,
>               "num_read_kb": 16973620,
>               "num_write": 111416,
>               "num_write_kb": 2459459,
>               "num_scrub_errors": 0,
>               "num_shallow_scrub_errors": 0,
>               "num_deep_scrub_errors": 0,
>               "num_objects_recovered": 48440,
>               "num_bytes_recovered": 10006953676,
>               "num_keys_recovered": 0},
>           "stat_cat_sum": {},
>           "up": [
>                 167,
>                 80],
>           "acting": [
>                 167,
>                 80,
>                 145]},
>       "empty": 0,
>       "dne": 0,
>       "incomplete": 0,
>       "last_epoch_started": 179611},
>   "recovery_state": [
>         { "name": "Started\/Primary\/Active",
>           "enter_time": "2014-08-21 09:42:07.473030",
>           "might_have_unfound": [],
>           "recovery_progress": { "backfill_target": -1,
>               "waiting_on_backfill": 0,
>               "backfill_pos": "0\/\/0\/\/-1",
>               "backfill_info": { "begin": "0\/\/0\/\/-1",
>                   "end": "0\/\/0\/\/-1",
>                   "objects": []},
>               "peer_backfill_info": { "begin": "0\/\/0\/\/-1",
>                   "end": "0\/\/0\/\/-1",
>                   "objects": []},
>               "backfills_in_flight": [],
>               "pull_from_peer": [],
>               "pushing": []},
>           "scrub": { "scrubber.epoch_start": "0",
>               "scrubber.active": 0,
>               "scrubber.block_writes": 0,
>               "scrubber.finalizing": 0,
>               "scrubber.waiting_on": 0,
>               "scrubber.waiting_on_whom": []}},
>         { "name": "Started",
>           "enter_time": "2014-08-21 09:42:06.410951"}]}
>
> --
> Regards
> Dominik
>
> 2014-08-18 23:27 GMT+02:00 Dominik Mostowiec <dominikmostowiec at gmail.com>:
>> After replace broken disk and ceph osd in it, cluster:
>> ceph health detail
>> HEALTH_WARN 2 pgs stuck unclean; recovery 60/346857819 degraded (0.000%)
>> pg 3.884 is stuck unclean for 570722.873270, current state
>> active+remapped, last acting [143,261,314]
>> pg 3.154a is stuck unclean for 577659.917066, current state
>> active+remapped, last acting [85,224,64]
>> recovery 60/346857819 degraded (0.000%)
>>
>> What can be wrong?
>> It is possible this is caused by 'ceph osd reweight-by-utilization' ?
>>
>> More info:
>> ceph -v
>> ceph version 0.67.9 (ba340a97c3dafc9155023da8d515eecc675c619a)
>>
>> Enabled tunnables:
>> # begin crush map
>> tunable choose_local_tries 0
>> tunable choose_local_fallback_tries 0
>> tunable choose_total_tries 50
>> tunable chooseleaf_descend_once 1
>>
>> df osd:
>> 143 - 78%
>> 261 - 78%
>> 314 - 80%
>>
>> 85 - 76%
>> 224  76%
>> 64 - 75%
>>
>> ceph osd dump | grep -i pool
>> pool 0 'data' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 64 pgp_num 64 last_change 28459 owner 0
>> crash_replay_interval 45
>> pool 1 'metadata' rep size 3 min_size 1 crush_ruleset 1 object_hash
>> rjenkins pg_num 64 pgp_num 64 last_change 28460 owner 0
>> pool 2 'rbd' rep size 3 min_size 1 crush_ruleset 2 object_hash
>> rjenkins pg_num 64 pgp_num 64 last_change 28461 owner 0
>> pool 3 '.rgw.buckets' rep size 3 min_size 1 crush_ruleset 0
>> object_hash rjenkins pg_num 8192 pgp_num 8192 last_change 73711 owner
>> 0
>> pool 4 '.log' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 2048 pgp_num 2048 last_change 90517 owner 0
>> pool 5 '.rgw' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 128 pgp_num 128 last_change 72467 owner 0
>> pool 6 '.users.uid' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 28465 owner 0
>> pool 7 '.users' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 28466 owner 0
>> pool 8 '.usage' rep size 2 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 28467 owner
>> 18446744073709551615
>> pool 9 '.intent-log' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 28468 owner
>> 18446744073709551615
>> pool 10 '.rgw.control' rep size 3 min_size 1 crush_ruleset 0
>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 33485 owner
>> 18446744073709551615
>> pool 11 '.rgw.gc' rep size 3 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 33487 owner
>> 18446744073709551615
>> pool 12 '.rgw.root' rep size 2 min_size 1 crush_ruleset 0 object_hash
>> rjenkins pg_num 8 pgp_num 8 last_change 44540 owner 0
>> pool 13 '' rep size 2 min_size 1 crush_ruleset 0 object_hash rjenkins
>> pg_num 8 pgp_num 8 last_change 46912 owner 0
>>
>> ceph pg 3.884 query
>> { "state": "active+remapped",
>>   "epoch": 160655,
>>   "up": [
>>         143],
>>   "acting": [
>>         143,
>>         261,
>>         314],
>>   "info": { "pgid": "3.884",
>>       "last_update": "160655'111533",
>>       "last_complete": "160655'111533",
>>       "log_tail": "159997'108532",
>>       "last_backfill": "MAX",
>>       "purged_snaps": "[]",
>>       "history": { "epoch_created": 4,
>>           "last_epoch_started": 160261,
>>           "last_epoch_clean": 160261,
>>           "last_epoch_split": 11488,
>>           "same_up_since": 160252,
>>           "same_interval_since": 160260,
>>           "same_primary_since": 160252,
>>           "last_scrub": "155516'107396",
>>           "last_scrub_stamp": "2014-08-06 03:15:18.193611",
>>           "last_deep_scrub": "155516'107293",
>>           "last_deep_scrub_stamp": "2014-08-03 06:45:59.215397",
>>           "last_clean_scrub_stamp": "2014-08-06 03:15:18.193611"},
>>       "stats": { "version": "160655'111533",
>>           "reported_seq": "856860",
>>           "reported_epoch": "160655",
>>           "state": "active+remapped",
>>           "last_fresh": "2014-08-18 23:06:47.068588",
>>           "last_change": "2014-08-17 21:12:29.452628",
>>           "last_active": "2014-08-18 23:06:47.068588",
>>           "last_clean": "2014-08-12 08:44:00.293916",
>>           "last_became_active": "2013-10-25 14:54:55.902442",
>>           "last_unstale": "2014-08-18 23:06:47.068588",
>>           "mapping_epoch": 160258,
>>           "log_start": "159997'108532",
>>           "ondisk_log_start": "159997'108532",
>>           "created": 4,
>>           "last_epoch_clean": 160261,
>>           "parent": "0.0",
>>           "parent_split_bits": 0,
>>           "last_scrub": "155516'107396",
>>           "last_scrub_stamp": "2014-08-06 03:15:18.193611",
>>           "last_deep_scrub": "155516'107293",
>>           "last_deep_scrub_stamp": "2014-08-03 06:45:59.215397",
>>           "last_clean_scrub_stamp": "2014-08-06 03:15:18.193611",
>>           "log_size": 3001,
>>           "ondisk_log_size": 3001,
>>           "stats_invalid": "0",
>>           "stat_sum": { "num_bytes": 2750235192,
>>               "num_objects": 12015,
>>               "num_object_clones": 0,
>>               "num_object_copies": 0,
>>               "num_objects_missing_on_primary": 0,
>>               "num_objects_degraded": 0,
>>               "num_objects_unfound": 0,
>>               "num_read": 708045,
>>               "num_read_kb": 39418032,
>>               "num_write": 120983,
>>               "num_write_kb": 2383937,
>>               "num_scrub_errors": 0,
>>               "num_shallow_scrub_errors": 0,
>>               "num_deep_scrub_errors": 0,
>>               "num_objects_recovered": 44904,
>>               "num_bytes_recovered": 7915543525,
>>               "num_keys_recovered": 0},
>>           "stat_cat_sum": {},
>>           "up": [
>>                 143],
>>           "acting": [
>>                 143,
>>                 261,
>>                 314]},
>>       "empty": 0,
>>       "dne": 0,
>>       "incomplete": 0,
>>       "last_epoch_started": 160261},
>>   "recovery_state": [
>>         { "name": "Started\/Primary\/Active",
>>           "enter_time": "2014-08-17 21:12:29.452429",
>>           "might_have_unfound": [],
>>           "recovery_progress": { "backfill_target": -1,
>>               "waiting_on_backfill": 0,
>>               "backfill_pos": "0\/\/0\/\/-1",
>>               "backfill_info": { "begin": "0\/\/0\/\/-1",
>>                   "end": "0\/\/0\/\/-1",
>>                   "objects": []},
>>               "peer_backfill_info": { "begin": "0\/\/0\/\/-1",
>>                   "end": "0\/\/0\/\/-1",
>>                   "objects": []},
>>               "backfills_in_flight": [],
>>               "pull_from_peer": [],
>>               "pushing": []},
>>           "scrub": { "scrubber.epoch_start": "0",
>>               "scrubber.active": 0,
>>               "scrubber.block_writes": 0,
>>               "scrubber.finalizing": 0,
>>               "scrubber.waiting_on": 0,
>>               "scrubber.waiting_on_whom": []}},
>>         { "name": "Started",
>>           "enter_time": "2014-08-17 21:12:28.436021"}]}
>>
>> ---
>> Regards
>> Dominik
>>
>> 2014-08-17 21:57 GMT+02:00 Dominik Mostowiec <dominikmostowiec at gmail.com>:
>>> Hi,
>>> After ceph osd out ( 1 osd ) cluster stopped rebalancing on
>>> 10621 active+clean, 2 active+remapped, 1 active+degraded+remapped;
>>>
>>> My crushmap is clean, there is not 'empty' device's.
>>> grep device /tmp/crush1.txt  | grep -v osd | grep -v '^#' | wc -l
>>> 0
>>>
>>> Can You help me with this?
>>>
>>>   "up": [
>>>         73],
>>>   "acting": [
>>>         73,
>>>         102],
>>> I have only one copy of this PG ?
>>>
>>> More info:
>>> --
>>> ceph health detail
>>> HEALTH_WARN 1 pgs degraded; 3 pgs stuck unclean; recovery
>>> 12008/346501095 degraded (0.003%)
>>> pg 3.884 is stuck unclean for 478441.392837, current state
>>> active+remapped, last acting [143,261,314]
>>> pg 3.154a is stuck unclean for 485378.436630, current state
>>> active+remapped, last acting [85,224,64]
>>> pg 3.cc7 is stuck unclean for 116231.803324, current state
>>> active+degraded+remapped, last acting [73,102]
>>> pg 3.cc7 is active+degraded+remapped, acting [73,102]
>>> recovery 12008/346501095 degraded (0.003%)
>>> --
>>> ceph pg dump | grep 3.cc7
>>> dumped all in format plain
>>> 3.cc7   12014   0       12012   0       2845541648      3870    3870
>>>  active+degraded+remapped        2014-08-17 21:08:04.155348
>>> 160273'273322   160273:1044675  [73]    [73,102]159997'270388
>>> 2014-08-13 05:23:48.386184      159997'270388   2014-08-13
>>> 05:23:48.386184
>>> --
>>> grep '3.cc7' /var/log/ceph/ceph-osd.73.log
>>> 2014-08-17 21:06:47.494511 7f788a625700 20 osd.73 160241  kicking pg 3.cc7
>>> 2014-08-17 21:06:47.494513 7f788a625700 30 osd.73 pg_epoch: 160241
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160026 n=12016 ec=4 les/c 160026/160026
>>> 160024/160025/153162) [73]/[73,102] r=0 lpr=160025 mlcod 160241'273319
>>> active+degraded+remapped] lock
>>> 2014-08-17 21:06:47.494522 7f788a625700 10 osd.73 pg_epoch: 160241
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160026 n=12016 ec=4 les/c 160026/160026
>>> 160024/160025/153162) [73]/[73,102] r=0 lpr=160025 mlcod 160241'273319
>>> active+degraded+remapped] on_shutdown
>>> 2014-08-17 21:06:47.494530 7f788a625700 10 osd.73 pg_epoch: 160241
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160026 n=12016 ec=4 les/c 160026/160026
>>> 160024/160025/153162) [73]/[73,102] r=0 lpr=160025 mlcod 160241'273319
>>> active+degraded+remapped] clear_primary_state
>>> 2014-08-17 21:06:47.494541 7f788a625700 10 osd.73 pg_epoch: 160241
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160026 n=12016 ec=4 les/c 160026/160026
>>> 160024/160025/153162) [73]/[73,102] r=0 lpr=160025 luod=0'0 mlcod 0'0
>>> active+degraded+remapped] cancel_recovery
>>> 2014-08-17 21:06:47.494548 7f788a625700 10 osd.73 pg_epoch: 160241
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160026 n=12016 ec=4 les/c 160026/160026
>>> 160024/160025/153162) [73]/[73,102] r=0 lpr=160025 luod=0'0 mlcod 0'0
>>> active+degraded+remapped] clear_recovery_state
>>> 2014-08-17 21:07:00.758061 7f9819814700  1 osd.73 pg_epoch: 160244
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160026 n=12016 ec=4 les/c 160026/160026
>>> 160244/160244/160244) [73]/[73,102] r=0 lpr=160244 pi=160025-160243/2
>>> lcod 0'0 mlcod 0'0 remapped] state<Start>: transitioning to Primary
>>> 2014-08-17 21:07:51.121028 7f9819814700  1 osd.73 pg_epoch: 160246
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160245 n=12016 ec=4 les/c 160245/160245
>>> 160244/160246/160244) [73] r=0 lpr=160246 pi=160244-160245/1 lcod 0'0
>>> mlcod 0'0 inactive] state<Start>: transitioning to Primary
>>> 2014-08-17 21:08:02.995105 7f9818011700  1 osd.73 pg_epoch: 160248
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160247 n=12016 ec=4 les/c 160247/160247
>>> 160244/160248/160244) [73]/[73,102] r=0 lpr=160248 pi=160246-160247/1
>>> lcod 0'0 mlcod 0'0 remapped] state<Start>: transitioning to Primary
>>> --
>>> grep '3.cc7' /var/log/ceph/ceph-osd.102.log
>>> 2014-08-17 21:06:47.554359 7f630df7a700  1 osd.102 pg_epoch: 160242
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160026 n=12016 ec=4 les/c 160026/160026
>>> 160242/160242/160242) []/[102] r=0 lpr=160242 pi=158292-160241/12 lcod
>>> 160241'273319 mlcod 0'0 remapped] state<Start>: transitioning to
>>> Primary
>>> 2014-08-17 21:07:00.772420 7f630b775700  1 osd.102 pg_epoch: 160244
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160243 n=12016 ec=4 les/c 160243/160243
>>> 160244/160244/160244) [73]/[73,102] r=1 lpr=160244 pi=160242-160243/1
>>> lcod 160241'273319 remapped NOTIFY] state<Start>: transitioning to
>>> Stray
>>> 2014-08-17 21:07:50.832077 7f62f878a700 20 osd.102 160245  kicking pg 3.cc7
>>> 2014-08-17 21:07:50.832079 7f62f878a700 30 osd.102 pg_epoch: 160245
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160245 n=12016 ec=4 les/c 160245/160245
>>> 160244/160244/160244) [73]/[73,102] r=1 lpr=160244 pi=160242-160243/1
>>> luod=0'0 lcod 160241'273319 active+remapped] lock
>>> 2014-08-17 21:07:50.832089 7f62f878a700 10 osd.102 pg_epoch: 160245
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160245 n=12016 ec=4 les/c 160245/160245
>>> 160244/160244/160244) [73]/[73,102] r=1 lpr=160244 pi=160242-160243/1
>>> luod=0'0 lcod 160241'273319 active+remapped] on_shutdown
>>> 2014-08-17 21:07:50.832099 7f62f878a700 10 osd.102 pg_epoch: 160245
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160245 n=12016 ec=4 les/c 160245/160245
>>> 160244/160244/160244) [73]/[73,102] r=1 lpr=160244 pi=160242-160243/1
>>> luod=0'0 lcod 160241'273319 active+remapped] clear_primary_state
>>> 2014-08-17 21:07:50.832109 7f62f878a700 10 osd.102 pg_epoch: 160245
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160245 n=12016 ec=4 les/c 160245/160245
>>> 160244/160244/160244) [73]/[73,102] r=1 lpr=160244 pi=160242-160243/1
>>> luod=0'0 lcod 160241'273319 active+remapped] cancel_recovery
>>> 2014-08-17 21:07:50.832117 7f62f878a700 10 osd.102 pg_epoch: 160245
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160245 n=12016 ec=4 les/c 160245/160245
>>> 160244/160244/160244) [73]/[73,102] r=1 lpr=160244 pi=160242-160243/1
>>> luod=0'0 lcod 160241'273319 active+remapped] clear_recovery_state
>>> 2014-08-17 21:08:02.979471 7f3d54953700  1 osd.102 pg_epoch: 160248
>>> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
>>> local-les=160245 n=12016 ec=4 les/c 160245/160245
>>> 160244/160248/160244) [73]/[73,102] r=1 lpr=160248 pi=160242-160247/3
>>> lcod 0'0 remapped NOTIFY] state<Start>: transitioning to Stray
>>> --
>>> ceph pg 3.cc7 query:
>>> { "state": "active+degraded+remapped",
>>>   "epoch": 160273,
>>>   "up": [
>>>         73],
>>>   "acting": [
>>>         73,
>>>         102],
>>>   "info": { "pgid": "3.cc7",
>>>       "last_update": "160273'273322",
>>>       "last_complete": "160273'273322",
>>>       "log_tail": "155516'269452",
>>>       "last_backfill": "MAX",
>>>       "purged_snaps": "[]",
>>>       "history": { "epoch_created": 4,
>>>           "last_epoch_started": 160249,
>>>           "last_epoch_clean": 160249,
>>>           "last_epoch_split": 11503,
>>>           "same_up_since": 160244,
>>>           "same_interval_since": 160248,
>>>           "same_primary_since": 160244,
>>>           "last_scrub": "159997'270388",
>>>           "last_scrub_stamp": "2014-08-13 05:23:48.386184",
>>>           "last_deep_scrub": "159997'270388",
>>>           "last_deep_scrub_stamp": "2014-08-13 05:23:48.386184",
>>>           "last_clean_scrub_stamp": "2014-08-13 05:23:48.386184"},
>>>       "stats": { "version": "160273'273322",
>>>           "reported_seq": "1044675",
>>>           "reported_epoch": "160273",
>>>           "state": "active+degraded+remapped",
>>>           "last_fresh": "2014-08-17 21:25:34.935269",
>>>           "last_change": "2014-08-17 21:08:04.155348",
>>>           "last_active": "2014-08-17 21:25:34.935269",
>>>           "last_clean": "2014-08-16 13:20:49.883438",
>>>           "last_became_active": "2013-10-25 13:05:26.849618",
>>>           "last_unstale": "2014-08-17 21:25:34.935269",
>>>           "mapping_epoch": 160246,
>>>           "log_start": "155516'269452",
>>>           "ondisk_log_start": "155516'269452",
>>>           "created": 4,
>>>           "last_epoch_clean": 160249,
>>>           "parent": "0.0",
>>>           "parent_split_bits": 0,
>>>           "last_scrub": "159997'270388",
>>>           "last_scrub_stamp": "2014-08-13 05:23:48.386184",
>>>           "last_deep_scrub": "159997'270388",
>>>           "last_deep_scrub_stamp": "2014-08-13 05:23:48.386184",
>>>           "last_clean_scrub_stamp": "2014-08-13 05:23:48.386184",
>>>           "log_size": 3870,
>>>           "ondisk_log_size": 3870,
>>>           "stats_invalid": "0",
>>>           "stat_sum": { "num_bytes": 2845541648,
>>>               "num_objects": 12014,
>>>               "num_object_clones": 0,
>>>               "num_object_copies": 0,
>>>               "num_objects_missing_on_primary": 0,
>>>               "num_objects_degraded": 0,
>>>               "num_objects_unfound": 0,
>>>               "num_read": 723032,
>>>               "num_read_kb": 24658206,
>>>               "num_write": 118401,
>>>               "num_write_kb": 2360009,
>>>               "num_scrub_errors": 0,
>>>               "num_shallow_scrub_errors": 0,
>>>               "num_deep_scrub_errors": 0,
>>>               "num_objects_recovered": 55614,
>>>               "num_bytes_recovered": 10782825899,
>>>               "num_keys_recovered": 0},
>>>           "stat_cat_sum": {},
>>>           "up": [
>>>                 73],
>>>           "acting": [
>>>                 73,
>>>                 102]},
>>>       "empty": 0,
>>>       "dne": 0,
>>>       "incomplete": 0,
>>>       "last_epoch_started": 160249},
>>>   "recovery_state": [
>>>         { "name": "Started\/Primary\/Active",
>>>           "enter_time": "2014-08-17 21:08:04.154871",
>>>           "might_have_unfound": [],
>>>           "recovery_progress": { "backfill_target": -1,
>>>               "waiting_on_backfill": 0,
>>>               "backfill_pos": "0\/\/0\/\/-1",
>>>               "backfill_info": { "begin": "0\/\/0\/\/-1",
>>>                   "end": "0\/\/0\/\/-1",
>>>                   "objects": []},
>>>               "peer_backfill_info": { "begin": "0\/\/0\/\/-1",
>>>                   "end": "0\/\/0\/\/-1",
>>>                   "objects": []},
>>>               "backfills_in_flight": [],
>>>               "pull_from_peer": [],
>>>               "pushing": []},
>>>           "scrub": { "scrubber.epoch_start": "0",
>>>               "scrubber.active": 0,
>>>               "scrubber.block_writes": 0,
>>>               "scrubber.finalizing": 0,
>>>               "scrubber.waiting_on": 0,
>>>               "scrubber.waiting_on_whom": []}},
>>>         { "name": "Started",
>>>           "enter_time": "2014-08-17 21:08:02.995104"}]}
>>>
>>>
>>> --
>>> Regards
>>> Dominik
>>
>>
>>
>> --
>> Pozdrawiam
>> Dominik
>
>
>
> --
> Pozdrawiam
> Dominik



-- 
Pozdrawiam
Dominik


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux