active+remapped after remove osd via ceph osd out

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



After replace broken disk and ceph osd in it, cluster:
ceph health detail
HEALTH_WARN 2 pgs stuck unclean; recovery 60/346857819 degraded (0.000%)
pg 3.884 is stuck unclean for 570722.873270, current state
active+remapped, last acting [143,261,314]
pg 3.154a is stuck unclean for 577659.917066, current state
active+remapped, last acting [85,224,64]
recovery 60/346857819 degraded (0.000%)

What can be wrong?
It is possible this is caused by 'ceph osd reweight-by-utilization' ?

More info:
ceph -v
ceph version 0.67.9 (ba340a97c3dafc9155023da8d515eecc675c619a)

Enabled tunnables:
# begin crush map
tunable choose_local_tries 0
tunable choose_local_fallback_tries 0
tunable choose_total_tries 50
tunable chooseleaf_descend_once 1

df osd:
143 - 78%
261 - 78%
314 - 80%

85 - 76%
224  76%
64 - 75%

ceph osd dump | grep -i pool
pool 0 'data' rep size 3 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 64 pgp_num 64 last_change 28459 owner 0
crash_replay_interval 45
pool 1 'metadata' rep size 3 min_size 1 crush_ruleset 1 object_hash
rjenkins pg_num 64 pgp_num 64 last_change 28460 owner 0
pool 2 'rbd' rep size 3 min_size 1 crush_ruleset 2 object_hash
rjenkins pg_num 64 pgp_num 64 last_change 28461 owner 0
pool 3 '.rgw.buckets' rep size 3 min_size 1 crush_ruleset 0
object_hash rjenkins pg_num 8192 pgp_num 8192 last_change 73711 owner
0
pool 4 '.log' rep size 3 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 2048 pgp_num 2048 last_change 90517 owner 0
pool 5 '.rgw' rep size 3 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 128 pgp_num 128 last_change 72467 owner 0
pool 6 '.users.uid' rep size 3 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 8 pgp_num 8 last_change 28465 owner 0
pool 7 '.users' rep size 3 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 8 pgp_num 8 last_change 28466 owner 0
pool 8 '.usage' rep size 2 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 8 pgp_num 8 last_change 28467 owner
18446744073709551615
pool 9 '.intent-log' rep size 3 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 8 pgp_num 8 last_change 28468 owner
18446744073709551615
pool 10 '.rgw.control' rep size 3 min_size 1 crush_ruleset 0
object_hash rjenkins pg_num 8 pgp_num 8 last_change 33485 owner
18446744073709551615
pool 11 '.rgw.gc' rep size 3 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 8 pgp_num 8 last_change 33487 owner
18446744073709551615
pool 12 '.rgw.root' rep size 2 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 8 pgp_num 8 last_change 44540 owner 0
pool 13 '' rep size 2 min_size 1 crush_ruleset 0 object_hash rjenkins
pg_num 8 pgp_num 8 last_change 46912 owner 0

ceph pg 3.884 query
{ "state": "active+remapped",
  "epoch": 160655,
  "up": [
        143],
  "acting": [
        143,
        261,
        314],
  "info": { "pgid": "3.884",
      "last_update": "160655'111533",
      "last_complete": "160655'111533",
      "log_tail": "159997'108532",
      "last_backfill": "MAX",
      "purged_snaps": "[]",
      "history": { "epoch_created": 4,
          "last_epoch_started": 160261,
          "last_epoch_clean": 160261,
          "last_epoch_split": 11488,
          "same_up_since": 160252,
          "same_interval_since": 160260,
          "same_primary_since": 160252,
          "last_scrub": "155516'107396",
          "last_scrub_stamp": "2014-08-06 03:15:18.193611",
          "last_deep_scrub": "155516'107293",
          "last_deep_scrub_stamp": "2014-08-03 06:45:59.215397",
          "last_clean_scrub_stamp": "2014-08-06 03:15:18.193611"},
      "stats": { "version": "160655'111533",
          "reported_seq": "856860",
          "reported_epoch": "160655",
          "state": "active+remapped",
          "last_fresh": "2014-08-18 23:06:47.068588",
          "last_change": "2014-08-17 21:12:29.452628",
          "last_active": "2014-08-18 23:06:47.068588",
          "last_clean": "2014-08-12 08:44:00.293916",
          "last_became_active": "2013-10-25 14:54:55.902442",
          "last_unstale": "2014-08-18 23:06:47.068588",
          "mapping_epoch": 160258,
          "log_start": "159997'108532",
          "ondisk_log_start": "159997'108532",
          "created": 4,
          "last_epoch_clean": 160261,
          "parent": "0.0",
          "parent_split_bits": 0,
          "last_scrub": "155516'107396",
          "last_scrub_stamp": "2014-08-06 03:15:18.193611",
          "last_deep_scrub": "155516'107293",
          "last_deep_scrub_stamp": "2014-08-03 06:45:59.215397",
          "last_clean_scrub_stamp": "2014-08-06 03:15:18.193611",
          "log_size": 3001,
          "ondisk_log_size": 3001,
          "stats_invalid": "0",
          "stat_sum": { "num_bytes": 2750235192,
              "num_objects": 12015,
              "num_object_clones": 0,
              "num_object_copies": 0,
              "num_objects_missing_on_primary": 0,
              "num_objects_degraded": 0,
              "num_objects_unfound": 0,
              "num_read": 708045,
              "num_read_kb": 39418032,
              "num_write": 120983,
              "num_write_kb": 2383937,
              "num_scrub_errors": 0,
              "num_shallow_scrub_errors": 0,
              "num_deep_scrub_errors": 0,
              "num_objects_recovered": 44904,
              "num_bytes_recovered": 7915543525,
              "num_keys_recovered": 0},
          "stat_cat_sum": {},
          "up": [
                143],
          "acting": [
                143,
                261,
                314]},
      "empty": 0,
      "dne": 0,
      "incomplete": 0,
      "last_epoch_started": 160261},
  "recovery_state": [
        { "name": "Started\/Primary\/Active",
          "enter_time": "2014-08-17 21:12:29.452429",
          "might_have_unfound": [],
          "recovery_progress": { "backfill_target": -1,
              "waiting_on_backfill": 0,
              "backfill_pos": "0\/\/0\/\/-1",
              "backfill_info": { "begin": "0\/\/0\/\/-1",
                  "end": "0\/\/0\/\/-1",
                  "objects": []},
              "peer_backfill_info": { "begin": "0\/\/0\/\/-1",
                  "end": "0\/\/0\/\/-1",
                  "objects": []},
              "backfills_in_flight": [],
              "pull_from_peer": [],
              "pushing": []},
          "scrub": { "scrubber.epoch_start": "0",
              "scrubber.active": 0,
              "scrubber.block_writes": 0,
              "scrubber.finalizing": 0,
              "scrubber.waiting_on": 0,
              "scrubber.waiting_on_whom": []}},
        { "name": "Started",
          "enter_time": "2014-08-17 21:12:28.436021"}]}

---
Regards
Dominik

2014-08-17 21:57 GMT+02:00 Dominik Mostowiec <dominikmostowiec at gmail.com>:
> Hi,
> After ceph osd out ( 1 osd ) cluster stopped rebalancing on
> 10621 active+clean, 2 active+remapped, 1 active+degraded+remapped;
>
> My crushmap is clean, there is not 'empty' device's.
> grep device /tmp/crush1.txt  | grep -v osd | grep -v '^#' | wc -l
> 0
>
> Can You help me with this?
>
>   "up": [
>         73],
>   "acting": [
>         73,
>         102],
> I have only one copy of this PG ?
>
> More info:
> --
> ceph health detail
> HEALTH_WARN 1 pgs degraded; 3 pgs stuck unclean; recovery
> 12008/346501095 degraded (0.003%)
> pg 3.884 is stuck unclean for 478441.392837, current state
> active+remapped, last acting [143,261,314]
> pg 3.154a is stuck unclean for 485378.436630, current state
> active+remapped, last acting [85,224,64]
> pg 3.cc7 is stuck unclean for 116231.803324, current state
> active+degraded+remapped, last acting [73,102]
> pg 3.cc7 is active+degraded+remapped, acting [73,102]
> recovery 12008/346501095 degraded (0.003%)
> --
> ceph pg dump | grep 3.cc7
> dumped all in format plain
> 3.cc7   12014   0       12012   0       2845541648      3870    3870
>  active+degraded+remapped        2014-08-17 21:08:04.155348
> 160273'273322   160273:1044675  [73]    [73,102]159997'270388
> 2014-08-13 05:23:48.386184      159997'270388   2014-08-13
> 05:23:48.386184
> --
> grep '3.cc7' /var/log/ceph/ceph-osd.73.log
> 2014-08-17 21:06:47.494511 7f788a625700 20 osd.73 160241  kicking pg 3.cc7
> 2014-08-17 21:06:47.494513 7f788a625700 30 osd.73 pg_epoch: 160241
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160026 n=12016 ec=4 les/c 160026/160026
> 160024/160025/153162) [73]/[73,102] r=0 lpr=160025 mlcod 160241'273319
> active+degraded+remapped] lock
> 2014-08-17 21:06:47.494522 7f788a625700 10 osd.73 pg_epoch: 160241
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160026 n=12016 ec=4 les/c 160026/160026
> 160024/160025/153162) [73]/[73,102] r=0 lpr=160025 mlcod 160241'273319
> active+degraded+remapped] on_shutdown
> 2014-08-17 21:06:47.494530 7f788a625700 10 osd.73 pg_epoch: 160241
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160026 n=12016 ec=4 les/c 160026/160026
> 160024/160025/153162) [73]/[73,102] r=0 lpr=160025 mlcod 160241'273319
> active+degraded+remapped] clear_primary_state
> 2014-08-17 21:06:47.494541 7f788a625700 10 osd.73 pg_epoch: 160241
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160026 n=12016 ec=4 les/c 160026/160026
> 160024/160025/153162) [73]/[73,102] r=0 lpr=160025 luod=0'0 mlcod 0'0
> active+degraded+remapped] cancel_recovery
> 2014-08-17 21:06:47.494548 7f788a625700 10 osd.73 pg_epoch: 160241
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160026 n=12016 ec=4 les/c 160026/160026
> 160024/160025/153162) [73]/[73,102] r=0 lpr=160025 luod=0'0 mlcod 0'0
> active+degraded+remapped] clear_recovery_state
> 2014-08-17 21:07:00.758061 7f9819814700  1 osd.73 pg_epoch: 160244
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160026 n=12016 ec=4 les/c 160026/160026
> 160244/160244/160244) [73]/[73,102] r=0 lpr=160244 pi=160025-160243/2
> lcod 0'0 mlcod 0'0 remapped] state<Start>: transitioning to Primary
> 2014-08-17 21:07:51.121028 7f9819814700  1 osd.73 pg_epoch: 160246
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160245 n=12016 ec=4 les/c 160245/160245
> 160244/160246/160244) [73] r=0 lpr=160246 pi=160244-160245/1 lcod 0'0
> mlcod 0'0 inactive] state<Start>: transitioning to Primary
> 2014-08-17 21:08:02.995105 7f9818011700  1 osd.73 pg_epoch: 160248
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160247 n=12016 ec=4 les/c 160247/160247
> 160244/160248/160244) [73]/[73,102] r=0 lpr=160248 pi=160246-160247/1
> lcod 0'0 mlcod 0'0 remapped] state<Start>: transitioning to Primary
> --
> grep '3.cc7' /var/log/ceph/ceph-osd.102.log
> 2014-08-17 21:06:47.554359 7f630df7a700  1 osd.102 pg_epoch: 160242
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160026 n=12016 ec=4 les/c 160026/160026
> 160242/160242/160242) []/[102] r=0 lpr=160242 pi=158292-160241/12 lcod
> 160241'273319 mlcod 0'0 remapped] state<Start>: transitioning to
> Primary
> 2014-08-17 21:07:00.772420 7f630b775700  1 osd.102 pg_epoch: 160244
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160243 n=12016 ec=4 les/c 160243/160243
> 160244/160244/160244) [73]/[73,102] r=1 lpr=160244 pi=160242-160243/1
> lcod 160241'273319 remapped NOTIFY] state<Start>: transitioning to
> Stray
> 2014-08-17 21:07:50.832077 7f62f878a700 20 osd.102 160245  kicking pg 3.cc7
> 2014-08-17 21:07:50.832079 7f62f878a700 30 osd.102 pg_epoch: 160245
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160245 n=12016 ec=4 les/c 160245/160245
> 160244/160244/160244) [73]/[73,102] r=1 lpr=160244 pi=160242-160243/1
> luod=0'0 lcod 160241'273319 active+remapped] lock
> 2014-08-17 21:07:50.832089 7f62f878a700 10 osd.102 pg_epoch: 160245
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160245 n=12016 ec=4 les/c 160245/160245
> 160244/160244/160244) [73]/[73,102] r=1 lpr=160244 pi=160242-160243/1
> luod=0'0 lcod 160241'273319 active+remapped] on_shutdown
> 2014-08-17 21:07:50.832099 7f62f878a700 10 osd.102 pg_epoch: 160245
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160245 n=12016 ec=4 les/c 160245/160245
> 160244/160244/160244) [73]/[73,102] r=1 lpr=160244 pi=160242-160243/1
> luod=0'0 lcod 160241'273319 active+remapped] clear_primary_state
> 2014-08-17 21:07:50.832109 7f62f878a700 10 osd.102 pg_epoch: 160245
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160245 n=12016 ec=4 les/c 160245/160245
> 160244/160244/160244) [73]/[73,102] r=1 lpr=160244 pi=160242-160243/1
> luod=0'0 lcod 160241'273319 active+remapped] cancel_recovery
> 2014-08-17 21:07:50.832117 7f62f878a700 10 osd.102 pg_epoch: 160245
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160245 n=12016 ec=4 les/c 160245/160245
> 160244/160244/160244) [73]/[73,102] r=1 lpr=160244 pi=160242-160243/1
> luod=0'0 lcod 160241'273319 active+remapped] clear_recovery_state
> 2014-08-17 21:08:02.979471 7f3d54953700  1 osd.102 pg_epoch: 160248
> pg[3.cc7( v 160241'273320 (155516'269452,160241'273320]
> local-les=160245 n=12016 ec=4 les/c 160245/160245
> 160244/160248/160244) [73]/[73,102] r=1 lpr=160248 pi=160242-160247/3
> lcod 0'0 remapped NOTIFY] state<Start>: transitioning to Stray
> --
> ceph pg 3.cc7 query:
> { "state": "active+degraded+remapped",
>   "epoch": 160273,
>   "up": [
>         73],
>   "acting": [
>         73,
>         102],
>   "info": { "pgid": "3.cc7",
>       "last_update": "160273'273322",
>       "last_complete": "160273'273322",
>       "log_tail": "155516'269452",
>       "last_backfill": "MAX",
>       "purged_snaps": "[]",
>       "history": { "epoch_created": 4,
>           "last_epoch_started": 160249,
>           "last_epoch_clean": 160249,
>           "last_epoch_split": 11503,
>           "same_up_since": 160244,
>           "same_interval_since": 160248,
>           "same_primary_since": 160244,
>           "last_scrub": "159997'270388",
>           "last_scrub_stamp": "2014-08-13 05:23:48.386184",
>           "last_deep_scrub": "159997'270388",
>           "last_deep_scrub_stamp": "2014-08-13 05:23:48.386184",
>           "last_clean_scrub_stamp": "2014-08-13 05:23:48.386184"},
>       "stats": { "version": "160273'273322",
>           "reported_seq": "1044675",
>           "reported_epoch": "160273",
>           "state": "active+degraded+remapped",
>           "last_fresh": "2014-08-17 21:25:34.935269",
>           "last_change": "2014-08-17 21:08:04.155348",
>           "last_active": "2014-08-17 21:25:34.935269",
>           "last_clean": "2014-08-16 13:20:49.883438",
>           "last_became_active": "2013-10-25 13:05:26.849618",
>           "last_unstale": "2014-08-17 21:25:34.935269",
>           "mapping_epoch": 160246,
>           "log_start": "155516'269452",
>           "ondisk_log_start": "155516'269452",
>           "created": 4,
>           "last_epoch_clean": 160249,
>           "parent": "0.0",
>           "parent_split_bits": 0,
>           "last_scrub": "159997'270388",
>           "last_scrub_stamp": "2014-08-13 05:23:48.386184",
>           "last_deep_scrub": "159997'270388",
>           "last_deep_scrub_stamp": "2014-08-13 05:23:48.386184",
>           "last_clean_scrub_stamp": "2014-08-13 05:23:48.386184",
>           "log_size": 3870,
>           "ondisk_log_size": 3870,
>           "stats_invalid": "0",
>           "stat_sum": { "num_bytes": 2845541648,
>               "num_objects": 12014,
>               "num_object_clones": 0,
>               "num_object_copies": 0,
>               "num_objects_missing_on_primary": 0,
>               "num_objects_degraded": 0,
>               "num_objects_unfound": 0,
>               "num_read": 723032,
>               "num_read_kb": 24658206,
>               "num_write": 118401,
>               "num_write_kb": 2360009,
>               "num_scrub_errors": 0,
>               "num_shallow_scrub_errors": 0,
>               "num_deep_scrub_errors": 0,
>               "num_objects_recovered": 55614,
>               "num_bytes_recovered": 10782825899,
>               "num_keys_recovered": 0},
>           "stat_cat_sum": {},
>           "up": [
>                 73],
>           "acting": [
>                 73,
>                 102]},
>       "empty": 0,
>       "dne": 0,
>       "incomplete": 0,
>       "last_epoch_started": 160249},
>   "recovery_state": [
>         { "name": "Started\/Primary\/Active",
>           "enter_time": "2014-08-17 21:08:04.154871",
>           "might_have_unfound": [],
>           "recovery_progress": { "backfill_target": -1,
>               "waiting_on_backfill": 0,
>               "backfill_pos": "0\/\/0\/\/-1",
>               "backfill_info": { "begin": "0\/\/0\/\/-1",
>                   "end": "0\/\/0\/\/-1",
>                   "objects": []},
>               "peer_backfill_info": { "begin": "0\/\/0\/\/-1",
>                   "end": "0\/\/0\/\/-1",
>                   "objects": []},
>               "backfills_in_flight": [],
>               "pull_from_peer": [],
>               "pushing": []},
>           "scrub": { "scrubber.epoch_start": "0",
>               "scrubber.active": 0,
>               "scrubber.block_writes": 0,
>               "scrubber.finalizing": 0,
>               "scrubber.waiting_on": 0,
>               "scrubber.waiting_on_whom": []}},
>         { "name": "Started",
>           "enter_time": "2014-08-17 21:08:02.995104"}]}
>
>
> --
> Regards
> Dominik



-- 
Pozdrawiam
Dominik


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux