Re: OSD not coming back up again

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> Op 11 augustus 2016 om 10:46 schreef Willem Jan Withagen <wjw@xxxxxxxxxxx>:
> 
> 
> On 11-8-2016 08:26, Wido den Hollander wrote:
> > 
> >> Op 11 augustus 2016 om 2:40 schreef Willem Jan Withagen <wjw@xxxxxxxxxxx>:
> >>
> >>
> >> Hi
> >>
> >> During testing with cephtool-test-mon.sh
> >>
> >> 3 OSDs are started, and then the code executes:
> >> ====
> >>   ceph osd set noup
> >>   ceph osd down 0
> >>   ceph osd dump | grep 'osd.0 down'
> >>   ceph osd unset noup
> >> ====
> >>
> >> And in 1000 secs osd.0 is not coming back up.
> >>
> >> Below some details, but where should I start looking?
> >>
> > Can you use the admin socket to query osd.0?
> > 
> > ceph daemon osd.0 status
> > 
> > What does that tell you?
> 
> Good suggestion:
> {
>     "cluster_fsid": "0f7dc8d3-7010-4094-aa9a-a0694e7c05a6",
>     "osd_fsid": "621c9b39-b8ea-412a-a9db-115c97eafe7d",
>     "whoami": 0,
>     "state": "waiting_for_healthy",
>     "oldest_map": 1,
>     "newest_map": 176,
>     "num_pgs": 8
> }
> 
> Right before setting the osd to down, the newest map is 173.
> So some maps have been exchanged....
> 
> How does the OSD decide that it is healthy?
> If it gets peer (ping) messages that is is up?
> 

Not 100% sure, but waiting for healthy is something I haven't seen before.

Is it incrementing the newest_map when the cluster advances?

> > Maybe try debug_osd = 20
> 
> But then still I need to know what to look for, since 20 generates
> serious output.
> 
True, it will. But I don't know exactly what to look for. debug_osd = 20 might reveal more information there.

It might be a very simple log line which tells you what is going on.

Wido

> Thanx,
> --WjW
> 
> > 
> > Wido
> > 
> >> Thanx
> >> --WjW
> >>
> >>
> >> ceph -s gives:
> >>
> >>     cluster 9b2500f8-44fb-40d1-91bc-ed522e9db5c6
> >>      health HEALTH_WARN
> >>             8 pgs degraded
> >>             8 pgs stuck unclean
> >>             8 pgs undersized
> >>      monmap e1: 3 mons at
> >> {a=127.0.0.1:7202/0,b=127.0.0.1:7203/0,c=127.0.0.1:7204/0}
> >>             election epoch 6, quorum 0,1,2 a,b,c
> >>      osdmap e179: 3 osds: 2 up, 2 in; 8 remapped pgs
> >>             flags sortbitwise,require_jewel_osds,require_kraken_osds
> >>       pgmap v384: 8 pgs, 1 pools, 0 bytes data, 0 objects
> >>             248 GB used, 198 GB / 446 GB avail
> >>                    8 active+undersized+degraded
> >>
> >> And the pgmap version is slowly growing.....
> >>
> >> This set of lines is repeated over and over in the osd.0.log
> >>
> >> 2016-08-11 02:31:48.710152 b2f4d00  1 -- 127.0.0.1:0/25528 -->
> >> 127.0.0.1:6806/25709 -- osd_ping(ping e175 stamp 2016-08-11
> >> 02:31:48.710144) v2 -- ?+0 0xb42bc00 con 0xb12ba40
> >> 2016-08-11 02:31:48.710188 b2f4d00  1 -- 127.0.0.1:0/25528 -->
> >> 127.0.0.1:6807/25709 -- osd_ping(ping e175 stamp 2016-08-11
> >> 02:31:48.710144) v2 -- ?+0 0xb42cc00 con 0xb12bb20
> >> 2016-08-11 02:31:48.710214 b2f4d00  1 -- 127.0.0.1:0/25528 -->
> >> 127.0.0.1:6810/25910 -- osd_ping(ping e175 stamp 2016-08-11
> >> 02:31:48.710144) v2 -- ?+0 0xb42a400 con 0xb12bc00
> >> 2016-08-11 02:31:48.710240 b2f4d00  1 -- 127.0.0.1:0/25528 -->
> >> 127.0.0.1:6811/25910 -- osd_ping(ping e175 stamp 2016-08-11
> >> 02:31:48.710144) v2 -- ?+0 0xb42c000 con 0xb12c140
> >> 2016-08-11 02:31:48.710604 b412480  1 -- 127.0.0.1:0/25528 <== osd.1
> >> 127.0.0.1:6806/25709 284 ==== osd_ping(ping_reply e179 stamp 2016-08-11
> >> 02:31:48.710144) v2 ==== 47+0+0 (281956571 0 0) 0xb42d800 con 0xb12ba40
> >> 2016-08-11 02:31:48.710665 b486900  1 -- 127.0.0.1:0/25528 <== osd.2
> >> 127.0.0.1:6810/25910 283 ==== osd_ping(ping_reply e179 stamp 2016-08-11
> >> 02:31:48.710144) v2 ==== 47+0+0 (281956571 0 0) 0xb42d200 con 0xb12bc00
> >> 2016-08-11 02:31:48.710683 b412480  1 -- 127.0.0.1:0/25528 <== osd.1
> >> 127.0.0.1:6806/25709 285 ==== osd_ping(you_died e179 stamp 2016-08-11
> >> 02:31:48.710144) v2 ==== 47+0+0 (1545205378 0 0) 0xb42d800 con 0xb12ba40
> >> 2016-08-11 02:31:48.710780 b412000  1 -- 127.0.0.1:0/25528 <== osd.1
> >> 127.0.0.1:6807/25709 284 ==== osd_ping(ping_reply e179 stamp 2016-08-11
> >> 02:31:48.710144) v2 ==== 47+0+0 (281956571 0 0) 0xb42da00 con 0xb12bb20
> >> 2016-08-11 02:31:48.710789 b486900  1 -- 127.0.0.1:0/25528 <== osd.2
> >> 127.0.0.1:6810/25910 284 ==== osd_ping(you_died e179 stamp 2016-08-11
> >> 02:31:48.710144) v2 ==== 47+0+0 (1545205378 0 0) 0xb42d200 con 0xb12bc00
> >> 2016-08-11 02:31:48.710821 b486d80  1 -- 127.0.0.1:0/25528 <== osd.2
> >> 127.0.0.1:6811/25910 283 ==== osd_ping(ping_reply e179 stamp 2016-08-11
> >> 02:31:48.710144) v2 ==== 47+0+0 (281956571 0 0) 0xb42d400 con 0xb12c140
> >> 2016-08-11 02:31:48.710973 b412000  1 -- 127.0.0.1:0/25528 <== osd.1
> >> 127.0.0.1:6807/25709 285 ==== osd_ping(you_died e179 stamp 2016-08-11
> >> 02:31:48.710144) v2 ==== 47+0+0 (1545205378 0 0) 0xb42da00 con 0xb12bb20
> >> 2016-08-11 02:31:48.711028 b486d80  1 -- 127.0.0.1:0/25528 <== osd.2
> >> 127.0.0.1:6811/25910 284 ==== osd_ping(you_died e179 stamp 2016-08-11
> >> 02:31:48.710144) v2 ==== 47+0+0 (1545205378 0 0) 0xb42d400 con 0xb12c140
> >> --
> >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> >> the body of a message to majordomo@xxxxxxxxxxxxxxx
> >> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux