Hi,
I was testing ceph cluster map changes and I got to stuck state which
seems to be indefinite.
First my description what I have done.
I'm testing special case with only one copy of pg's (pool size = 1).
All pg's was on one osd.0. I created second osd.1 and modified cluster
map to transfer one pool (metadata) to the newly created osd.1
PG's started to remap and "objects degraded number" was dropping - so
everything looked normal.
During that recovery process I restarted both osd daemons.
After that I noticed, that pg's, that should be remapped had "stale"
state - stale+active+remapped+backfilling and other object with stale
state .
I tried to run ceph pg force_create_pg on one pg, that should be
remapped, but nothing changed (that is 1 stuck / creating PG below in
ceph health)
Command rados -p metadata ls hangs so data are unavailable, but it
should be there.
What should I do in this state to get it working?
ceph -s below:
cluster 93418692-8e2e-4689-a237-ed5b47f39f72
health HEALTH_WARN 52 pgs backfill; 1 pgs backfilling; 63 pgs
stale; 1 pgs stuck inactive; 63 pgs stuck stale; 54 pgs stuck unclean;
recovery 107232/1881806 objects degraded (5.698%); mon.imatic-mce low
disk space
monmap e1: 1 mons at {imatic-mce=192.168.11.165:6789/0}, election
epoch 1, quorum 0 imatic-mce
mdsmap e450: 1/1/1 up {0=imatic-mce=up:active}
osdmap e275: 2 osds: 2 up, 2 in
pgmap v51624: 448 pgs, 4 pools, 790 GB data, 1732 kobjects
804 GB used, 2915 GB / 3720 GB avail
107232/1881806 objects degraded (5.698%)
52 stale+active+remapped+wait_backfill
1 creating
1 stale+active+remapped+backfilling
10 stale+active+clean
384 active+clean
Last message in OSD log's:
2014-11-07 22:17:45.402791 deb4db70 0 -- 192.168.11.165:6804/29564 >>
192.168.11.165:6807/29939 pipe(0x9d52f00 sd=213 :53216 s=2 pgs=1 cs=1
l=0 c=0x2c7f58c0).fault with nothing to send, going to standby
Thank you for help
With regards
Jan Pekar, ceph fan
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com