I don't see anything related to lost objects in your output. I just see waiting on backfill, backfill_toofull, remapped, and so forth. You can read a bit about what is going on here: http://ceph.com/docs/next/rados/operations/monitoring-osd-pg/
Keep us posted as to the recovery, and let me know what I can do to improve the docs for scenarios like this.
On Sat, Apr 20, 2013 at 10:52 AM, Marco Aroldi <marco.aroldi@xxxxxxxxx> wrote:
John,
thanks for the quick reply.
Below you can see my ceph osd tree
The problem is caused not by the failure itself, but by the "renamed"
bunch of devices.
It was like a deadly 15-puzzle
I think that the solution was to mount the devices in fstab using UUID
(/dev/disk/by-uuid) instead of /dev/sdX
However, yes I have an entry in my ceph.conf (devs = /dev/sdX1 --
osd_journal = /dev/sdX2) *and* an entry in my fstab for each OSD
The node with failed disk is s103 (osd.59)
Now i have 5 osd from s203 up and in to try to let ceph rebalance
data... but is still a bloody mess.
Look at ceph -w output: is reported a total of 110TB: is wrong... al
drives are 2TB and i have 49 drives up and in -- total 98Tb
I think that 110TB (55 osd) was the size before cluster became inaccessible
# id weight type name up/down reweight
-1 130 root default
-9 65 room p1
-3 44 rack r14
-4 22 host s101
11 2 osd.11 up 1
12 2 osd.12 up 1
13 2 osd.13 up 1
14 2 osd.14 up 1
15 2 osd.15 up 1
16 2 osd.16 up 1
17 2 osd.17 up 1
18 2 osd.18 up 1
19 2 osd.19 up 1
20 2 osd.20 up 1
21 2 osd.21 up 1
-6 22 host s102
33 2 osd.33 up 1
34 2 osd.34 up 1
35 2 osd.35 up 1
36 2 osd.36 up 1
37 2 osd.37 up 1
38 2 osd.38 up 1
39 2 osd.39 up 1
40 2 osd.40 up 1
41 2 osd.41 up 1
42 2 osd.42 up 1
43 2 osd.43 up 1
-13 21 rack r10
-12 21 host s103
55 2 osd.55 up 0
56 2 osd.56 up 0
57 2 osd.57 up 0
58 2 osd.58 up 0
59 2 osd.59 down 0
60 2 osd.60 down 0
61 2 osd.61 down 0
62 2 osd.62 up 0
63 2 osd.63 up 0
64 1.5 osd.64 up 0
65 1.5 osd.65 down 0
-10 65 room p2
-7 22 rack r20
-5 22 host s202
22 2 osd.22 up 1
23 2 osd.23 up 1
24 2 osd.24 up 1
25 2 osd.25 up 1
26 2 osd.26 up 1
27 2 osd.27 up 1
28 2 osd.28 up 1
29 2 osd.29 up 1
30 2 osd.30 up 1
31 2 osd.31 up 1
32 2 osd.32 up 1
-8 22 rack r22
-2 22 host s201
0 2 osd.0 up 1
1 2 osd.1 up 1
2 2 osd.2 up 1
3 2 osd.3 up 1
4 2 osd.4 up 1
5 2 osd.5 up 1
6 2 osd.6 up 1
7 2 osd.7 up 1
8 2 osd.8 up 1
9 2 osd.9 up 1
10 2 osd.10 up 1
-14 21 rack r21
-11 21 host s203
44 2 osd.44 up 1
45 2 osd.45 up 1
46 2 osd.46 up 1
47 2 osd.47 up 1
48 2 osd.48 up 1
49 2 osd.49 up 0
50 2 osd.50 up 0
51 2 osd.51 up 0
52 1.5 osd.52 up 0
53 1.5 osd.53 up 0
54 2 osd.54 up 0
ceph -w
2013-04-20 19:46:48.608988 mon.0 [INF] pgmap v1352767: 17280 pgs: 58
active, 12581 active+clean, 1686 active+remapped+wait_backfill, 24
active+degraded+wait_backfill, 224
active+remapped+wait_backfill+backfill_toofull, 1061
active+recovery_wait, 4
active+degraded+wait_backfill+backfill_toofull, 629 peering, 626
active+remapped, 72 active+remapped+backfilling, 89 active+degraded,
14 active+remapped+backfill_toofull, 1 active+clean+scrubbing, 8
active+degraded+remapped+wait_backfill, 20
active+recovery_wait+remapped, 5
active+degraded+remapped+wait_backfill+backfill_toofull, 162
remapped+peering, 1 active+degraded+remapped+backfilling, 2
active+degraded+remapped+backfill_toofull, 13 active+recovering; 49777
GB data, 72863 GB used, 40568 GB / 110 TB avail; 2965687/21848501
degraded (13.574%); recovering 5 o/s, 16363B/s
2013/4/20 John Wilkins <john.wilkins@xxxxxxxxxxx>:
> Marco,
>
> If you do a "ceph tree" can you see if your OSDs are all up? You seem to
> have at least one problem related to the backfill OSDs being too full, and
> some which are near full or full for the purposes of storage. See the
> following in the documentation to see if this helps:
>
> http://ceph.com/docs/master/rados/configuration/mon-config-ref/#storage-capacity
> http://ceph.com/docs/master/rados/configuration/osd-config-ref/#backfilling
> http://ceph.com/docs/master/rados/operations/troubleshooting-osd/#no-free-drive-space
>
> Before you start deleting data as a remedy, you'd want to at least try to
> get the OSDs back up and running first.
>
> If rebooting changed the drive names, you might look here:
> http://ceph.com/docs/master/rados/configuration/osd-config-ref/#general-settings
>
> We have default settings for OSD and journal paths, which you could override
> if you can locate the data and journal sources on the renamed drives. If you
> mounted them, but didn't add them to the fstab, that might be the source of
> the problem. I'd rather see you use the default paths, as it would be easier
> to troubleshoot later. So did you mount the drives, but not add the mount
> points to fstab?
>
> John
>
>
>
>
> On Sat, Apr 20, 2013 at 8:46 AM, Marco Aroldi <marco.aroldi@xxxxxxxxx>
> wrote:
>>
>> Hi,
>> due a harware failure during expanding ceph, I'm in big trouble
>> because the cephfs doesn't mount anymore.
>> I was adding a couple storage nodes, but a disk has failed and after a
>> reboot the OS (ubuntu 12.04) renamed the remaining devices, so the
>> entire node has been screwed out.
>>
>> Now, from the "sane new node", I'm taking some new osd up and in
>> because the cluster is near full and I can't revert completely the
>> situation as before
>>
>> *I can* afford data loss, but i need to regain access to the filesystem
>>
>> My setup:
>> 3 mon + 3 mds
>> 4 storage nodes (i was adding no. 5 and 6)
>>
>> Ceph 0.56.4
>>
>>
>> ceph health:
>> HEALTH_ERR 2008 pgs backfill; 246 pgs backfill_toofull; 74 pgs
>> backfilling; 134 pgs degraded; 790 pgs peering; 10 pgs recovering;
>> 1116 pgs recovery_wait; 790 pgs stuck inactive; 4782 pgs stuck
>> unclean; recovery 3049459/21926624 degraded (13.908%); recovering 6
>> o/s, 16316KB/s; 4 full osd(s); 30 near full osd(s); full,noup,nodown
>> flag(s) set
>>
>>
>>
>> ceph mds dump:
>> dumped mdsmap epoch 44
>> epoch 44
>> flags 0
>> created 2013-03-18 14:42:29.330548
>> modified 2013-04-20 17:14:32.969332
>> tableserver 0
>> root 0
>> session_timeout 60
>> session_autoclose 300
>> last_failure 43
>> last_failure_osd_epoch 18160
>> compat compat={},rocompat={},incompat={1=base v0.20,2=client
>> writeable ranges,3=default file layouts on dirs,4=dir inode in
>> separate object}
>> max_mds 1
>> in 0
>> up {0=6376}
>> failed
>> stopped
>> data_pools [0]
>> metadata_pool 1
>> 6376: 192.168.21.11:6800/13457 'm1' mds.0.9 up:replay seq 1
>> 5945: 192.168.21.13:6800/12999 'm3' mds.-1.0 up:standby seq 1
>> 5963: 192.168.21.12:6800/22454 'm2' mds.-1.0 up:standby seq 1
>>
>>
>>
>> ceph mon dump:
>> epoch 1
>> fsid d634f7b3-8a8a-4893-bdfb-a95ccca7fddd
>> last_changed 2013-03-18 14:39:42.253923
>> created 2013-03-18 14:39:42.253923
>> 0: 192.168.21.11:6789/0 mon.m1
>> 1: 192.168.21.12:6789/0 mon.m2
>> 2: 192.168.21.13:6789/0 mon.m3
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
>
> --
> John Wilkins
> Senior Technical Writer
> Intank
> john.wilkins@xxxxxxxxxxx
> (415) 425-9599
> http://inktank.com
John Wilkins
Senior Technical Writer
Intank
john.wilkins@xxxxxxxxxxx
(415) 425-9599
http://inktank.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com