2016-03-29 14:50 GMT+08:00 Christian Balzer <chibi@xxxxxxx>: > > Hello, > > On Tue, 29 Mar 2016 14:00:44 +0800 lin zhou wrote: > >> Hi,Christian. >> When I re-add these OSD(0,3,9,12,15),the high latency occur again.the >> default reweight of these OSD is 0.0 >> > That makes no sense, at a crush weight (not reweight) of 0 they should not > get used at all. > > When you deleted the other OSD (6?) because of high latency, was your only > reason/data point the "ceph osd perf" output? because this is a near-product environment,so when the osd latency and the system latency is high.I delete these osd to let it work first. >> root@node-65:~# ceph osd tree >> # id weight type name up/down reweight >> -1 103.7 root default >> -2 8.19 host node-65 >> 18 2.73 osd.18 up 1 >> 21 0 osd.21 up 1 >> 24 2.73 osd.24 up 1 >> 27 2.73 osd.27 up 1 >> 30 0 osd.30 up 1 >> 33 0 osd.33 up 1 >> 0 0 osd.0 up 1 >> 3 0 osd.3 up 1 >> 6 0 osd.6 down 0 >> 9 0 osd.9 up 1 >> 12 0 osd.12 up 1 >> 15 0 osd.15 up 1 >> >> ceph osd perf: >> 0 9825 10211 >> 3 9398 9775 >> 9 35852 36904 >> 12 24716 25626 >> 15 18893 19633 >> > This could very well be old, stale data. > Still, these are some seriously bad numbers, if they are real. > > Do the these perf numbers change at all? My guess would be no. Yes,the are never changed. > >> but iostat of these device is empty. > Unsurprising, as they should not be used by Ceph with a weight of 0. > atop gives you an even better, complete view. > >> smartctl say nothing error found in these OSD device. >> > What exactly are these devices (model please), 3TB SATA drives Ia assume? > How are they connect (controller)? Yes,3TB SATA,Model Number: WDC WD3000FYYZ-01UL1B1 and today,I try to set osd.0 reweight to 0.1;and then check.some useful data found. avg-cpu: %user %nice %system %iowait %steal %idle 1.63 0.00 0.48 16.15 0.00 81.75 Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s avgrq-sz avgqu-sz await r_await w_await svctm %util sda 0.00 0.00 0.00 2.00 0.00 1.00 1024.00 39.85 1134.00 0.00 1134.00 500.00 100.00 sda1 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 sda2 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 sda3 0.00 0.00 0.00 0.00 0.00 0.00 0.00 1.00 0.00 0.00 0.00 0.00 100.40 sda4 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 sda5 0.00 0.00 0.00 2.00 0.00 1.00 1024.00 32.32 1134.00 0.00 1134.00 502.00 100.40 sda6 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.66 0.00 0.00 0.00 0.00 66.40 sda7 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 0.00 sda8 0.00 0.00 0.00 0.00 0.00 0.00 0.00 1.00 0.00 0.00 0.00 0.00 100.00 sda9 0.00 0.00 0.00 0.00 0.00 0.00 0.00 1.00 0.00 0.00 0.00 0.00 100.00 sda10 0.00 0.00 0.00 0.00 0.00 0.00 0.00 1.00 0.00 0.00 0.00 0.00 100.00 ^C^C^C^C^C^C^C^C^C root@node-65:~# ls -l /var/lib/ceph/osd/ceph-0 total 62924048 -rw-r--r-- 1 root root 487 Oct 12 16:49 activate.monmap -rw-r--r-- 1 root root 3 Oct 12 16:49 active -rw-r--r-- 1 root root 37 Oct 12 16:49 ceph_fsid drwxr-xr-x 280 root root 8192 Mar 30 11:58 current -rw-r--r-- 1 root root 37 Oct 12 16:49 fsid lrwxrwxrwx 1 root root 9 Oct 12 16:49 journal -> /dev/sda5 -rw------- 1 root root 56 Oct 12 16:49 keyring -rw-r--r-- 1 root root 21 Oct 12 16:49 magic -rw-r--r-- 1 root root 6 Oct 12 16:49 ready -rw-r--r-- 1 root root 4 Oct 12 16:49 store_version -rw-r--r-- 1 root root 42 Oct 12 16:49 superblock -rw-r--r-- 1 root root 64424509440 Mar 30 10:20 testfio.file -rw-r--r-- 1 root root 0 Mar 28 09:54 upstart -rw-r--r-- 1 root root 2 Oct 12 16:49 whoami the journal of osd.0 is sda5,it is so busy.and cpu wait in top is 30%.the system is slow. So,maybe this is the problem of sda5?it is INTEL SSDSC2BB120G4 we use two SSD for journal and system. root@node-65:~# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT sda 8:0 0 111.8G 0 disk ├─sda1 8:1 0 22M 0 part ├─sda2 8:2 0 191M 0 part /boot ├─sda3 8:3 0 43.9G 0 part / ├─sda4 8:4 0 3.8G 0 part [SWAP] ├─sda5 8:5 0 10.2G 0 part ├─sda6 8:6 0 10.2G 0 part ├─sda7 8:7 0 10.2G 0 part ├─sda8 8:8 0 10.2G 0 part ├─sda9 8:9 0 10.2G 0 part └─sda10 8:10 0 10.2G 0 part sdb 8:16 0 111.8G 0 disk ├─sdb1 8:17 0 24M 0 part ├─sdb2 8:18 0 10.2G 0 part ├─sdb3 8:19 0 10.2G 0 part ├─sdb4 8:20 0 10.2G 0 part ├─sdb5 8:21 0 10.2G 0 part ├─sdb6 8:22 0 10.2G 0 part ├─sdb7 8:23 0 10.2G 0 part └─sdb8 8:24 0 50.1G 0 part > Christian >> 2016-03-29 13:22 GMT+08:00 lin zhou <hnuzhoulin2@xxxxxxxxx>: >> > Thanks.I try this method just like ceph document say. >> > But I just test osd.6 in this way,and the leveldb of osd.6 is >> > broken.so it can not start. >> > >> > When I try this for other osd,it works. >> > >> > 2016-03-29 8:22 GMT+08:00 Christian Balzer <chibi@xxxxxxx>: >> >> On Mon, 28 Mar 2016 18:36:14 +0800 lin zhou wrote: >> >> >> >>> > Hello, >> >>> > >> >>> > On Sun, 27 Mar 2016 13:41:57 +0800 lin zhou wrote: >> >>> > >> >>> > > Hi,guys. >> >>> > > some days ago,one osd have a large latency seeing in ceph osd >> >>> > > perf.and this device make this node a high cpu await. >> >>> > The thing to do at that point would have been look at things with >> >>> > atop or iostat to verify that it was the device itself that was >> >>> > slow and not because it was genuinely busy due to uneven activity >> >>> > maybe. As well as a quick glance at SMART of course. >> >>> >> >>> Thanks.I will follow this when I face this problem next time. >> >>> >> >>> > > So,I delete this osd ad then check this device. >> >>> > If that device (HDD, SSD, which model?) slowed down your cluster, >> >>> > you should not have deleted it. >> >>> > The best method would have been to set your cluster to noout and >> >>> > stop that specific OSD. >> >>> > >> >>> > When you say "delete", what exact steps did you take? >> >>> > Did this include removing it from the crush map? >> >>> >> >>> Yes,I delete it from crush map.delete its auth,and rm osd. >> >>> >> >> >> >> Google is your friend, if you deleted it like in the link below you >> >> should be be able to re-add it the same way: >> >> http://lists.ceph.com/pipermail/ceph-users-ceph.com/2015-June/002345.html >> >> >> >> Christian >> >> >> >>> > > But nothing error found. >> >>> > > >> >>> > > And now I want to re-add this device into cluster with it's data. >> >>> > > >> >>> > All the data was already replicated elsewhere if you >> >>> > deleted/removed the OSD, you're likely not going to save much if >> >>> > any data movement by re-adding it. >> >>> >> >>> Yes,the cluster finished rebalance.but I face a problem of one >> >>> unfound object. And in the output of pg query in recovery_state >> >>> say,this osd is down,but other odds are ok. >> >>> So I want to recover this osd to recover this unfound object. >> >>> >> >>> and mark_unfound_lost revert/delete do not work: >> >>> Error EINVAL: pg has 1 unfound objects but we haven't probed all >> >>> sources, >> >>> >> >>> detail see: >> >>> http://lists.ceph.com/pipermail/ceph-users-ceph.com/2016-March/008452.html >> >>> >> >>> Thanks again. >> >>> >> >>> > > >> >>> > > I try to using ceph-osd to add it,but it can not start.log are >> >>> > > paste in : >> >>> > > https://gist.github.com/hnuzhoulin/836f9e633b90041e89ad >> >>> > > >> >>> > > so what's the recommend steps. >> >>> > That depends on how you deleted it, but at this point your data is >> >>> > likely to be mostly stale anyway, so I'd start from scratch. >> >>> >> >>> > Christian >> >>> > -- >> >>> > Christian Balzer Network/Systems Engineer >> >>> > chibi@xxxxxxx Global OnLine Japan/Rakuten Communications >> >>> > http://www.gol.com/ >> >>> > >> >>> >> >> >> >> >> >> -- >> >> Christian Balzer Network/Systems Engineer >> >> chibi@xxxxxxx Global OnLine Japan/Rakuten Communications >> >> http://www.gol.com/ >> > > > -- > Christian Balzer Network/Systems Engineer > chibi@xxxxxxx Global OnLine Japan/Rakuten Communications > http://www.gol.com/ _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com