Do you have pg_autoscaler enabled or the balancer module?
Zitat von Ml Ml <mliebherr99@xxxxxxxxxxxxxx>:
Hello,
ceph is stuck since 4 days with 0.064% misplaced and i dunno why. Can
anyone help me to get it fixed?
I did restart some OSDs and reweight them again to get some data
moving but that did not help.
root@node01:~ # ceph -s
cluster:
id: 251c937e-0b55-48c1-8f34-96e84e4023d4
health: HEALTH_WARN
1803/2799972 objects misplaced (0.064%)
mon node02 is low on available space
services:
mon: 3 daemons, quorum node01,node02,node03
mgr: node03(active), standbys: node01, node02
osd: 16 osds: 16 up, 16 in; 1 remapped pgs
data:
pools: 1 pools, 512 pgs
objects: 933.32k objects, 2.68TiB
usage: 9.54TiB used, 5.34TiB / 14.9TiB avail
pgs: 1803/2799972 objects misplaced (0.064%)
511 active+clean
1 active+clean+remapped
io:
client: 131KiB/s rd, 8.57MiB/s wr, 28op/s rd, 847op/s wr
root@node01:~ # ceph health detail
HEALTH_WARN 1803/2800179 objects misplaced (0.064%); mon node02 is low
on available space
OBJECT_MISPLACED 1803/2800179 objects misplaced (0.064%)
MON_DISK_LOW mon node02 is low on available space
mon.node02 has 28% avail
root@node01:~ # ceph versions
{
"mon": {
"ceph version 12.2.13 (98af9a6b9a46b2d562a0de4b09263d70aeb1c9dd)
luminous (stable)": 3
},
"mgr": {
"ceph version 12.2.13 (98af9a6b9a46b2d562a0de4b09263d70aeb1c9dd)
luminous (stable)": 3
},
"osd": {
"ceph version 12.2.13 (98af9a6b9a46b2d562a0de4b09263d70aeb1c9dd)
luminous (stable)": 16
},
"mds": {},
"overall": {
"ceph version 12.2.13 (98af9a6b9a46b2d562a0de4b09263d70aeb1c9dd)
luminous (stable)": 22
}
}
root@node02:~ # df -h
Filesystem Size Used Avail Use% Mounted on
udev 63G 0 63G 0% /dev
tmpfs 13G 1.3G 12G 11% /run
/dev/sda3 46G 31G 14G 70% /
tmpfs 63G 57M 63G 1% /dev/shm
tmpfs 5.0M 0 5.0M 0% /run/lock
tmpfs 63G 0 63G 0% /sys/fs/cgroup
/dev/sda1 922M 206M 653M 24% /boot
/dev/fuse 30M 144K 30M 1% /etc/pve
/dev/sde1 93M 5.4M 88M 6% /var/lib/ceph/osd/ceph-11
/dev/sdf1 93M 5.4M 88M 6% /var/lib/ceph/osd/ceph-14
/dev/sdc1 889G 676G 214G 77% /var/lib/ceph/osd/ceph-3
/dev/sdb1 889G 667G 222G 76% /var/lib/ceph/osd/ceph-2
/dev/sdd1 93M 5.4M 88M 6% /var/lib/ceph/osd/ceph-7
tmpfs 13G 0 13G 0% /run/user/0
root@node02:~ # ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 14.34781 root default
-2 4.25287 host node01
0 hdd 0.85999 osd.0 up 0.80005 1.00000
1 hdd 0.86749 osd.1 up 0.85004 1.00000
6 hdd 0.87270 osd.6 up 0.90002 1.00000
12 hdd 0.78000 osd.12 up 0.95001 1.00000
13 hdd 0.87270 osd.13 up 0.95001 1.00000
-3 3.91808 host node02
2 hdd 0.70000 osd.2 up 0.80005 1.00000
3 hdd 0.59999 osd.3 up 0.85004 1.00000
7 hdd 0.87270 osd.7 up 0.85004 1.00000
11 hdd 0.87270 osd.11 up 0.75006 1.00000
14 hdd 0.87270 osd.14 up 0.85004 1.00000
-4 6.17686 host node03
4 hdd 0.87000 osd.4 up 1.00000 1.00000
5 hdd 0.87000 osd.5 up 1.00000 1.00000
8 hdd 0.87270 osd.8 up 1.00000 1.00000
10 hdd 0.87270 osd.10 up 1.00000 1.00000
15 hdd 0.87270 osd.15 up 1.00000 1.00000
16 hdd 1.81879 osd.16 up 1.00000 1.00000
root@node01:~ # ceph osd df tree
ID CLASS WEIGHT REWEIGHT SIZE USE DATA OMAP META
AVAIL %USE VAR PGS TYPE NAME
-1 14.55780 - 14.9TiB 9.45TiB 7.46TiB 1.47GiB 23.2GiB
5.43TiB 63.52 1.00 - root default
-2 4.27286 - 4.35TiB 3.15TiB 2.41TiB 486MiB 7.62GiB
1.21TiB 72.32 1.14 - host node01
0 hdd 0.85999 0.80005 888GiB 619GiB 269GiB 92.3MiB 0B
269GiB 69.72 1.10 89 osd.0
1 hdd 0.86749 0.85004 888GiB 641GiB 248GiB 109MiB 0B
248GiB 72.12 1.14 92 osd.1
6 hdd 0.87270 0.90002 894GiB 634GiB 632GiB 98.9MiB 2.65GiB
259GiB 70.99 1.12 107 osd.6
12 hdd 0.79999 0.95001 894GiB 664GiB 661GiB 94.4MiB 2.52GiB
230GiB 74.31 1.17 112 osd.12
13 hdd 0.87270 0.95001 894GiB 665GiB 663GiB 91.7MiB 2.46GiB
229GiB 74.43 1.17 112 osd.13
-3 4.10808 - 4.35TiB 3.17TiB 2.18TiB 479MiB 6.99GiB
1.18TiB 72.86 1.15 - host node02
2 hdd 0.78999 0.75006 888GiB 654GiB 235GiB 95.6MiB 0B
235GiB 73.57 1.16 94 osd.2
3 hdd 0.70000 0.80005 888GiB 737GiB 151GiB 114MiB 0B
151GiB 82.98 1.31 105 osd.3
7 hdd 0.87270 0.85004 894GiB 612GiB 610GiB 88.9MiB 2.43GiB
281GiB 68.50 1.08 103 osd.7
11 hdd 0.87270 0.75006 894GiB 576GiB 574GiB 81.8MiB 2.19GiB
317GiB 64.47 1.01 97 osd.11
14 hdd 0.87270 0.85004 894GiB 669GiB 666GiB 98.8MiB 2.37GiB
225GiB 74.85 1.18 112 osd.14
-4 6.17686 - 6.17TiB 3.13TiB 2.86TiB 541MiB 8.58GiB
3.04TiB 50.73 0.80 - host node03
4 hdd 0.87000 1.00000 888GiB 504GiB 384GiB 124MiB 0B
384GiB 56.72 0.89 72 osd.4
5 hdd 0.87000 1.00000 888GiB 520GiB 368GiB 96.2MiB 0B
368GiB 58.57 0.92 75 osd.5
8 hdd 0.87270 1.00000 894GiB 508GiB 505GiB 80.2MiB 2.07GiB
386GiB 56.80 0.89 85 osd.8
10 hdd 0.87270 1.00000 894GiB 374GiB 373GiB 51.9MiB 1.73GiB
519GiB 41.88 0.66 63 osd.10
15 hdd 0.87270 1.00000 894GiB 504GiB 502GiB 60.1MiB 1.99GiB
390GiB 56.37 0.89 84 osd.15
16 hdd 1.81879 1.00000 1.82TiB 797GiB 794GiB 129MiB 2.79GiB
1.04TiB 42.77 0.67 134 osd.16
TOTAL 14.9TiB 9.45TiB 7.46TiB 1.47GiB 23.2GiB
5.43TiB 63.52
MIN/MAX VAR: 0.66/1.31 STDDEV: 11.59
root@node02:~ # df -h
Filesystem Size Used Avail Use% Mounted on
udev 63G 0 63G 0% /dev
tmpfs 13G 1.4G 12G 11% /run
/dev/sda3 46G 31G 14G 70% /
tmpfs 63G 63M 63G 1% /dev/shm
tmpfs 5.0M 0 5.0M 0% /run/lock
tmpfs 63G 0 63G 0% /sys/fs/cgroup
/dev/sda1 922M 206M 653M 24% /boot
/dev/fuse 30M 140K 30M 1% /etc/pve
/dev/sde1 93M 5.4M 88M 6% /var/lib/ceph/osd/ceph-11
/dev/sdf1 93M 5.4M 88M 6% /var/lib/ceph/osd/ceph-14
/dev/sdc1 889G 738G 152G 83% /var/lib/ceph/osd/ceph-3
/dev/sdb1 889G 654G 235G 74% /var/lib/ceph/osd/ceph-2
/dev/sdd1 93M 5.4M 88M 6% /var/lib/ceph/osd/ceph-7
tmpfs 13G 0 13G 0% /run/user/0
Any idea?
Thanks,
Michael
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx