Hi to All, I've the following warning message (WARN) in my cluster: ceph@ceph-node04:~$ sudo ceph status cluster 50ae3778-dfe3-4492-9628-54a8918ede92 health HEALTH_WARN too few pgs per osd (3 < min 20) monmap e1: 1 mons at {ceph-node01=10.1.1.151:6789/0}, election epoch 2, quorum 0 ceph-node01 osdmap e259: 55 osds: 55 up, 55 in pgmap v703: 192 pgs, 3 pools, 0 bytes data, 0 objects 2579 MB used, 7328 GB / 7331 GB avail 192 active+clean ceph@ceph-node04:~$ It's a new cluster setup, the OSD tree is the following: ceph@ceph-node04:~$ sudo ceph osd tree # id weight type name up/down reweight -1 7.27 root default -2 1.15 host ceph-node01 12 0.06999 osd.12 up 1 13 0.06999 osd.13 up 1 14 0.06999 osd.14 up 1 15 0.06999 osd.15 up 1 16 0.06999 osd.16 up 1 17 0.06999 osd.17 up 1 18 0.06999 osd.18 up 1 19 0.06999 osd.19 up 1 20 0.06999 osd.20 up 1 21 0.45 osd.21 up 1 22 0.06999 osd.22 up 1 -3 1.53 host ceph-node02 23 0.06999 osd.23 up 1 24 0.06999 osd.24 up 1 25 0.06999 osd.25 up 1 26 0.06999 osd.26 up 1 27 0.06999 osd.27 up 1 28 0.06999 osd.28 up 1 29 0.06999 osd.29 up 1 30 0.06999 osd.30 up 1 31 0.06999 osd.31 up 1 32 0.45 osd.32 up 1 33 0.45 osd.33 up 1 -4 1.53 host ceph-node03 34 0.06999 osd.34 up 1 35 0.06999 osd.35 up 1 36 0.06999 osd.36 up 1 37 0.06999 osd.37 up 1 38 0.06999 osd.38 up 1 39 0.06999 osd.39 up 1 40 0.06999 osd.40 up 1 41 0.06999 osd.41 up 1 42 0.06999 osd.42 up 1 43 0.45 osd.43 up 1 44 0.45 osd.44 up 1 -5 1.53 host ceph-node04 0 0.06999 osd.0 up 1 1 0.06999 osd.1 up 1 2 0.06999 osd.2 up 1 3 0.06999 osd.3 up 1 4 0.06999 osd.4 up 1 5 0.06999 osd.5 up 1 6 0.06999 osd.6 up 1 7 0.06999 osd.7 up 1 8 0.06999 osd.8 up 1 9 0.45 osd.9 up 1 10 0.45 osd.10 up 1 -6 1.53 host ceph-node05 11 0.06999 osd.11 up 1 45 0.06999 osd.45 up 1 46 0.06999 osd.46 up 1 47 0.06999 osd.47 up 1 48 0.06999 osd.48 up 1 49 0.06999 osd.49 up 1 50 0.06999 osd.50 up 1 51 0.06999 osd.51 up 1 52 0.06999 osd.52 up 1 53 0.45 osd.53 up 1 54 0.45 osd.54 up 1 ceph@ceph-node04:~$ Someone could give me a hand to resolved this situation. German Anders |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com