> Please, paste your `ceph osd df tree` and `ceph osd dump | head -n 12`. $ sudo ceph osd df tree ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS TYPE NAME -8 639.98883 - 639T 327T 312T 51.24 1.00 - root default -10 111.73999 - 111T 58509G 55915G 51.13 1.00 - host bison 78 hdd_fast 0.90900 1.00000 930G 1123M 929G 0.12 0.00 0 osd.78 79 hdd_fast 0.81799 1.00000 837G 1123M 836G 0.13 0.00 0 osd.79 20 hdd 9.09499 0.95000 9313G 4980G 4333G 53.47 1.04 204 osd.20 28 hdd 9.09499 1.00000 9313G 4612G 4700G 49.53 0.97 200 osd.28 29 hdd 9.09499 1.00000 9313G 4848G 4465G 52.05 1.02 211 osd.29 33 hdd 9.09499 1.00000 9313G 4759G 4553G 51.10 1.00 207 osd.33 34 hdd 9.09499 1.00000 9313G 4613G 4699G 49.54 0.97 195 osd.34 35 hdd 9.09499 0.89250 9313G 4954G 4359G 53.19 1.04 206 osd.35 36 hdd 9.09499 1.00000 9313G 4724G 4588G 50.73 0.99 200 osd.36 37 hdd 9.09499 1.00000 9313G 5013G 4300G 53.83 1.05 214 osd.37 38 hdd 9.09499 0.92110 9313G 4962G 4350G 53.28 1.04 206 osd.38 39 hdd 9.09499 1.00000 9313G 4960G 4353G 53.26 1.04 214 osd.39 40 hdd 9.09499 1.00000 9313G 5022G 4291G 53.92 1.05 216 osd.40 41 hdd 9.09499 0.88235 9313G 5037G 4276G 54.09 1.06 203 osd.41 7 ssd 0.87299 1.00000 893G 18906M 875G 2.07 0.04 124 osd.7 -7 102.74084 - 102T 54402G 50805G 51.71 1.01 - host bonnie 0 hdd 7.27699 0.87642 7451G 4191G 3259G 56.25 1.10 175 osd.0 1 hdd 7.27699 0.86200 7451G 3837G 3614G 51.49 1.01 163 osd.1 2 hdd 7.27699 0.74664 7451G 3920G 3531G 52.61 1.03 169 osd.2 11 hdd 7.27699 0.77840 7451G 3983G 3467G 53.46 1.04 169 osd.11 13 hdd 9.09499 0.76595 9313G 4894G 4419G 52.55 1.03 201 osd.13 14 hdd 9.09499 1.00000 9313G 4350G 4963G 46.71 0.91 189 osd.14 16 hdd 9.09499 0.92635 9313G 4879G 4434G 52.39 1.02 204 osd.16 18 hdd 9.09499 0.67932 9313G 4634G 4678G 49.76 0.97 190 osd.18 22 hdd 9.09499 0.93053 9313G 5085G 4228G 54.60 1.07 218 osd.22 31 hdd 9.09499 0.88536 9313G 5152G 4160G 55.33 1.08 221 osd.31 42 hdd 9.09499 0.84232 9313G 4796G 4516G 51.51 1.01 199 osd.42 43 hdd 9.09499 0.87662 9313G 4656G 4657G 50.00 0.98 191 osd.43 6 ssd 0.87299 1.00000 894G 20643M 874G 2.25 0.04 134 osd.6 -6 102.74100 - 102T 53627G 51580G 50.97 0.99 - host capone 3 hdd 7.27699 0.84938 7451G 4028G 3422G 54.07 1.06 171 osd.3 4 hdd 7.27699 0.83890 7451G 3909G 3542G 52.46 1.02 167 osd.4 5 hdd 7.27699 1.00000 7451G 3389G 4061G 45.49 0.89 151 osd.5 9 hdd 7.27699 1.00000 7451G 3710G 3740G 49.80 0.97 161 osd.9 15 hdd 9.09499 1.00000 9313G 4952G 4360G 53.18 1.04 206 osd.15 17 hdd 9.09499 0.95000 9313G 4865G 4448G 52.24 1.02 202 osd.17 23 hdd 9.09499 1.00000 9313G 4984G 4329G 53.52 1.04 223 osd.23 24 hdd 9.09499 1.00000 9313G 4847G 4466G 52.05 1.02 202 osd.24 25 hdd 9.09499 0.89929 9313G 4909G 4404G 52.71 1.03 205 osd.25 30 hdd 9.09499 0.92787 9313G 4740G 4573G 50.90 0.99 202 osd.30 74 hdd 9.09499 0.93146 9313G 4709G 4603G 50.57 0.99 199 osd.74 75 hdd 9.09499 1.00000 9313G 4559G 4753G 48.96 0.96 194 osd.75 8 ssd 0.87299 1.00000 893G 19593M 874G 2.14 0.04 129 osd.8 -16 102.74100 - 102T 53985G 51222G 51.31 1.00 - host elefant 19 hdd 7.27699 1.00000 7451G 3665G 3786G 49.19 0.96 152 osd.19 21 hdd 7.27699 0.89539 7451G 4102G 3349G 55.05 1.07 169 osd.21 64 hdd 7.27699 0.89275 7451G 3956G 3494G 53.10 1.04 171 osd.64 65 hdd 7.27699 0.92513 7451G 3976G 3475G 53.36 1.04 171 osd.65 66 hdd 9.09499 1.00000 9313G 4674G 4638G 50.20 0.98 199 osd.66 67 hdd 9.09499 1.00000 9313G 4737G 4575G 50.87 0.99 201 osd.67 68 hdd 9.09499 0.89973 9313G 4946G 4366G 53.11 1.04 211 osd.68 69 hdd 9.09499 1.00000 9313G 4648G 4665G 49.91 0.97 204 osd.69 70 hdd 9.09499 0.89526 9313G 4907G 4405G 52.69 1.03 209 osd.70 71 hdd 9.09499 0.84923 9313G 4690G 4622G 50.37 0.98 198 osd.71 72 hdd 9.09499 0.87547 9313G 4976G 4336G 53.43 1.04 211 osd.72 73 hdd 9.09499 1.00000 9313G 4683G 4630G 50.29 0.98 200 osd.73 10 ssd 0.87299 1.00000 893G 19158M 875G 2.09 0.04 126 osd.10 -14 110.01300 - 110T 58498G 54157G 51.93 1.01 - host flodhest 27 hdd 9.09499 1.00000 9313G 4602G 4710G 49.42 0.96 199 osd.27 32 hdd 9.09499 0.92557 9313G 5028G 4285G 53.99 1.05 215 osd.32 54 hdd 9.09499 0.90724 9313G 4897G 4415G 52.59 1.03 203 osd.54 55 hdd 9.09499 1.00000 9313G 4867G 4446G 52.26 1.02 198 osd.55 56 hdd 9.09499 1.00000 9313G 4827G 4485G 51.84 1.01 202 osd.56 57 hdd 9.09499 0.93675 9313G 4783G 4530G 51.36 1.00 204 osd.57 58 hdd 9.09499 0.93192 9313G 4599G 4713G 49.39 0.96 194 osd.58 59 hdd 9.09499 1.00000 9313G 5029G 4284G 54.00 1.05 221 osd.59 60 hdd 9.09499 1.00000 9313G 5010G 4303G 53.79 1.05 210 osd.60 61 hdd 9.09499 1.00000 9313G 4773G 4539G 51.26 1.00 205 osd.61 62 hdd 9.09499 1.00000 9313G 5004G 4308G 53.74 1.05 209 osd.62 63 hdd 9.09499 0.93686 9313G 5055G 4258G 54.28 1.06 207 osd.63 77 ssd 0.87299 1.00000 893G 18472M 875G 2.02 0.04 127 osd.77 -12 110.01300 - 110T 56762G 55893G 50.39 0.98 - host yak 12 hdd 9.09499 1.00000 9313G 4916G 4396G 52.79 1.03 211 osd.12 26 hdd 9.09499 1.00000 9313G 4792G 4520G 51.46 1.00 209 osd.26 44 hdd 9.09499 0.91795 9313G 4921G 4391G 52.85 1.03 201 osd.44 45 hdd 9.09499 1.00000 9313G 4764G 4549G 51.15 1.00 206 osd.45 46 hdd 9.09499 0.93283 9313G 4811G 4502G 51.66 1.01 198 osd.46 47 hdd 9.09499 0.91283 9313G 4790G 4523G 51.44 1.00 205 osd.47 48 hdd 9.09499 0.90672 9313G 4902G 4410G 52.64 1.03 205 osd.48 49 hdd 9.09499 1.00000 9313G 4343G 4969G 46.64 0.91 192 osd.49 50 hdd 9.09499 1.00000 9313G 4387G 4925G 47.11 0.92 176 osd.50 51 hdd 9.09499 1.00000 9313G 4522G 4790G 48.56 0.95 195 osd.51 52 hdd 9.09499 1.00000 9313G 4587G 4725G 49.26 0.96 212 osd.52 53 hdd 9.09499 1.00000 9313G 4997G 4315G 53.66 1.05 211 osd.53 76 ssd 0.87299 1.00000 893G 22043M 872G 2.41 0.05 128 osd.76 TOTAL 639T 327T 312T 51.24 MIN/MAX VAR: 0.00/1.10 STDDEV: 16.25 $ sudo ceph osd dump | head -n 12 epoch 49887 fsid dbc33946-ba1f-477c-84df-c63a3c9c91a6 created 2018-05-09 17:14:17.800686 modified 2018-12-25 18:14:44.508127 flags sortbitwise,recovery_deletes,purged_snapdirs crush_version 446 full_ratio 0.95 backfillfull_ratio 0.9 nearfull_ratio 0.9 require_min_compat_client hammer min_compat_client hammer require_osd_release luminous _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com