Hi Joe and Mehmet!
Thanks for your responses!
The requested outputs at the end of the message.
But to make my question more clear:
What we are actually after, is not about CURRENT usage of our OSDs, but
stats on total GBs written in the cluster, per OSD, and read/write ratio.
With those numbers, we would be able to identify suitable replacement
SSDs for our current HDDs, and select specifically for OUR typical use.
(taking into account endurance, speed, price, etc, etc)
And it seems smartctl on our seagate ST4000NM0034 drives do not give us
data on total bytes written or read. (...or are we simply not looking in
the right place..?)
Requested outputs below:
root@node1:~# ceph osd df tree
ID CLASS WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS TYPE NAME
-1 87.35376 - 87.3TiB 49.1TiB 38.2TiB 56.22 1.00 - root default
-2 29.11688 - 29.1TiB 16.4TiB 12.7TiB 56.23 1.00 - host node1
0 hdd 3.64000 1.00000 3.64TiB 2.01TiB 1.62TiB 55.34 0.98 137 osd.0
1 hdd 3.64000 1.00000 3.64TiB 2.09TiB 1.54TiB 57.56 1.02 141 osd.1
2 hdd 3.63689 1.00000 3.64TiB 1.92TiB 1.72TiB 52.79 0.94 128 osd.2
3 hdd 3.64000 1.00000 3.64TiB 2.07TiB 1.57TiB 56.90 1.01 143 osd.3
12 hdd 3.64000 1.00000 3.64TiB 2.15TiB 1.48TiB 59.18 1.05 138 osd.12
13 hdd 3.64000 1.00000 3.64TiB 1.99TiB 1.64TiB 54.80 0.97 131 osd.13
14 hdd 3.64000 1.00000 3.64TiB 1.93TiB 1.70TiB 53.13 0.94 127 osd.14
15 hdd 3.64000 1.00000 3.64TiB 2.19TiB 1.45TiB 60.10 1.07 143 osd.15
-3 29.12000 - 29.1TiB 16.4TiB 12.7TiB 56.22 1.00 - host node2
4 hdd 3.64000 1.00000 3.64TiB 2.11TiB 1.53TiB 57.97 1.03 142 osd.4
5 hdd 3.64000 1.00000 3.64TiB 1.97TiB 1.67TiB 54.11 0.96 134 osd.5
6 hdd 3.64000 1.00000 3.64TiB 2.12TiB 1.51TiB 58.40 1.04 142 osd.6
7 hdd 3.64000 1.00000 3.64TiB 1.97TiB 1.66TiB 54.28 0.97 128 osd.7
16 hdd 3.64000 1.00000 3.64TiB 2.00TiB 1.64TiB 54.90 0.98 133 osd.16
17 hdd 3.64000 1.00000 3.64TiB 2.33TiB 1.30TiB 64.14 1.14 153 osd.17
18 hdd 3.64000 1.00000 3.64TiB 1.97TiB 1.67TiB 54.07 0.96 132 osd.18
19 hdd 3.64000 1.00000 3.64TiB 1.89TiB 1.75TiB 51.93 0.92 124 osd.19
-4 29.11688 - 29.1TiB 16.4TiB 12.7TiB 56.22 1.00 - host node3
8 hdd 3.64000 1.00000 3.64TiB 1.79TiB 1.85TiB 49.24 0.88 123 osd.8
9 hdd 3.64000 1.00000 3.64TiB 2.17TiB 1.47TiB 59.72 1.06 144 osd.9
10 hdd 3.64000 1.00000 3.64TiB 2.40TiB 1.24TiB 65.88 1.17 157 osd.10
11 hdd 3.64000 1.00000 3.64TiB 2.06TiB 1.58TiB 56.64 1.01 133 osd.11
20 hdd 3.64000 1.00000 3.64TiB 2.19TiB 1.45TiB 60.23 1.07 148 osd.20
21 hdd 3.64000 1.00000 3.64TiB 1.74TiB 1.90TiB 47.80 0.85 115 osd.21
22 hdd 3.64000 1.00000 3.64TiB 2.05TiB 1.59TiB 56.27 1.00 138 osd.22
23 hdd 3.63689 1.00000 3.64TiB 1.96TiB 1.67TiB 54.01 0.96 130 osd.23
TOTAL 87.3TiB 49.1TiB 38.2TiB 56.22
MIN/MAX VAR: 0.85/1.17 STDDEV: 4.08
root@node1:~# ceph osd status
+----+------+-------+-------+--------+---------+--------+---------+-----------+
| id | host | used | avail | wr ops | wr data | rd ops | rd data | state |
+----+------+-------+-------+--------+---------+--------+---------+-----------+
| 0 | node1 | 2061G | 1663G | 38 | 5168k | 3 | 1491k | exists,up |
| 1 | node1 | 2143G | 1580G | 4 | 1092k | 9 | 2243k | exists,up |
| 2 | node1 | 1965G | 1758G | 20 | 3643k | 5 | 1758k | exists,up |
| 3 | node1 | 2119G | 1605G | 17 | 99.5k | 4 | 3904k | exists,up |
| 4 | node2 | 2158G | 1565G | 12 | 527k | 1 | 2632k | exists,up |
| 5 | node2 | 2014G | 1709G | 15 | 239k | 0 | 889k | exists,up |
| 6 | node2 | 2174G | 1549G | 11 | 1677k | 5 | 1931k | exists,up |
| 7 | node2 | 2021G | 1702G | 2 | 597k | 0 | 1638k | exists,up |
| 8 | node3 | 1833G | 1890G | 4 | 564k | 4 | 5595k | exists,up |
| 9 | node3 | 2223G | 1500G | 6 | 1124k | 10 | 4864k | exists,up |
| 10 | node3 | 2453G | 1270G | 8 | 1257k | 3 | 1447k | exists,up |
| 11 | node3 | 2109G | 1614G | 14 | 2889k | 3 | 1449k | exists,up |
| 12 | node1 | 2204G | 1520G | 17 | 1596k | 4 | 1806k | exists,up |
| 13 | node1 | 2040G | 1683G | 15 | 2526k | 0 | 819k | exists,up |
| 14 | node1 | 1978G | 1745G | 11 | 1713k | 8 | 3489k | exists,up |
| 15 | node1 | 2238G | 1485G | 25 | 5151k | 5 | 2715k | exists,up |
| 16 | node2 | 2044G | 1679G | 2 | 43.3k | 1 | 3371k | exists,up |
| 17 | node2 | 2388G | 1335G | 14 | 1736k | 9 | 5315k | exists,up |
| 18 | node2 | 2013G | 1710G | 8 | 1907k | 2 | 2004k | exists,up |
| 19 | node2 | 1934G | 1790G | 15 | 2115k | 4 | 3248k | exists,up |
| 20 | node3 | 2243G | 1481G | 15 | 3292k | 1 | 1763k | exists,up |
| 21 | node3 | 1780G | 1944G | 8 | 1636k | 0 | 86.4k | exists,up |
| 22 | node3 | 2095G | 1628G | 23 | 5012k | 4 | 1654k | exists,up |
| 23 | node3 | 2011G | 1712G | 9 | 1662k | 1 | 2457k | exists,up |
+----+------+-------+-------+--------+---------+--------+---------+-----------+
Thanks!
MJ
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx