Hi, So I have my small test cluster with 2 OSD nodes. In each OSD node, there is one disk (5 GB) and one cosd. When I mount the FS on a client, I see it's 10 GB in total. Fine. I copied a small set of multimedia files (135 MB) and that's all. So why do I see the space used is getting higher and higher ? root@client:/mnt/osd$ df -h . Filesystem Size Used Avail Use% Mounted on 10.1.56.232:6789:/ 10G 5.2G 3.9G 57% /mnt/osd and like 2 hours later : root@client:/mnt/osd$ df -h . Filesystem Size Used Avail Use% Mounted on 10.1.56.232:6789:/ 10G 6.8G 2.3G 76% /mnt/osd I'm using btrfs to store the files and I made no creation/modification of files; i just run some bench with "ceph osd tell [12] bench" Does the bench create files and do note delete them ? Extract of ceph -w : root@test2:/data/mon1# ceph -w 2011-06-11 19:16:09.186416 7fbefd01e700 -- :/14931 >> 10.1.56.233:6789/0 pipe(0x1e58540 sd=4 pgs=0 cs=0 l=0).fault first fault 2011-06-11 19:16:12.191439 pg v1990: 594 pgs: 594 active+clean; 2066 MB data, 6906 MB used, 2286 MB / 10240 MB avail 2011-06-11 19:16:12.193479 mds e48: 1/1/1 up {0=test4=up:active}, 1 up:standby 2011-06-11 19:16:12.193551 osd e44: 2 osds: 2 up, 2 in 2011-06-11 19:16:12.193658 log 2011-06-12 00:55:42.389017 osd2 10.1.56.237:6800/1744 307 : [INF] bench: wrote 1024 MB in blocks of 4096 KB in 12.438478 sec at 84300 KB/sec 2011-06-11 19:16:12.193757 mon e1: 3 mons at {0=10.1.56.231:6789/0,1=10.1.56.232:6789/0,2=10.1.56.233:6789/0} Attached is my ceph.conf Thanks ! Wilfrid
Attachment:
ceph.conf
Description: Binary data