Ah, The fix for this is 92a49fb0f79f3300e6e50ddf56238e70678e4202, which first appeared in the 3.9 kernel. The mainline 3.8 stable kernel is EOL, but Canonical is still maintaining one for ubuntu. I can send a note to them. sage On Thu, 13 Jun 2013, Da Chun wrote: > Sage, > > I have the same issue with ceph 0.61.3 on Ubuntu 13.04. > > ceph@ceph-node4:~/mycluster$ df -h > Filesystem Size Used Avail Use% Mounted on > /dev/mapper/ubuntu1304--64--vg-root 15G 1.5G 13G 11% / > none 4.0K 0 4.0K 0% /sys/fs/cgroup > udev 487M 4.0K 487M 1% /dev > tmpfs 100M 284K 100M 1% /run > none 5.0M 0 5.0M 0% /run/lock > none 498M 0 498M 0% /run/shm > none 100M 0 100M 0% /run/user > /dev/sda1 228M 34M 183M 16% /boot > /dev/sdc1 14G 4.4G 9.7G 32% > /var/lib/ceph/osd/ceph-3 > /dev/sdb1 9.0G 1.6G 7.5G 18% > /var/lib/ceph/osd/ceph-0 > 172.18.46.34:6789:/ 276M 94M 183M 34% /mnt/mycephfs > ##### which should be about 70G. > ceph@ceph-node4:~/mycluster$ uname -a > Linux ceph-node4 3.8.0-19-generic #30-Ubuntu SMP Wed May 1 16:35:23 UTC 2013 > x86_64 x86_64 x86_64 GNU/Linux > > > ------------------ Original ------------------ > From: "Sage Weil"<sage@xxxxxxxxxxx>; > Date: Wed, Jun 12, 2013 11:45 PM > To: "Markus Goldberg"<goldberg@xxxxxxxxxxxxxxxxx>; > Cc: "ceph-users"<ceph-users@xxxxxxxxxxxxxx>; > Subject: Re: ceph mount: Only 240 GB , should be 60TB > > Hi Markus, > > What version of the kernel are you using on the client? There is an > annoying compatibility issue with older glibc that makes representing > large values for statfs(2) (df) difficult. We switched this behavior to > hopefully do things the better/"more right" way for the future, but it's > possible you have an odd version or combination that gives goofy results. > > sage > > > On Wed, 12 Jun 2013, Markus Goldberg wrote: > > > Hi, > > this is cuttlefish 0.63 on Ubuntu 13.04, underlying OSD-FS is btrfs, 3 > > servers, each of them 20TB (Raid6-array) > > > > When i mount at the client (or at one of the servers) the mounted > filesystem > > is only 240GB but it should be 60TB. > > > > root@bd-0:~# cat /etc/ceph/ceph.conf > > [global] > > fsid = e0dbf70d-af59-42a5-b834-7ad739a7f89b > > mon_initial_members = bd-0, bd-1, bd-2 > > mon_host = ###.###.###.20,###.###.###.21,###.###.###.22 > > auth_supported = cephx > > public_network = ###.###.###.0/24 > > cluster_network = 192.168.1.0/24 > > osd_mkfs_type = btrfs > > osd_mkfs_options_btrfs = -n 32k -l 32k > > osd_mount_options_btrfs = rw,noatime,nodiratime,autodefrag > > osd_journal_size = 10240 > > > > root@bd-0:~# > > > > df on one of the servers: > > root@bd-0:~# df -h > > Filesystem Size Used Avail Use% Mounted on > > /dev/sda1 39G 4,5G 32G 13% / > > none 4,0K 0 4,0K 0% /sys/fs/cgroup > > udev 16G 12K 16G 1% /dev > > tmpfs 3,2G 852K 3,2G 1% /run > > none 5,0M 4,0K 5,0M 1% /run/lock > > none 16G 0 16G 0% /run/shm > > none 100M 0 100M 0% /run/user > > /dev/sdc1 20T 6,6M 20T 1% /var/lib/ceph/osd/ceph-0 > > root@bd-0:~# > > root@bd-0:~# ceph -s > > health HEALTH_OK > > monmap e1: 3 mons at > >{bd-0=###.###.###.20:6789/0,bd-1=###.###.###.21:6789/0,bd-2=###.###.###.22: > 6789/0}, > > election epoch 66, quorum 0,1,2 bd-0,bd-1,bd-2 > > osdmap e109: 3 osds: 3 up, 3 in > > pgmap v848: 192 pgs: 192 active+clean; 23239 bytes data, 16020 KB > used, > > 61402 GB / 61408 GB avail > > mdsmap e56: 1/1/1 up {0=bd-1=up:active}, 2 up:standby > > > > root@bd-0:~# > > > > > > at the client: > > root@bs4:~# > > root@bs4:~# mount -t ceph ###.###.###.20:6789:/ /mnt/myceph -v -o > > name=admin,secretfile=/etc/ceph/admin.secret > > parsing options: rw,name=admin,secretfile=/etc/ceph/admin.secret > > root@bs4:~# df -h > > Dateisystem Gr???e Benutzt Verf. Verw% Eingeh??ngt auf > > /dev/sda1 28G 3,0G 24G 12% / > > none 4,0K 0 4,0K 0% /sys/fs/cgroup > > udev 998M 4,0K 998M 1% /dev > > tmpfs 201M 708K 200M 1% /run > > none 5,0M 0 5,0M 0% /run/lock > > none 1002M 84K 1002M 1% /run/shm > > none 100M 0 100M 0% /run/user > > ###.###.###.20:6789:/ 240G 25M 240G 1% /mnt/myceph > > root@bs4:~# > > root@bs4:~# cd /mnt/myceph > > root@bs4:/mnt/myceph# mkdir Test > > root@bs4:/mnt/myceph# cd Test > > root@bs4:/mnt/myceph/Test# touch testfile > > root@bs4:/mnt/myceph/Test# ls -la > > insgesamt 0 > > drwxr-xr-x 1 root root 0 Jun 12 2013 . > > drwxr-xr-x 1 root root 0 Jun 12 10:17 .. > > -rw-r--r-- 1 root root 0 Jun 12 10:18 testfile > > root@bs4:/mnt/myceph/Test# pwd > > /mnt/myceph/Test > > root@bs4:/mnt/myceph/Test# df -h . > > Dateisystem Gr???e Benutzt Verf. Verw% Eingeh??ngt auf > > ###.###.###.20:6789:/ 240G 25M 240G 1% /mnt/myceph > > > > > > BTW /dev/sda on the servers are 256GB-SSDs > > > > > > Can anyone please help ? > > > > Thank you, Markus > > > > -- > > MfG, > > Markus Goldberg > > > > ------------------------------------------------------------------------ > > Markus Goldberg | Universit?t Hildesheim > > | Rechenzentrum > > Tel +49 5121 883212 | Marienburger Platz 22, D-31141 Hildesheim, Germany > > Fax +49 5121 883205 | email goldberg@xxxxxxxxxxxxxxxxx > > ------------------------------------------------------------------------ > > > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@xxxxxxxxxxxxxx > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com