ceph mount: Only 240 GB , should be 60TB

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,
this is cuttlefish 0.63 on Ubuntu 13.04, underlying OSD-FS is btrfs, 3 servers, each of them 20TB (Raid6-array)

When i mount at the client (or at one of the servers) the mounted filesystem is only 240GB but it should be 60TB.

root@bd-0:~# cat /etc/ceph/ceph.conf
[global]
fsid = e0dbf70d-af59-42a5-b834-7ad739a7f89b
mon_initial_members = bd-0, bd-1, bd-2
mon_host = ###.###.###.20,###.###.###.21,###.###.###.22
auth_supported = cephx
public_network = ###.###.###.0/24
cluster_network = 192.168.1.0/24
osd_mkfs_type = btrfs
osd_mkfs_options_btrfs = -n 32k -l 32k
osd_mount_options_btrfs = rw,noatime,nodiratime,autodefrag
osd_journal_size = 10240

root@bd-0:~#

df on one of the servers:
root@bd-0:~# df -h
Filesystem      Size  Used Avail Use% Mounted on
/dev/sda1        39G  4,5G   32G  13% /
none            4,0K     0  4,0K   0% /sys/fs/cgroup
udev             16G   12K   16G   1% /dev
tmpfs           3,2G  852K  3,2G   1% /run
none            5,0M  4,0K  5,0M   1% /run/lock
none             16G     0   16G   0% /run/shm
none            100M     0  100M   0% /run/user
/dev/sdc1        20T  6,6M   20T   1% /var/lib/ceph/osd/ceph-0
root@bd-0:~#
root@bd-0:~# ceph -s
   health HEALTH_OK
monmap e1: 3 mons at {bd-0=###.###.###.20:6789/0,bd-1=###.###.###.21:6789/0,bd-2=###.###.###.22:6789/0}, election epoch 66, quorum 0,1,2 bd-0,bd-1,bd-2
   osdmap e109: 3 osds: 3 up, 3 in
pgmap v848: 192 pgs: 192 active+clean; 23239 bytes data, 16020 KB used, 61402 GB / 61408 GB avail
   mdsmap e56: 1/1/1 up {0=bd-1=up:active}, 2 up:standby

root@bd-0:~#


at the client:
root@bs4:~#
root@bs4:~# mount -t ceph ###.###.###.20:6789:/ /mnt/myceph -v -o name=admin,secretfile=/etc/ceph/admin.secret
parsing options: rw,name=admin,secretfile=/etc/ceph/admin.secret
root@bs4:~# df -h
Dateisystem           GröÃe Benutzt Verf. Verw% Eingehängt auf
/dev/sda1               28G    3,0G   24G   12% /
none                   4,0K       0  4,0K    0% /sys/fs/cgroup
udev                   998M    4,0K  998M    1% /dev
tmpfs                  201M    708K  200M    1% /run
none                   5,0M       0  5,0M    0% /run/lock
none                  1002M     84K 1002M    1% /run/shm
none                   100M       0  100M    0% /run/user
###.###.###.20:6789:/  240G     25M  240G    1% /mnt/myceph
root@bs4:~#
root@bs4:~# cd /mnt/myceph
root@bs4:/mnt/myceph# mkdir Test
root@bs4:/mnt/myceph# cd Test
root@bs4:/mnt/myceph/Test# touch testfile
root@bs4:/mnt/myceph/Test# ls -la
insgesamt 0
drwxr-xr-x 1 root root 0 Jun 12  2013 .
drwxr-xr-x 1 root root 0 Jun 12 10:17 ..
-rw-r--r-- 1 root root 0 Jun 12 10:18 testfile
root@bs4:/mnt/myceph/Test# pwd
/mnt/myceph/Test
root@bs4:/mnt/myceph/Test# df -h .
Dateisystem           GröÃe Benutzt Verf. Verw% Eingehängt auf
###.###.###.20:6789:/  240G     25M  240G    1% /mnt/myceph


BTW /dev/sda on the servers are 256GB-SSDs


Can anyone please help ?

Thank you,  Markus

--
MfG,
  Markus Goldberg

------------------------------------------------------------------------
Markus Goldberg     | Universität Hildesheim
                    | Rechenzentrum
Tel +49 5121 883212 | Marienburger Platz 22, D-31141 Hildesheim, Germany
Fax +49 5121 883205 | email goldberg@xxxxxxxxxxxxxxxxx
------------------------------------------------------------------------


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux