Hi Luis,
What i mean , we have three OSD with Harddisk size each 1TB and two pool (poolA and poolB) with replica 2. Here writing behavior is the confusion for us. Our assumptions is below.
PoolA -- may write with OSD1 and OSD2 (is this correct)
PoolB -- may write with OSD3 and OSD1 (is this correct)
suppose the hard disk size got full , then how many OSD's need to be added and How will be the writing behavior to new OSD's
After added few osd's
PoolA -- may write with OSD4 and OSD5 (is this correct)
PoolB -- may write with OSD5 and OSD6 (is this correct)
Regards
Prabu
---- On Mon, 17 Aug 2015 19:41:53 +0530 Luis Periquito <periquito@xxxxxxxxx> wrote ----
I don't understand your question? You created a 1G RBD/disk and it's full. You are able to grow it though - but that's a Linux management issue, not ceph.As everything is thin-provisioned you can create a RBD with an arbitrary size - I've create one with 1PB when the cluster only had 600G/Raw available.On Mon, Aug 17, 2015 at 1:18 PM, gjprabu <gjprabu@xxxxxxxxxxxx> wrote:Hi All,Anybody can help on this issue.Hi All,Also please find osd information.ceph osd dump | grep 'replicated size'pool 2 'repo' replicated size 2 min_size 2 crush_ruleset 0 object_hash rjenkins pg_num 126 pgp_num 126 last_change 21573 flags hashpspool stripe_width 0RegardsPrabuHi All,We need to test three OSD and one image with replica 2(size 1GB). While testing data is not writing above 1GB. Is there any option to write on third OSD.ceph osd pool get repo pg_numpg_num: 126# rbd showmappedid pool image snap device0 rbd integdownloads - /dev/rbd0 -- Already one2 repo integrepotest - /dev/rbd2 -- newly created[root@hm2 repository]# df -ThFilesystem Type Size Used Avail Use% Mounted on/dev/sda5 ext4 289G 18G 257G 7% /devtmpfs devtmpfs 252G 0 252G 0% /devtmpfs tmpfs 252G 0 252G 0% /dev/shmtmpfs tmpfs 252G 538M 252G 1% /runtmpfs tmpfs 252G 0 252G 0% /sys/fs/cgroup/dev/sda2 ext4 488M 212M 241M 47% /boot/dev/sda4 ext4 1.9T 20G 1.8T 2% /var/dev/mapper/vg0-zoho ext4 8.6T 1.7T 6.5T 21% /zoho/dev/rbd0 ocfs2 977G 101G 877G 11% /zoho/build/downloads/dev/rbd2 ocfs2 1000M 1000M 0 100% /zoho/build/repository@:~$ scp -r sample.txt root@integ-hm2:/zoho/build/repository/root@integ-hm2's password:sample.txt 100% 1024MB 4.5MB/s 03:48scp: /zoho/build/repository//sample.txt: No space left on deviceRegardsPrabuDear Team,We are using two ceph OSD with replica 2 and it is working properly. Here my doubt is (Pool A -image size will be 10GB) and its replicated with two OSD, what will happen suppose if the size reached the limit, Is there any chance to make the data to continue writing in another two OSD's.RegardsPrabu_______________________________________________ceph-users mailing list
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com