Re: Arbiter brick size estimation

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 03/05/2016 03:45 PM, Oleksandr Natalenko wrote:
In order to estimate GlusterFS arbiter brick size, I've deployed test setup 
with replica 3 arbiter 1 volume within one node. Each brick is located on 
separate HDD (XFS with inode size == 512). Using GlusterFS v3.7.6 + memleak 
patches. Volume options are kept default.

Here is the script that creates files and folders in mounted volume: [1]

The script creates 1M of files of random size (between 1 and 32768 bytes) and 
some amount of folders. After running it I've got 1036637 folders. So, in 
total it is 2036637 files and folders.

The initial used space on each brick is 42M . After running script I've got:

replica brick 1 and 2: 19867168 kbytes == 19G
arbiter brick: 1872308 kbytes == 1.8G

The amount of inodes on each brick is 3139091. So here goes estimation.

Dividing arbiter used space by files+folders we get:

(1872308 - 42000)/2036637 == 899 bytes per file or folder

Dividing arbiter used space by inodes we get:

(1872308 - 42000)/3139091 == 583 bytes per inode

Not sure about what calculation is correct.

I think the first one is right because you still haven't used up all the inodes.(2036637 used vs. the max. permissible 3139091). But again this is an approximation because not all files would be 899 bytes. For example if there are a thousand files present in a directory, then du <dirname> would be more than du <file> because the directory will take some disk space to store the dentries.

 I guess we should consider the one 
that accounts inodes because of .glusterfs/ folder data.

Nevertheless, in contrast, documentation [2] says it should be 4096 bytes per 
file. Am I wrong with my calculations?

The 4KB is a conservative estimate considering the fact that though the arbiter brick does not store data, it still keeps a copy of both user and gluster xattrs. For example, if the application sets a lot of xattrs, it can consume a data block if they cannot be accommodated on the inode itself.  Also there is the .glusterfs folder like you said which would take up some space. Here is what I tried on an XFS brick:
[root@ravi4 brick]# touch file

[root@ravi4 brick]# ls -l file
-rw-r--r-- 1 root root 0 Mar  8 12:54 file

[root@ravi4 brick]# du file

0       file

[root@ravi4 brick]# for i in {1..100}

> do
> setfattr -n user.value$i -v value$i file
> done

[root@ravi4 brick]# ll -l file

-rw-r--r-- 1 root root 0 Mar  8 12:54 file

[root@ravi4 brick]# du -h file

4.0K    file

Hope this helps,
Ravi


Pranith?

[1] http://termbin.com/ka9x
[2] http://gluster.readthedocs.org/en/latest/Administrator%20Guide/arbiter-volumes-and-quorum/
_______________________________________________
Gluster-devel mailing list
Gluster-devel@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-devel


_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux