Re: Distributed volume going to Read only mode if any of the Brick is not available

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Team,
Anyone can suggest my below query, so that I can get clear idea

Regards,
Varad

On 19 May 2015 20:28, "Varadharajan S" <rajanvaradhu@xxxxxxxxx> wrote:

FYI

On 19 May 2015 20:25, "Varadharajan S" <rajanvaradhu@xxxxxxxxx> wrote:

Hi,
Replication means, I won't get space. Distribution is not like striping right? If one brick is not available in the volume, other bricks can distribute data in between. If I do any tuning will get solution?

On 19 May 2015 20:02, "Atin Mukherjee" <atin.mukherjee83@xxxxxxxxx> wrote:


On 19 May 2015 17:10, "Varadharajan S" <rajanvaradhu@xxxxxxxxx> wrote:
>
> Hi,
>
> We are using Ubuntu 14.04 server and for storage purpose we configured gluster 3.5 as distributed volume and find the below details,
>
> 1).4 Servers - 14.04 Ubuntu Server and each server disks free spaces are  configured as ZFS raiddz2 volume
>
> 2). Each server has /pool/gluster zfs volume and capacity as - 5 TB,8 TB,6 TB and 10 TB
>
> 3). Bricks are - rep1,rep2,rep3 and st1 and all the bricks are connected as Distributed Volume and mounted on each system as,
>  
>   For E.x in rep1 -> mount -t glusterfs  rep1:/glustervol  /data.
>                   rep2  -> mount  -t glusterfs  rep2:/glustervol  /data
>                   rep3  -> mount  -t glusterfs  rep3:/glustervol  /data
>                   st1    ->  mount  -t glusterfs  st1:/glustervol  /data
>
> So we get /data is having  around 29 TB and all our applications data's are stored in /data mount point.
>
> Details about volume:
>
> volume glustervol-client-0
>     type protocol/client
>     option send-gids true
>     option password b217da9d1d8b-bb55
>     option username 9d76-4553-8c75
>     option transport-type tcp
>     option remote-subvolume /pool/gluster
>     option remote-host rep1
>     option ping-timeout 42
> end-volume
>
> volume glustervol-client-1
>     type protocol/client
>     option send-gids true
>     option password b217da9d1d8b-bb55
>     option username jkd76-4553-5347
>     option transport-type tcp
>     option remote-subvolume /pool/gluster
>     option remote-host rep2
>     option ping-timeout 42
> end-volume
>
> volume glustervol-client-2
>     type protocol/client
>     option send-gids true
>     option password b217da9d1d8b-bb55
>     option username 19d7-5a190c2
>     option transport-type tcp
>     option remote-subvolume /pool/gluster
>     option remote-host rep3
>     option ping-timeout 42
> end-volume
>
> volume glustervol-client-3
>     type protocol/client
>     option send-gids true
>     option password b217da9d1d8b-bb55
>     option username c75-5436b5a168347
>     option transport-type tcp
>     option remote-subvolume /pool/gluster
>     option remote-host st1
>
>     option ping-timeout 42
> end-volume
>
> volume glustervol-dht
>     type cluster/distribute
>     subvolumes glustervol-client-0 glustervol-client-1 glustervol-client-2 glustervol-client-3
> end-volume
>
> volume glustervol-write-behind
>     type performance/write-behind
>     subvolumes glustervol-dht
> end-volume
>
> volume glustervol-read-ahead
>     type performance/read-ahead
>     subvolumes glustervol-write-behind
> end-volume
>
> volume glustervol-io-cache
>     type performance/io-cache
>     subvolumes glustervol-read-ahead
> end-volume
>
> volume glustervol-quick-read
>     type performance/quick-read
>     subvolumes glustervol-io-cache
> end-volume
>
> volume glustervol-open-behind
>     type performance/open-behind
>     subvolumes glustervol-quick-read
> end-volume
>
> volume glustervol-md-cache
>     type performance/md-cache
>     subvolumes glustervol-open-behind
> end-volume
>
> volume glustervol
>     type debug/io-stats
>     option count-fop-hits off
>     option latency-measurement off
>     subvolumes glustervol-md-cache
> end-volume
>
>
> ap@rep3:~$ sudo gluster volume info
>  
> Volume Name: glustervol
> Type: Distribute
> Volume ID: 165b-XXXXX
> Status: Started
> Number of Bricks: 4
> Transport-type: tcp
> Bricks:
> Brick1: rep1:/pool/gluster
> Brick2: rep2:/pool/gluster
> Brick3: rep3:/pool/gluster
> Brick4: st1:/pool/gluster
>
> Problem:
>
> If we shutdown any of the bricks , the volume size is reduced (this is ok) but from the other servers , i can see my mount point /data but it's only listing contents and i can't write or edit any single files/folders.
>
> Solution Required:
>
> If anyone brick is not available, From other servers should allow for Write and edit functions
This is expected since you are using distributed volume. You wouldn't be able to write/edit files belonging to the brick which is down. Solution would be to migrate to distributed replicate volume.
>
> Please let us know, what can i try further ?
>
> Regards,
> Varad
>
>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users@xxxxxxxxxxx
> http://www.gluster.org/mailman/listinfo/gluster-users

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux