Re: Brick and Subvolume Info

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




comments inline

From: "Gino Lisignoli" <glisignoli@xxxxxxxxx>
To: gluster-users@xxxxxxxxxxx
Sent: Wednesday, November 22, 2017 3:49:02 AM
Subject: Brick and Subvolume Info

Hello

I have a Distributed-Replicate volume and I would like to know if it is possible to see what sub-volume a brick belongs to, eg:

A Distributed-Replicate volume containing:

Number of Bricks: 2 x 2 = 4
Brick1: node1.localdomain:/mnt/data1/brick1
Brick2: node2.localdomain:/mnt/data1/brick1
Brick3: node1.localdomain:/mnt/data2/brick2
Brick4: node2.localdomain:/mnt/data2/brick2

Is it possible to list the bricks in sub-volumes, showing which bricks are mirrors of each other? My assumption is that the order the bricks are listed determines what sub-volumes they are in.
>>>No, we don't have this utility yet. However, your assumption is correct.  In your case, Brick1 and Brick2 belong to subvolume one and Brick3 and Brick4 to second sub volume.


Is it also possible to get the status of bricks in their sub-volume to determine their health? Eg, are the bricks in a state where I can offline one of the nodes, replace faulty drives and then online it again?
>>> You can use "gluster v status <volname>" and can get the status  of each brick, if they are UP or not.
When you talk about "health", you should also consider "gluster v heal <volname >" info . to see if there is anything pending to heal or not.
To your next question, I would rather talk about bricks. You can kill/offline a brick and replace it without any issue. If you are not doing any IO then it should be easy job.
If you are offline a NODE and 2 bricks of any one sub volume are placed on that node then that subvolume will  not be active/used. At this point if you are creating/writing a file and that hashes to that sub volume,
you might see IO error on mount point. If you NOT doing any IO that should be fine.

If you want to remove 2 bricks of a same subvolume, I would suggest to go via remove bricks path.
- First remove both the bricks of a subvolume using "gluster volume remove-bricks" command.
- Add 2 new bricks and run rebalance.

----
Ashish


_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux