Hi All, We are adding a new brick (91TB) to our existing gluster volume (328 TB). The new brick is on a new physical server and we want to make sure that we are doing this correctly (the existing volume had 2 bricks on a single physical server). Both are running glusterfs 3.7.11. The steps we've followed are
The follow-up question is how long should the rebalance using fix-layout take? Some additional information # gluster volume info Volume Name: data-volume Type: Distribute Volume ID: c162161e-2a2d-4dac-b015-f31fd89ceb18 Status: Started Number of Bricks: 3 Transport-type: tcp Bricks: Brick1: mseas-data2:/mnt/brick1 Brick2: mseas-data2:/mnt/brick2 Brick3: mseas-data3:/export/sda/brick3 Options Reconfigured: diagnostics.client-log-level: ERROR network.inode-lru-limit: 50000 performance.md-cache-timeout: 60 performance.open-behind: off disperse.eager-lock: off auth.allow: * server.allow-insecure: on nfs.exports-auth-enable: on diagnostics.brick-sys-log-level: WARNING performance.readdir-ahead: on nfs.disable: on nfs.export-volumes: off Thanks -- -=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=- Pat Haley Email: phaley@xxxxxxx Center for Ocean Engineering Phone: (617) 253-6824 Dept. of Mechanical Engineering Fax: (617) 253-8125 MIT, Room 5-213 http://web.mit.edu/phaley/www/ 77 Massachusetts Avenue Cambridge, MA 02139-4301 |
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users