Re: Wrong volume size with df

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Sure!

> 1 - output of gluster volume heal <volname> info

Brick pod-sjc1-gluster1:/data/brick1/gv0
Status: Connected
Number of entries: 0

Brick pod-sjc1-gluster2:/data/brick1/gv0
Status: Connected
Number of entries: 0

Brick pod-sjc1-gluster1:/data/brick2/gv0
Status: Connected
Number of entries: 0

Brick pod-sjc1-gluster2:/data/brick2/gv0
Status: Connected
Number of entries: 0

Brick pod-sjc1-gluster1:/data/brick3/gv0
Status: Connected
Number of entries: 0

Brick pod-sjc1-gluster2:/data/brick3/gv0
Status: Connected
Number of entries: 0

> 2 - /var/log/glusterfs - provide log file with mountpoint-volumename.log

Attached

> 3 - output of gluster volume <volname> info

[root@pod-sjc1-gluster2 ~]# gluster volume info
 
Volume Name: gv0
Type: Distributed-Replicate
Volume ID: d490a9ec-f9c8-4f10-a7f3-e1b6d3ced196
Status: Started
Snapshot Count: 13
Number of Bricks: 3 x 2 = 6
Transport-type: tcp
Bricks:
Brick1: pod-sjc1-gluster1:/data/brick1/gv0
Brick2: pod-sjc1-gluster2:/data/brick1/gv0
Brick3: pod-sjc1-gluster1:/data/brick2/gv0
Brick4: pod-sjc1-gluster2:/data/brick2/gv0
Brick5: pod-sjc1-gluster1:/data/brick3/gv0
Brick6: pod-sjc1-gluster2:/data/brick3/gv0
Options Reconfigured:
performance.cache-refresh-timeout: 60
performance.stat-prefetch: on
server.allow-insecure: on
performance.flush-behind: on
performance.rda-cache-limit: 32MB
network.tcp-window-size: 1048576
performance.nfs.io-threads: on
performance.write-behind-window-size: 4MB
performance.nfs.write-behind-window-size: 512MB
performance.io-cache: on
performance.quick-read: on
features.cache-invalidation: on
features.cache-invalidation-timeout: 600
performance.cache-invalidation: on
performance.md-cache-timeout: 600
network.inode-lru-limit: 90000
performance.cache-size: 4GB
server.event-threads: 16
client.event-threads: 16
features.barrier: disable
transport.address-family: inet
nfs.disable: on
performance.client-io-threads: on
cluster.lookup-optimize: on
server.outstanding-rpc-limit: 1024
auto-delete: enable
You have new mail in /var/spool/mail/root

> 4 - output of gluster volume <volname> status

[root@pod-sjc1-gluster2 ~]# gluster volume status gv0
Status of volume: gv0
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
Brick pod-sjc1-gluster1:/data/
brick1/gv0                                  49152     0          Y       3198 
Brick pod-sjc1-gluster2:/data/
brick1/gv0                                  49152     0          Y       4018 
Brick pod-sjc1-gluster1:/data/
brick2/gv0                                  49153     0          Y       3205 
Brick pod-sjc1-gluster2:/data/
brick2/gv0                                  49153     0          Y       4029 
Brick pod-sjc1-gluster1:/data/
brick3/gv0                                  49154     0          Y       3213 
Brick pod-sjc1-gluster2:/data/
brick3/gv0                                  49154     0          Y       4036 
Self-heal Daemon on localhost               N/A       N/A        Y       17869
Self-heal Daemon on pod-sjc1-gluster1.exava
ult.com                                     N/A       N/A        Y       3183 
 
Task Status of Volume gv0
------------------------------------------------------------------------------
There are no active volume tasks
 

> 5 - Also, could you try unmount the volume and mount it again and check the size?

I have done this a few times but it doesn't seem to help.

On Thu, Dec 21, 2017 at 11:18 AM, Ashish Pandey <aspandey@xxxxxxxxxx> wrote:

Could youplease provide following -

1 - output of gluster volume heal <volname> info
2 - /var/log/glusterfs - provide log file with mountpoint-volumename.log
3 - output of gluster volume <volname> info
4 - output of gluster volume <volname> status
5 - Also, could you try unmount the volume and mount it again and check the size?








From: "Teknologeek Teknologeek" <teknologeek06@xxxxxxxxx>
To: gluster-users@xxxxxxxxxxx
Sent: Wednesday, December 20, 2017 2:54:40 AM
Subject: Wrong volume size with df


I have a glusterfs setup with distributed disperse volumes 5 * ( 4 + 2 ).

After a server crash, "gluster peer status" reports all peers as connected.

"gluster volume status detail" shows that all bricks are up and running with the right size, but when I use df from a client mount point, the size displayed is about 1/6 of the total size.

When browsing the data, they seem to be ok tho.

I need some help to understand what's going on as i can't delete the volume and recreate it from scratch.

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users


_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users

Attachment: glusterfs-mount.log
Description: Binary data

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux