Re: cannot access to CIFS export

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




On Mon, Feb 9, 2015 at 10:35 PM, Volnei Puttini <volnei@xxxxxxxxxxxxxxx> wrote:
Hi Alessandro,

My system:

CentOS 7

samba-vfs-glusterfs-4.1.1-37.el7_0.x86_64
This is the rpm that ships glusterfs.so which Samba searched for and did not find.
 
samba-winbind-4.1.1-37.el7_0.x86_64
samba-libs-4.1.1-37.el7_0.x86_64
samba-common-4.1.1-37.el7_0.x86_64
samba-winbind-modules-4.1.1-37.el7_0.x86_64
samba-winbind-clients-4.1.1-37.el7_0.x86_64
samba-4.1.1-37.el7_0.x86_64
samba-client-4.1.1-37.el7_0.x86_64

glusterfs 3.6.2 built on Jan 22 2015 12:59:57

Try this, work fine for me:

[GFSVOL]
    browseable = No
    comment = Gluster share of volume gfsvol
    path = /
    read _only_ = No
    guest ok = Yes
    kernel share modes = No
    posix locking = No
    vfs objects = glusterfs
    glusterfs:loglevel = 7
    glusterfs:logfile = /var/log/samba/glusterfs-gfstest.log
    glusterfs:volume = vgtest
    glusterfs:volfile_server = 192.168.2.21



On 09-02-2015 14:45, RASTELLI Alessandro wrote:

Hi,

I’ve created and started a new replica volume “downloadstat” with CIFS export enabled on GlusterFS 3.6.2.

I can see the following piece has been added automatically to smb.conf:

[gluster-downloadstat]

comment = For samba share of volume downloadstat

vfs objects = glusterfs

glusterfs:volume = downloadstat

glusterfs:logfile = /var/log/samba/glusterfs-downloadstat.%M.log

glusterfs:loglevel = 7

path = /

read _only_ = no

guest ok = yes

 

I restarted smb service, without errors.

When I try to access from Win7 client to “\\gluster01-mi\gluster-downloadstat” it asks me a login (which user do I need to put?) and then gives me error “The network path was not found”

The user you use to login depends on your Samba setup, if it is in AD environment , you would use one of the AD users. If it is not, then you will have to use a user added to Samba on the server
using smbpasswd -a <user>. Also, make sure that the same user exists on the node(s) and has rights on the bricks.

 

and on Gluster smb.log I see:

[2015/02/09 17:21:13.111639,  0] smbd/vfs.c:173(vfs_init_custom)

  error probing vfs module 'glusterfs': NT_STATUS_UNSUCCESSFUL

[2015/02/09 17:21:13.111709,  0] smbd/vfs.c:315(smbd_vfs_init)

  smbd_vfs_init: vfs_init_custom failed for glusterfs

[2015/02/09 17:21:13.111741,  0] smbd/service.c:902(make_connection_snum)

  vfs_init failed for service gluster-downloadstat

 

Can you explain how to fix?

Thanks

 

Alessandro

 

From: gluster-users-bounces@xxxxxxxxxxx [mailto:gluster-users-bounces@xxxxxxxxxxx] On Behalf Of David F. Robinson
Sent: domenica 8 febbraio 2015 18:19
To: Gluster Devel; gluster-users@xxxxxxxxxxx
Subject: cannot delete non-empty directory

 

I am seeing these messsages after I delete large amounts of data using gluster 3.6.2.

cannot delete non-empty directory: old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final

 

From the FUSE mount (as root), the directory shows up as empty:

 

# pwd
/backup/homegfs/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final


# ls -al
total 5
d--------- 2 root root    4106 Feb  6 13:55 .
drwxrws--- 3  601 dmiller   72 Feb  6 13:55 ..

 

However, when you look at the bricks, the files are still there (none on brick01bkp, all files are on brick02bkp).  All of the files are 0-length and have ------T permissions.

Any suggestions on how to fix this and how to prevent it from happening?

 

#  ls -al /data/brick*/homegfs_bkp/backup.0/old_shelf4/Aegis/\!\!\!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final
/data/brick01bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final:
total 4
d---------+ 2 root root  10 Feb  6 13:55 .
drwxrws---+ 3  601 raven 36 Feb  6 13:55 ..

 

/data/brick02bkp/homegfs_bkp/backup.0/old_shelf4/Aegis/!!!Programs/RavenCFD/Storage/Jimmy_Old/src_vj1.5_final:
total 8
d---------+ 3 root root  4096 Dec 31  1969 .
drwxrws---+ 3  601 raven   36 Feb  6 13:55 ..
---------T  5  601 raven    0 Nov 20 00:08 read_inset.f.gz
---------T  5  601 raven    0 Nov 20 00:08 readbc.f.gz
---------T  5  601 raven    0 Nov 20 00:08 readcn.f.gz
---------T  5  601 raven    0 Nov 20 00:08 readinp.f.gz
---------T  5  601 raven    0 Nov 20 00:08 readinp_v1_2.f.gz
---------T  5  601 raven    0 Nov 20 00:08 readinp_v1_3.f.gz
---------T  5  601 raven    0 Nov 20 00:08 rotatept.f.gz
d---------+ 2 root root   118 Feb  6 13:54 save1
---------T  5  601 raven    0 Nov 20 00:08 sepvec.f.gz
---------T  5  601 raven    0 Nov 20 00:08 shadow.f.gz
---------T  5  601 raven    0 Nov 20 00:08 snksrc.f.gz
---------T  5  601 raven    0 Nov 20 00:08 source.f.gz
---------T  5  601 raven    0 Nov 20 00:08 step.f.gz
---------T  5  601 raven    0 Nov 20 00:08 stoprog.f.gz
---------T  5  601 raven    0 Nov 20 00:08 summer6.f.gz
---------T  5  601 raven    0 Nov 20 00:08 totforc.f.gz
---------T  5  601 raven    0 Nov 20 00:08 tritet.f.gz
---------T  5  601 raven    0 Nov 20 00:08 wallrsd.f.gz
---------T  5  601 raven    0 Nov 20 00:08 wheat.f.gz
---------T  5  601 raven    0 Nov 20 00:08 write_inset.f.gz

 

 

This is using gluster 3.6.2 on a distributed gluster volume that resides on a single machine.  Both of the bricks are on one machine consisting of 2x RAID-6 arrays.

 

df -h | grep brick

/dev/mapper/vg01-lvol1                       88T   22T   66T  25% /data/brick01bkp
/dev/mapper/vg02-lvol1                       88T   22T   66T  26% /data/brick02bkp

 

# gluster volume info homegfs_bkp

Volume Name: homegfs_bkp
Type: Distribute
Volume ID: 96de8872-d957-4205-bf5a-076e3f35b294
Status: Started
Number of Bricks: 2
Transport-type: tcp
Bricks:
Brick1: gfsib01bkp.corvidtec.com:/data/brick01bkp/homegfs_bkp
Brick2: gfsib01bkp.corvidtec.com:/data/brick02bkp/homegfs_bkp
Options Reconfigured:
storage.owner-gid: 100
performance.io-thread-count: 32
server.allow-insecure: on
network.ping-timeout: 10
performance.cache-size: 128MB
performance.write-behind-window-size: 128MB
server.manage-gids: on
changelog.rollover-time: 15
changelog.fsync-interval: 3

 

 

 

===============================

David F. Robinson, Ph.D.

President - Corvid Technologies

704.799.6944 x101 [office]

704.252.1310 [cell]

704.799.7974 [fax]

 

 



_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users


_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users



--
Raghavendra Talur 

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux