Bryan, can you please post the glusterfs log files? Also please try with posix-locks loaded. thanks, avati 2008/12/4 Bryan Whitehead <driver at megahappy.net>: > glusterfs 1.3.12 > > Client conf: > volume beavis > type protocol/client > option transport-type tcp/client > option remote-host beavis > option remote-subvolume gluster1 > end-volume > > volume butthead > type protocol/client > option transport-type tcp/client > option remote-host butthead > option remote-subvolume gluster1 > end-volume > > volume mirror0 > type cluster/afr > subvolumes beavis butthead > end-volume > > Server conf: > volume gluster1 > type storage/posix > option directory /export/gluster1 > end-volume > > volume gluster1-locks > type features/posix-locks > option mandatory on > subvolumes gluster1 > end-volume > > volume gluster1-io-thr > type performance/io-threads > subvolumes gluster1-locks > end-volume > > volume gluster1-wb > type performance/write-behind > subvolumes gluster1-io-thr > end-volume > > volume gluster1-ra > type performance/read-ahead > subvolumes gluster1-wb > end-volume > > volume server > type protocol/server > option transport-type tcp/server > option auth.ip.gluster1.allow 10.* > subvolumes gluster1-ra > end-volume > > If you'd like a complete openldap error, I'll have to do another attempt > tonight. Previously, I had not used the posix-locks, io-threads, > write-behind, or read-ahead. My first attempt was with a server config that > looked like this: > > volume gluster1 > type storage/posix > option directory /export/gluster1 > end-volume > > volume server > type protocol/server > option transport-type tcp/server > option auth.ip.gluster1.allow 10.* > subvolumes gluster1 > end-volume > > But with the error from ldap, I thought that posix-locks might be needed. > > Anand Avati wrote: >> >> Can you give some more details - the version, spec files and log files? >> >> avati >> >> 2008/12/4 Bryan Whitehead <driver at megahappy.net>: >>> >>> I have a very simple setup of gluster. I basically have a mirror using >>> afr on the client side. When I attempt to start zimbra (which begins by >>> trying to start openldap) I get a weird error: >>> >>> backend_startup_one: bi_db_open failed! (-1) >>> >>> I have a functional zimbra setup. What I do is stop zimbra, rsync over >>> the installation, mv the existing directory aside, then symlink (or >>> create directory and directly mount from gluster). I cannot get zimbra >>> to startup. :( >>> >>> example: >>> >>> /opt/zimbra is where everything is. Program, data, everything. >>> >>> /etc/init.d/zimbra stop >>> time rsync -arv --delete /opt/* /import/gluster1/zimbra/opt/ >>> cd /opt >>> mv zimbra zimbra.works >>> ln -s /import/gluster1/zimbra/opt/zimbra >>> /etc/init.d/zimbra start >>> >>> if I undo the symlink (or mountpoint/directory) and move the >>> /opt/zimbra.works back to /opt/zimbra, I can start just fine like >>> nothing happened. >>> >>> -Bryan >>> >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://zresearch.com/cgi-bin/mailman/listinfo/gluster-users >>> >