Looks like RAMFS has the same issue TMPFS does... I'm looking into the RNA networks. Thanks- Jeremy On 11/18/2010 6:55 PM, Craig Carl wrote: > On 11/18/2010 04:33 PM, Jeremy Enos wrote: >> Post is almost a year old... ever any response here? Is it possible >> to export tmpfs locations w/ gluster? >> thx- >> >> Jeremy >> >> On 12/1/2009 8:14 AM, Alexander Beregalov wrote: >>> Hi >>> >>> Is it possible to start server on tmpfs ? >>> It is announced that stripe can be used over tmpfs, but stripe is >>> client plugin and server cannot start on tmpfs because lack of >>> xattrs. >>> >>> I am trying to setup a small fast storage for small files (compiling >>> purpose). >>> I made ext2 with xattr on ramdisk on 4 hosts, joined them with >>> replicate plugin and mounted it on one client. Also io-cache, >>> write-behind, quick-read and io threads were used on client side. >>> I compiled linux kernel, performance was 10 times worse than tmpfs >>> exported by NFS on one node. >>> >>> Any ideas? >>> >>> Servers: >>> >>> volume posix >>> type storage/posix # POSIX FS translator >>> option directory /mnt/ost # Export this directory >>> end-volume >>> >>> volume locks >>> type features/locks >>> option mandatory-locks on >>> subvolumes posix >>> end-volume >>> >>> volume brick >>> type performance/io-threads >>> option thread-count 4 # Four CPUs >>> subvolumes locks >>> end-volume >>> >>> volume server >>> type protocol/server >>> option transport-type tcp >>> option transport.socket.nodelay on >>> subvolumes brick >>> option auth.addr.brick.allow * # Allow access to "brick" volume >>> end-volume >>> >>> >>> Client: >>> >>> volume server1 >>> type protocol/client >>> option transport-type tcp >>> option remote-host<IP> >>> option transport.socket.nodelay on >>> option remote-subvolume brick # name of the remote volume >>> end-volume >>> >>> <the same for server[2-4]> >>> >>> volume replicated >>> type cluster/replicate >>> subvolumes server1 server2 server3 server4 >>> end-volume >>> >>> volume iocache >>> type performance/io-cache >>> option cache-size 1000MB # default is 32MB >>> option priority *.h:3,*.o:2,*:1 # default is '*:0' >>> option cache-timeout 1 # default is 1 second >>> subvolumes replicated >>> end-volume >>> >>> volume writeback >>> type performance/write-behind >>> option cache-size 500MB # default is equal to aggregate-size >>> option flush-behind off # default is 'off' >>> subvolumes iocache >>> end-volume >>> >>> volume quickread >>> type performance/quick-read >>> option cache-timeout 1 # default 1 second >>> option max-file-size 256KB # default 64Kb >>> subvolumes iocache >>> end-volume >>> >>> volume iothreads >>> type performance/io-threads >>> option thread-count 16 # default is 16 >>> subvolumes quickread >>> end-volume >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users > > Jeremy - > xattrs are required for any Gluster setup regardless of the volume > design, tmpfs doesn't support user xattrs. RamFS works well, if you > have the budget FisionIO cards are very fast and work well with > Gluster, so does the solution from RNA Networks. > (http://www.rnanetworks.com/) > > Thanks, > > Craig > > --> > Craig Carl > Senior Systems Engineer > Gluster > _______________________________________________ > Gluster-users mailing list > Gluster-users at gluster.org > http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >