I'm definitely interested in getting ramfs to work directly (instead of the hacky loopback on top of it tmpfs). I'll check again to be sure, but I'm pretty sure I got the same errors I did w/ tmpfs about xattrs when I tried to export ramfs. thx- Jeremy On 11/19/2010 7:27 PM, Craig Carl wrote: > Jeremy - > We use ramfs to test with here, it should work. If you want to > troubleshoot that issue please let me know. > > Craig > > > On 11/19/2010 05:12 PM, Jeremy Enos wrote: >> Looks like RAMFS has the same issue TMPFS does... I'm looking into >> the RNA networks. Thanks- >> >> Jeremy >> >> On 11/18/2010 6:55 PM, Craig Carl wrote: >>> On 11/18/2010 04:33 PM, Jeremy Enos wrote: >>>> Post is almost a year old... ever any response here? Is it >>>> possible to export tmpfs locations w/ gluster? >>>> thx- >>>> >>>> Jeremy >>>> >>>> On 12/1/2009 8:14 AM, Alexander Beregalov wrote: >>>>> Hi >>>>> >>>>> Is it possible to start server on tmpfs ? >>>>> It is announced that stripe can be used over tmpfs, but stripe is >>>>> client plugin and server cannot start on tmpfs because lack of >>>>> xattrs. >>>>> >>>>> I am trying to setup a small fast storage for small files >>>>> (compiling purpose). >>>>> I made ext2 with xattr on ramdisk on 4 hosts, joined them with >>>>> replicate plugin and mounted it on one client. Also io-cache, >>>>> write-behind, quick-read and io threads were used on client side. >>>>> I compiled linux kernel, performance was 10 times worse than tmpfs >>>>> exported by NFS on one node. >>>>> >>>>> Any ideas? >>>>> >>>>> Servers: >>>>> >>>>> volume posix >>>>> type storage/posix # POSIX FS translator >>>>> option directory /mnt/ost # Export this directory >>>>> end-volume >>>>> >>>>> volume locks >>>>> type features/locks >>>>> option mandatory-locks on >>>>> subvolumes posix >>>>> end-volume >>>>> >>>>> volume brick >>>>> type performance/io-threads >>>>> option thread-count 4 # Four CPUs >>>>> subvolumes locks >>>>> end-volume >>>>> >>>>> volume server >>>>> type protocol/server >>>>> option transport-type tcp >>>>> option transport.socket.nodelay on >>>>> subvolumes brick >>>>> option auth.addr.brick.allow * # Allow access to "brick" volume >>>>> end-volume >>>>> >>>>> >>>>> Client: >>>>> >>>>> volume server1 >>>>> type protocol/client >>>>> option transport-type tcp >>>>> option remote-host<IP> >>>>> option transport.socket.nodelay on >>>>> option remote-subvolume brick # name of the remote volume >>>>> end-volume >>>>> >>>>> <the same for server[2-4]> >>>>> >>>>> volume replicated >>>>> type cluster/replicate >>>>> subvolumes server1 server2 server3 server4 >>>>> end-volume >>>>> >>>>> volume iocache >>>>> type performance/io-cache >>>>> option cache-size 1000MB # default is 32MB >>>>> option priority *.h:3,*.o:2,*:1 # default is '*:0' >>>>> option cache-timeout 1 # default is 1 second >>>>> subvolumes replicated >>>>> end-volume >>>>> >>>>> volume writeback >>>>> type performance/write-behind >>>>> option cache-size 500MB # default is equal to aggregate-size >>>>> option flush-behind off # default is 'off' >>>>> subvolumes iocache >>>>> end-volume >>>>> >>>>> volume quickread >>>>> type performance/quick-read >>>>> option cache-timeout 1 # default 1 second >>>>> option max-file-size 256KB # default 64Kb >>>>> subvolumes iocache >>>>> end-volume >>>>> >>>>> volume iothreads >>>>> type performance/io-threads >>>>> option thread-count 16 # default is 16 >>>>> subvolumes quickread >>>>> end-volume >>>>> _______________________________________________ >>>>> Gluster-users mailing list >>>>> Gluster-users at gluster.org >>>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users at gluster.org >>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>> >>> Jeremy - >>> xattrs are required for any Gluster setup regardless of the >>> volume design, tmpfs doesn't support user xattrs. RamFS works well, >>> if you have the budget FisionIO cards are very fast and work well >>> with Gluster, so does the solution from RNA Networks. >>> (http://www.rnanetworks.com/) >>> >>> Thanks, >>> >>> Craig >>> >>> --> >>> Craig Carl >>> Senior Systems Engineer >>> Gluster >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users >>> >