Start a new volume with pre-existing directories

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Can you send the output of -

`gluster volume info all`
`gluster peer status`

from a gluster storage server and

`mount` from the client?

Craig


On 12/03/2010 02:50 AM, Daniel Zander wrote:
> Dear Craig,
>
> I'm afraid I wasn't able to start a self heal the way you suggested. I 
> tested the following:
>
> ----------
> WHAT I DID
> ----------
> I created on fs7:/storage/7
> user_7_1  user_7_2  user_7_3
>
> and on fs8:/storage/8
> user_8_1  user_8_2  user_8_3
>
> and filled all of the directories with some small files and 
> subdirectories.
>
> Then, on fs8:
> gluster volume create heal_me transport tcp 192.168.101.246:/storage/8 
> 192.168.101.247:/storage/7
>
> Then on fs8 and afterwards on fs7:
> mount -t glusterfs localhost:/heal_me /tempmount/
> cd /tempmount
> find . | xargs stat >>/dev/null 2>&1
> umount /tempmount
>
> All went well, no error messages or anything. The output of `find . | 
> xargs stat` is probably too long to post it here, but there were no 
> error messages or anything else that would look suspicious to me.
>
> -------
> RESULTS
> -------
> ls fs8:storage/8
> user_8_1  user_8_2  user_8_3
>
> ls fs7:/storage/7
> user_7_1  user_7_2  user_7_3  user_8_1  user_8_2  user_8_3
>
> ls client:/storage/gluster
> user_8_1  user_8_2  user_8_3
>
> ls fs7:/tempmount
> user_8_1  user_8_2  user_8_3
>
> ls fs8:/tempmount
> user_8_1  user_8_2  user_8_3
>
> Unmounting and remounting has no effect.
>
> Servers are both Ubuntu Server 10.4, client is CentOS 5, 64bits all 
> around.
>
> Thanks and regards,
> Daniel
>
>
> On 12/03/2010 10:10 AM, Craig Carl wrote:
>> Daniel -
>> If you want to export existing data you will need to run the self heal
>> process so extended attributes can get written. While this should work
>> without any issues it isn't an officially supported process, please make
>> sure you have complete and up to date backups.
>>
>> After you have setup and started the Gluster volume mount it locally on
>> one of the servers using `mount -t glusterfs localhost:/<volname> /<some
>> temporary mount>`. CD into the root of the mount point and run `find . |
>> xargs stat >>/dev/null 2>&1` to start a self heal.
>>
>> Also the command you used to create the volume should not have worked,
>> it is missing a volume name - gluster volume create <VOLNAME> transport
>> tcp fs7:/storage/7, fs8:/storage/8, typo maybe?
>>
>> Please let us know how it goes, and please let me know if you have any
>> other questions.
>>
>> Thanks,
>>
>> Craig
>>
>> -->
>> Craig Carl
>> Senior Systems Engineer; Gluster, Inc.
>> Cell - (408) 829-9953 (California, USA)
>> Office - (408) 770-1884
>> Gtalk - craig.carl at gmail.com
>> Twitter - @gluster
>> http://rackerhacker.com/2010/08/11/one-month-with-glusterfs-in-production/ 
>>
>>
>>
>>
>>
>> On 12/02/2010 11:38 PM, Daniel Zander wrote:
>>> Dear all,
>>>
>>> at our institute, we currently have 6 file servers, each one of them
>>> individually mounted via NFS on ~ 20 clients. The structure on the
>>> servers and the clients is the following:
>>>
>>> /storage/1/<user_directories> (NFS export from FS1)
>>> /storage/2/<user_directories> (NFS export from FS2)
>>> etc ...
>>>
>>> Recently, we decided that we would like to migrate this to glusterFS,
>>> so that we can have one big storage directory on the clients. Let's
>>> call it
>>>
>>> /gluster/<user_directories>
>>>
>>> I tried to set up a gluster volume with two empty fileservers and it
>>> worked without any problems. I could easily mount it on a client and
>>> use it (using the native glusterFS mount).
>>>
>>> If we now want to migrate the entire institute, it would be very
>>> convenient, if existing folders could be easily included into a new
>>> volume. I tried to do this, but I did not succeed.
>>>
>>> Here's a short description of what I tried:
>>>
>>> Existing folders:
>>> on fs7: /storage/7/user_1,user_2
>>> on fs8: /storage/8/user_3,user_4
>>>
>>> gluster volume create transport tcp fs7:/storage/7, fs8:/storage/8
>>>
>>> I hoped to see on the client:
>>> /gluster/user_1
>>> /gluster/user_2
>>> /gluster/user_3
>>> /gluster/user_4
>>>
>>> The creation was successful, the volume could be started and mounted.
>>> On the client, however, I could only find (via "ls /gluster") the
>>> directories user_1 and user_2. But when I tried "cd /gluster/user_3",
>>> it succeeded! Now "ls /gluster" showed me user_1, user_2 and user_3.
>>> Unfortunately, user_3's subdirectories and files were still invisible,
>>> but with the above mentioned trick, I could make them visible.
>>>
>>> This is however not an option, as there are too much users and too
>>> complicated file structures to do this manually. It anyhow seems like
>>> Voodoo to me.
>>>
>>> Is it possible to include all of the existing directories in the new
>>> glusterFS volume? If yes: how?
>>>
>>> Thank you in advance for your efforts,
>>> Regards,
>>> Daniel
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users at gluster.org
>>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users


[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux