On 09/14/2009 10:18 AM, Stephan von Krawczynski wrote:
Our "split brain" is no real split brain and looks like this: Logfiles are written every 5 mins. If you add a secondary server that has 14 days old logfiles on it you notice that about half of your data vanishes while not successful self heal is performed, because the old logfiles read from the secondary server overwrite the new logfiles on your primary while new data is added to them. This is a very simple and solvable situation. All you had to do to win the situation 100% is to compare the files' mtime.
I'd argue that "longest running glusterfsd" is the right self-heal solution here as well (as opposed to "latest mtime"). The secondary server should come up with zero uptime, and have the lowest precedence.
Whereas a true split brain is rare, our situation arises every time you add a server, maybe because you made a kernel update or needed a reboot for some reason. Your secondary comes back and kicks your ass. Even better, it is completely irrelevant which server gets re-added, as soon as you have old data on it you are busted.
This is interesting. I wondered about this reading the documentation. I came to the conclusion that there is supposed to be some sort of version attribute attached to the file that will resolve this situation. Are you doing something special - such as removing the volume from your configuration, and then re-adding it? I don't know how it works - but if the system simply goes down for a period - for kernel update or reboot - I am lead to believe that everything should be fine. Have you tried this?
You might argue to prevent that by simply deleting everything on a newly added server. But if you deal with TBs of data you really do not want to spend the time and network bandwidth to heal the data, when most of it is actually in good shape and only some MBs or GBs are outdated. Btw I know this is not what you call "split brain", but glusterfs thinks it is, and that is part of the problem. It cannot distinguish the cases. Your argument is broken anyways because in your situation you will loose the data no matter if you keep the current implementation or create a new "option favorite-child mtime" option. In the current implementation you will loose about every other file content summing up to 100% of the files being damaged, iff in a true split brain both servers get new data for their respective fileset and are mixed together later on. If the file comes from server A you lost all data added on server B during split brain and vice versa. Thinking about it it sounds as if the current implementation is the worst possible. There is really no good reason for distributing file access in a split brain detect situation. At least it should then choose the same child for following file access to prevent the 100% loss. Another idea would be switching split-brain files to read-only access. This would be the conservative approach of not loosing already written data - only new writes get lost this way.
I think you want the "hot add / hot remove" functionality on the roadmap for 2.2. If you are removing the volume while the system is down, then you are using it outside the design case for the solution at present.
I agree it should work - eventually - but I think your use is outside of intended scope at the moment.
Now, if your system is going down - no removal of the volume - and it comes back up with the behaviour you describe, then I am very concerned as well.
My opinion, anyways. Cheers, mark -- Mark Mielke<mark@xxxxxxxxx>