gluster-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Gluster-devel] solutions for split brain situation


From: Mark Mielke
Subject: Re: [Gluster-devel] solutions for split brain situation
Date: Mon, 14 Sep 2009 10:33:13 -0400
User-agent: Mozilla/5.0 (X11; U; Linux x86_64; en-US; rv:1.9.1.1) Gecko/20090814 Fedora/3.0-2.6.b3.fc11 Thunderbird/3.0b3

On 09/14/2009 10:18 AM, Stephan von Krawczynski wrote:
Our "split brain" is no real split brain and looks like this: Logfiles are
written every 5 mins. If you add a secondary server that has 14 days old
logfiles on it you notice that about half of your data vanishes while not
successful self heal is performed, because the old logfiles read from the
secondary server overwrite the new logfiles on your primary while new data is
added to them. This is a very simple and solvable situation. All you had to do
to win the situation 100% is to compare the files' mtime.

I'd argue that "longest running glusterfsd" is the right self-heal solution here as well (as opposed to "latest mtime"). The secondary server should come up with zero uptime, and have the lowest precedence.

Whereas a true split brain is rare, our situation arises every time you add a
server, maybe because you made a kernel update or needed a reboot for some
reason. Your secondary comes back and kicks your ass. Even better, it is
completely irrelevant which server gets re-added, as soon as you have old data
on it you are busted.

This is interesting. I wondered about this reading the documentation. I came to the conclusion that there is supposed to be some sort of version attribute attached to the file that will resolve this situation. Are you doing something special - such as removing the volume from your configuration, and then re-adding it? I don't know how it works - but if the system simply goes down for a period - for kernel update or reboot - I am lead to believe that everything should be fine. Have you tried this?

You might argue to prevent that by simply deleting everything on a newly added
server. But if you deal with TBs of data you really do not want to spend the
time and network bandwidth to heal the data, when most of it is actually in
good shape and only some MBs or GBs are outdated.
Btw I know this is not what you call "split brain", but glusterfs thinks it
is, and that is part of the problem. It cannot distinguish the cases.
Your argument is broken anyways because in your situation you will loose the
data no matter if you keep the current implementation or create a new "option
favorite-child mtime" option. In the current implementation you will loose
about every other file content summing up to 100% of the files being damaged,
iff in a true split brain both servers get new data for their respective
fileset and are mixed together later on. If the file comes from server A you
lost all data added on server B during split brain and vice versa.
Thinking about it it sounds as if the current implementation is the worst
possible. There is really no good reason for distributing file access in a
split brain detect situation. At least it should then choose the same child
for following file access to prevent the 100% loss.
Another idea would be switching split-brain files to read-only access. This
would be the conservative approach of not loosing already written data - only
new writes get lost this way.

I think you want the "hot add / hot remove" functionality on the roadmap for 2.2. If you are removing the volume while the system is down, then you are using it outside the design case for the solution at present.

I agree it should work - eventually - but I think your use is outside of intended scope at the moment.

Now, if your system is going down - no removal of the volume - and it comes back up with the behaviour you describe, then I am very concerned as well.

My opinion, anyways.

Cheers,
mark

--
Mark Mielke<address@hidden>





reply via email to

[Prev in Thread] Current Thread [Next in Thread]