Re: [PATCH RFC] vfs: make fstatat retry on ESTALE errors from getattrcall

From: Steve Dickson
Date: Tue Apr 17 2012 - 07:54:19 EST




On 04/15/2012 03:57 PM, Chuck Lever wrote:
>
> On Apr 15, 2012, at 3:03 PM, Bernd Schubert wrote:
>
>> On 04/13/2012 05:42 PM, Jeff Layton wrote:
>>> (note: please don't trim the CC list!)
>>>
>>> Indefinitely does make some sense (as Peter articulated in his original
>>> set). It's possible you could race several times in a row, or a server
>>> misconfiguration or something has happened and you have a transient
>>> error that will eventually recover. His assertion was that any limit on
>>> the number of retries is by definition wrong. For NFS, a fatal signal
>>> ought to interrupt things as well, so retrying indefinitely has some
>>> appeal there.
>>>
>>> OTOH, we do have to contend with filesystems that might return ESTALE
>>> persistently for other reasons and that might not respond to signals.
>>> Miklos pointed out that some FUSE fs' do this in his review of Peter's
>>> set.
>>>
>>> As a purely defensive coding measure, limiting the number of retries to
>>> something finite makes sense. If we're going to do that though, I'd
>>> probably recommend that we set the number of retries be something
>>> higher just so that this is more resilient in the face of multiple
>>> races. Those other fs' might "spin" a bit in that case but it is an
>>> error condition and IMO resiliency trumps performance -- at least in
>> this case.
>>
>> I am definitely voting against an infinite number of retries. I'm
>> working on FhGFS, which supports distributed meta data servers. So when
>> a file is moved around between directories, its file handle, which
>> contains the meta-data target id might become invalid.
>
> Doesn't Jeff's recovery mechanism resolve this situation? The client does a fresh lookup, so shouldn't it get the new FH at that point? If there is no possible way for the client to discover the new FH, then ESTALE recovery should probably be finite.
>
>> As NFSv3 is
>> stateless we cannot inform the client about that and must return ESTALE
>> then. NFSv4 is better, but I'm not sure how well invalidating a file
>> handle works. So retrying once on ESTALE might be a good idea, but
>> retrying forever is not.
>> Also, what about asymmetric HA servers? I believe to remember that also
>> resulted in ESTALE. So for example server1 exports /home and /scratch,
>> but on failure server2 can only take over /home and denies access to
>> /scratch.
>
> Retrying forever is bad only if we think there are cases where there is no possible recovery action for the client, or the ESTALE signals a condition that is not temporary.
>
> It is temporary, for instance, when an administrator takes an exported volume offline; at some point, that volume will be brought back online. Maybe it is better generally for the client to retry indefinitely instead of causing applications to fail.
>
> Retrying forever is exactly what we do for "hard" NFS mounts, for example, and for many types of NFSv4 state recovery. Philosophically, how is this situation different?
The analogy of mount looping and ESTATE loop is a bit of a stretch IMHO...
With mount looping the server has not responded so it make sense to wait for a response (aka the first response). With ESTALE looping the server has actively remove something the client has been working with... I see those as two different events so they need to be handled differently...

>
> It would be reasonable, at least, to insert a backoff delay in the retry logic.
If looping is going to be involved in ESTALE handling then a time based loop does make more sense than a simple for loop.. IMHO... Similar to Jeff's version 2 patch...

steved.

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/