Re: Major network performance regression in 3.7

From: Eric Dumazet
Date: Sun Jan 06 2013 - 11:45:32 EST


On Sun, 2013-01-06 at 16:51 +0100, Willy Tarreau wrote:
> Hi Eric,
>

> Oh sorry, I didn't really want to pollute the list with links and configs,
> especially during the initial report with various combined issues :-(
>
> The client is my old "inject" tool, available here :
>
> http://git.1wt.eu/web?p=inject.git
>
> The server is my "httpterm" tool, available here :
>
> http://git.1wt.eu/web?p=httpterm.git
> Use "-O3 -DENABLE_POLL -DENABLE_EPOLL -DENABLE_SPLICE" for CFLAGS.
>
> I'm starting httpterm this way :
> httpterm -D -L :8000 -P 256
> => it starts a server on port 8000, and sets pipe size to 256 kB. It
> uses SPLICE_F_MORE on output data but removing it did not fix the
> issue one of the early tests.
>
> Then I'm starting inject this way :
> inject -o 1 -u 1 -G 0:8000/?s=1g
> => 1 user, 1 object at a time, and fetch /?s=1g from the loopback.
> The server will then emit 1 GB of data using splice().
>
> It's possible to disable splicing on the server using -dS. The client
> "eats" data using recv(MSG_TRUNC) to avoid a useless copy.
>
> > TCP has very low defaults concerning initial window, and it appears you
> > set RCVBUF to even smaller values.
>
> Yes, you're right, my bootup scripts still change the default value, though
> I increase them to larger values during the tests (except the one where you
> saw win 8030 due to the default rmem set to 16060). I've been using this
> value in the past with older kernels because it allowed an integer number
> of segments to fit into the default window, and offered optimal performance
> with large numbers of concurrent connections. Since 2.6, tcp_moderate_rcvbuf
> works very well and this is not needed anymore.
>
> Anyway, it does not affect the test here. Good kernels are OK whatever the
> default value, and bad kernels are bad whatever the default value too.
>
> Hmmm finally it's this commit again :
>
> 2f53384 tcp: allow splice() to build full TSO packets
>
> I'm saying "again" because we already diagnosed a similar effect several
> months ago that was revealed by this patch and we fixed it with the
> following one, though I remember that we weren't completely sure it
> would fix everything :
>
> bad115c tcp: do_tcp_sendpages() must try to push data out on oom conditions
>
> Just out of curiosity, I tried to re-apply the patch above just after the
> first one but it did not change anything (after all it changed a symptom
> which appeared in different conditions).
>
> Interestingly, this commit (2f53384) significantly improved performance
> on spliced data over the loopback (more than 50% in this test). In 3.7,
> it seems to have no positive effect anymore. I reverted it using the
> following patch and now the problem is fixed (mtu=64k works fine now) :
>
> diff --git a/net/ipv4/tcp.c b/net/ipv4/tcp.c
> index e457c7a..61e4517 100644
> --- a/net/ipv4/tcp.c
> +++ b/net/ipv4/tcp.c
> @@ -935,7 +935,7 @@ wait_for_memory:
> }
>
> out:
> - if (copied && !(flags & MSG_SENDPAGE_NOTLAST))
> + if (copied)
> tcp_push(sk, flags, mss_now, tp->nonagle);
> return copied;
>
> Regards,
> Willy
>

Hmm, I'll have to check if this really can be reverted without hurting
vmsplice() again.



--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/