Nagle & delayed ACK strike again

Damien Miller djm at mindrot.org
Fri Dec 22 14:20:12 EST 2006


sorry to interrupt your argument:

revision 1.120
date: 2006/03/15 01:05:22;  author: djm;  state: Exp;  lines: +2 -3
   - dtucker at cvs.openbsd.org 2006/03/13 08:33:00
     [packet.c]
     Set TCP_NODELAY for all connections not just "interactive" ones.  Fixes
     poor performance and protocol stalls under some network conditions (mindrot
     bugs #556 and #981). Patch originally from markus@, ok djm@

is in OpenSSH from 4.4 onwards

On Fri, 22 Dec 2006, Miklos Szeredi wrote:

> > > To me it still looks like the use of Nagle is the exception, it has
> > > already been turned off in the server for
> > > 
> > >   - interactive sessions
> > 
> > For at least some interactive sessions.  In the telnet space at least, 
> > there is this constant back and forth happening bewteen wanting 
> > keystrokes to be nice and uniform, and not overwhelming slot terminal 
> > devices (eg barcode scanners) when applications on the server dump a 
> > bunch of stuff down stdio.
> 
> For ssh this is unconditional.  I've suggested adding NoDelay/
> NoNoDelay options, but somebody on this list vetoed that.
> 
> > >   - X11 forwarding
> > > 
> > > and it will need to be turned off for
> > > 
> > >   - SFTP transport
> > > 
> > >   - IP tunnelling
> > > 
> > >   - ???
> > > 
> > > Is there any transported protocol where Nagle does make sense?
> > 
> > Regular FTP is one, anything unidirectional.
> 
> Nagle doesn't help FTP or HTTP does it?  Anything that just pushes a
> big chunk of data will automatically end up with big packets.
> 
> So other than the disputed interactive session, Nagle doesn't seem to
> have any positive effects.
> 
> > It also depends on what one is trying to optimize.  If one is only 
> > interested in optimizing time, Nagle may not be the thing.  However, 
> > Nagle can optimize the ratio of data to data+headers and it can optimize 
> > the quanity of CPU consumed per unit of data transferred.
> 
> For a filesystem protocol obviously latency (and hence throughput) is
> the most important factor.
> 
> > Some netperf data for the unidirectional case, between a system in Palo 
> > Alto and one in Cupertino, sending-side CPU utilization included, 
> > similar things can happen to receive-side CPU:
> > 
> > TCP STREAM TEST from 0.0.0.0 (0.0.0.0) port 0 AF_INET to 
> > tardy.cup.hp.com (15.244.56.217) port 0 AF_INET
> > Recv  Send   Send                        Utilization      Service Demand
> > SocketSocket Message Elapsed             Send     Recv    Send    Recv
> > Size  Size   Size    Time     Throughput local    remote  local   remote
> > bytes bytes  bytes   secs.    10^6bits/s % S      % U     us/KB   us/KB
> > 
> > 131072 219136    512    10.10      74.59   8.78   -1.00   9.648   -1.000
> > 
> > raj at tardy:~/netperf2_work$ src/netperf -H tardy.cup.hp.com -c -- -m 512 
> > -s 128K -S 128K -D
> > TCP STREAM TEST from 0.0.0.0 (0.0.0.0) port 0 AF_INET to 
> > tardy.cup.hp.com (15.244.56.217) port 0 AF_INET : nodelay
> > Recv   Send   Send                       Utilization      Service Demand
> > Socket Socket Message Elapsed            Send     Recv    Send    Recv
> > Size   Size   Size    Time    Throughput local    remote  local   remote
> > bytes  bytes  bytes   secs.   10^6bits/s % S      % U     us/KB   us/KB
> > 
> > 131072 219136   512    10.02       69.21  20.56   -1.00   24.335  -1.000
> > 
> > The multiple concurrent request/response case is more nuanced and 
> > difficule to make.  Basically, it is a race between how many small 
> > requests (or responses) will be made at one time, the RTT between the 
> > systems, the standalone ACK timer on the receiver, and the service time 
> > on the receiver.
> > 
> > Here is some data with netperf TCP_RR between those two systems:
> > 
> > raj at tardy:~/netperf2_work$ src/netperf -H tardy.cup.hp.com -c -t TCP_RR 
> > -- -r 128,2048 -b 3
> > TCP REQUEST/RESPONSE TEST from 0.0.0.0 (0.0.0.0) port 0 AF_INET to 
> > tardy.cup.hp.com (15.244.56.217) port 0 AF_INET : first burst 3
> > Local /Remote
> > Socket Size  Request Resp. Elapsed Trans.  CPU    CPU    S.dem   S.dem
> > Send   Recv  Size    Size  Time    Rate    local  remote local   remote
> > bytes  bytes bytes   bytes secs.   per sec % S    % U    us/Tr   us/Tr
> > 
> > 16384  87380 128     2048  10.00   1106.42 4.74   -1.00  42.852  -1.000
> > 32768  32768
> > raj at tardy:~/netperf2_work$ src/netperf -H tardy.cup.hp.com -c -t TCP_RR 
> > -- -r 128,2048 -b 3 -D
> > TCP REQUEST/RESPONSE TEST from 0.0.0.0 (0.0.0.0) port 0 AF_INET to 
> > tardy.cup.hp.com (15.244.56.217) port 0 AF_INET : nodelay : first burst 3
> > Local /Remote
> > Socket Size  Request Resp. Elapsed Trans.   CPU    CPU    S.dem   S.dem
> > Send   Recv  Size    Size  Time    Rate     local  remote local   remote
> > bytes  bytes bytes   bytes secs.   per sec  % S    % U    us/Tr   us/Tr
> > 
> > 16384  87380 128     2048  10.01   2145.98  10.49  -1.00  48.875  -1.000
> > 32768  32768
> > 
> > 
> > Now, setting TCP_NODELAY did indeed produce a big jump in transactions 
> > per second.  Notice though how it also resulted in a 14% increase in CPU 
> > utilization per transaction.  Clearly the lunch was not free.
> > 
> > The percentage difference in transactions per second will converge the 
> > larger the number of outstanding transactions.  Taking the settings from 
> > above, where the first column is the size of the burst in netperf, the 
> > second is without TCP_NODELAY set, the third with:
> > 
> > raj at tardy:~/netperf2_work$ for i in 3 6 9 12 15 18 21 24 27; do echo $i 
> > `src/netperf -H tardy.cup.hp.com -t TCP_RR -l 4 -P 0 -v 0 -- -r 128,2048 
> > -b $i; src/netperf -H tardy.cup.hp.com -t TCP_RR -l 4 -P 0 -v 0 -- -r 
> > 128,2048 -b $i -D`; done
> > 3 1186.40 2218.63
> > 6 1952.53 3695.64
> > 9 2574.49 4833.47
> > 12 3194.71 4856.63
> > 15 3388.54 4784.26
> > 18 4215.70 5099.52
> > 21 4645.97 5170.89
> > 24 4918.16 5336.79
> > 27 4927.71 5448.78
> > 
> > If we increase the request size to 256 bytes, and the response to 8192 
> > (In all honesty I don't know what sizes sftp might use so I'm making 
> > wild guesses) we can see the convergence happen much sooner - it takes 
> > fewer of the 8192 byte responses to take the TCP connection to the 
> > bandwidth delay product of the link:
> > 
> > raj at tardy:~/netperf2_work$ for i in 3 6 9 12 15 18 21 24 27; do echo $i 
> > `src/netperf -H tardy.cup.hp.com -t TCP_RR -l 4 -P 0 -v 0 -- -r 256,8192 
> > -b $i -s 128K -S 128K; src/netperf -H tardy.cup.hp.com -t TCP_RR -l 4 -P 
> > 0 -v 0 -- -r 256,8192 -s 128K -S 128K -b $i -D`; done
> > 3 895.18 1279.38
> > 6 1309.11 1405.38
> > 9 1395.30 1325.44
> > 12 1256.75 1422.01
> > 15 1412.39 1413.64
> > 18 1400.04 1419.76
> > 21 1415.62 1422.79
> > 24 1419.56 1420.10
> > 27 1422.43 1379.72
> 
> In SFTP the WRIYR request/reply sizes are more like 64kB/32B, and the
> outstanding transactions are as many as the socket buffers will bear.
> 
> The slowdown is clearly due to 50ms outages from delayed ACK, which is
> totally broken, the network is just sitting there idle for no good
> reason whatsoever.
> 
> I can make new traces, but I guess they would be very similar to the
> ones I sent last time for the SFTP download case.
> 
> Miklos
> _______________________________________________
> openssh-unix-dev mailing list
> openssh-unix-dev at mindrot.org
> http://lists.mindrot.org/mailman/listinfo/openssh-unix-dev
> 


More information about the openssh-unix-dev mailing list