[ih] TCP RTT Estimator

vinton cerf vgcerf at gmail.com
Tue Mar 11 14:10:45 PDT 2025


jack'd recollections mirror my own.

v


On Tue, Mar 11, 2025 at 4:43 PM Jack Haverty via Internet-history <
internet-history at elists.isoc.org> wrote:

> On 3/11/25 07:05, David Finnigan via Internet-history wrote:
> > It looks like staff at RSRE (Royal Signals and Radar Establishment) took
> > the lead in experimenting with formulae and methods for dynamic
> > estimation of round trip times in TCP. Does anyone here have any further
> > insight or recollection into these experiments for estimating RTT, and
> > the development of the RTT formula?
> >
>
> IMHO the key factor was the state of the Internet at that time
> (1980ish).  The ARPANET was the primary "backbone" of The Internet in
> what I think of as the "fuzzy peach" stage of Internet evolution.   The
> ARPANET was the peach, and sites on the ARPANET were adding LANs of some
> type and connecting them with some kind of gateway to the ARPANET IMP.
>
> The exception to that structure was Europe, especially Peter Kirstein's
> group at UCL and John Laws group at RSRE.   They were interconnected
> somehow in the UK, but their access to the Internet was through a
> connection to a SATNET node (aka SIMP) at Goonhilly Downs.
>
> SATNET was connected to the ARPANET through one of the "core gateways"
> that we at BBN were responsible to run as a 24x7 operational network.
>
> The ARPANET was a packet network, but it presented a virtual circuit
> service to its users.  Everything that went in one end came out the
> other end, in order, with nothing missing, and nothing duplicated. TCPs
> at a US site talking to TCPs at another US site didn't have much work to
> do, since everything they sent would be received intact.   So RTT values
> could be set very high - I recall one common choice was 3 seconds.
>
> For the UK users however, things were quite different.  The "core
> gateways" at the time were very limited by their hardware
> configurations.  They didn't have much buffering space.   So they did
> drop datagrams, which of course had to be retransmitted by the host at
> the end of the TCP connection.  IIRC, at one point the ARPANET/SATNET
> gateway had exactly one datagram of buffer space.
>
> I don't recall anyone ever saying it, but I suspect that situation
> caused the UCL and RSRE crews to pay a lot of attention to TCP behavior,
> and try to figure out how best to deal with their skinny pipe across the
> Atlantic.
>
> At one point, someone (from UCL or RSRE, can't remember) reported an
> unexpected measurement.  They did frequent file transfers, often trying
> to "time" their transfers to happen at a time of day when UK and US
> traffic flows would be lowest.   But they observed that their transfers
> during "busy times" went much faster than similar transfers during
> "quiet times".  That made little sense of course.
>
> After digging around with XNET, SNMP, etc., we discovered the cause.
> That ARPANET/SATNET gateway had very few buffers.  The LANs at users'
> sites and the ARPANET path could deliver datagrams to that gateway
> faster than SATNET could take them.  So the buffers filled up and
> datagrams were discarded -- just as expected.
>
> During "quiet times", the TCP connection would deliver datagrams to the
> gateway in bursts (whatever the TCPs negotiated as a Window size).
> Buffers in the gateway would overflow and some of those datagrams were
> lost.  The sending TCP would retransmit, but only after the RTT timer
> expired, which was often set to 3 seconds. Result - slow FTPs.
>
> Conversely, during "busy times", the traffic through the ARPANET would
> be spread out in time.   With other users' traffic flows present,
> chances were better that someone else's datagram would be dropped
> instead.  Result - faster FTP transfers.
>
> AFAIK, none of this behavior was ever analyzed mathematically.  The
> mathematical model of an Internet seemed beyond the capability of
> queuing theory et al.  Progress was very much driven by experimentation
> and "let's try this" activity.
>
> The solution, or actually workaround, was to improve the gateway's
> hardware.  More memory meant more buffering was available.   That
> principle seems to have continued even today, but has caused other
> problems.  Google "buffer bloat" if you're curious.
>
> As far as I remember, there weren't any such problems reported with the
> various Packet Radio networks.   They tended to be used only
> occasionally, for tests and demos, where the SATNET linkage was used
> almost daily.
>
> The Laws and Kirstein groups in the UK were, IMHO, the first "real"
> users of TCP on The Internet, exploring paths not protected by ARPANET
> mechanisms.
>
> Jack Haverty
>
> --
> Internet-history mailing list
> Internet-history at elists.isoc.org
> https://elists.isoc.org/mailman/listinfo/internet-history
>


More information about the Internet-history mailing list