[ih] TCP RTT Estimator
Greg Skinner
gregskinner0 at icloud.com
Sun Mar 16 23:06:34 PDT 2025
On Mar 11, 2025, at 1:42 PM, Jack Haverty via Internet-history <internet-history at elists.isoc.org> wrote:
>
> IMHO the key factor was the state of the Internet at that time (1980ish). The ARPANET was the primary "backbone" of The Internet in what I think of as the "fuzzy peach" stage of Internet evolution. The ARPANET was the peach, and sites on the ARPANET were adding LANs of some type and connecting them with some kind of gateway to the ARPANET IMP.
>
> The exception to that structure was Europe, especially Peter Kirstein's group at UCL and John Laws group at RSRE. They were interconnected somehow in the UK, but their access to the Internet was through a connection to a SATNET node (aka SIMP) at Goonhilly Downs.
>
> SATNET was connected to the ARPANET through one of the "core gateways" that we at BBN were responsible to run as a 24x7 operational network.
>
> The ARPANET was a packet network, but it presented a virtual circuit service to its users. Everything that went in one end came out the other end, in order, with nothing missing, and nothing duplicated. TCPs at a US site talking to TCPs at another US site didn't have much work to do, since everything they sent would be received intact. So RTT values could be set very high - I recall one common choice was 3 seconds.
>
> For the UK users however, things were quite different. The "core gateways" at the time were very limited by their hardware configurations. They didn't have much buffering space. So they did drop datagrams, which of course had to be retransmitted by the host at the end of the TCP connection. IIRC, at one point the ARPANET/SATNET gateway had exactly one datagram of buffer space.
>
> I don't recall anyone ever saying it, but I suspect that situation caused the UCL and RSRE crews to pay a lot of attention to TCP behavior, and try to figure out how best to deal with their skinny pipe across the Atlantic.
>
> At one point, someone (from UCL or RSRE, can't remember) reported an unexpected measurement. They did frequent file transfers, often trying to "time" their transfers to happen at a time of day when UK and US traffic flows would be lowest. But they observed that their transfers during "busy times" went much faster than similar transfers during "quiet times". That made little sense of course.
>
> After digging around with XNET, SNMP, etc., we discovered the cause. That ARPANET/SATNET gateway had very few buffers. The LANs at users' sites and the ARPANET path could deliver datagrams to that gateway faster than SATNET could take them. So the buffers filled up and datagrams were discarded -- just as expected.
>
> During "quiet times", the TCP connection would deliver datagrams to the gateway in bursts (whatever the TCPs negotiated as a Window size). Buffers in the gateway would overflow and some of those datagrams were lost. The sending TCP would retransmit, but only after the RTT timer expired, which was often set to 3 seconds. Result - slow FTPs.
>
> Conversely, during "busy times", the traffic through the ARPANET would be spread out in time. With other users' traffic flows present, chances were better that someone else's datagram would be dropped instead. Result - faster FTP transfers.
>
> AFAIK, none of this behavior was ever analyzed mathematically. The mathematical model of an Internet seemed beyond the capability of queuing theory et al. Progress was very much driven by experimentation and "let's try this" activity.
>
> The solution, or actually workaround, was to improve the gateway's hardware. More memory meant more buffering was available. That principle seems to have continued even today, but has caused other problems. Google "buffer bloat" if you're curious.
>
> As far as I remember, there weren't any such problems reported with the various Packet Radio networks. They tended to be used only occasionally, for tests and demos, where the SATNET linkage was used almost daily.
>
> The Laws and Kirstein groups in the UK were, IMHO, the first "real" users of TCP on The Internet, exploring paths not protected by ARPANET mechanisms.
>
> Jack Haverty
>
> --
> Internet-history mailing list
> Internet-history at elists.isoc.org
> https://elists.isoc.org/mailman/listinfo/internet-history
There was a packet radio network at the Ft. Bragg, North Carolina site, with an SRI field office supporting the XVIII Airborne Corps. [1] The TIU implementation run at that site was written by Jim Mathis. (IEN 98 and RFC 801 provide more information. The latter notes that the TIU was used daily for communications between Ft. Bragg users and ISID. A test program called PTIME used to determine “user-visible throughput” also appears in [1].) Some ISI staff supported this testbed also. [2] By 1986, the database applications such as the Tactical Reporting System were running on Sun workstations, and mostly used Sun RPC over TCP.
--gregbo
[1] https://pdos.csail.mit.edu/archive/decouto/papers/frankel82.pdf
[2] https://apps.dtic.mil/sti/tr/pdf/ADA157991.pdf
More information about the Internet-history
mailing list