[ih] ruggedized Honeywell 516 ARPANET IMP cabinet top lifting hooks (Was: The IMP Lights story (Was: Nit-picking an origin story))

the keyboard of geoff goodfellow geoff at iconia.com
Tue Aug 26 10:15:59 PDT 2025


when yours truly was visiting MIT in the 70's there seemed to be a similar
"practice" of taking things up to the top of the Green Building (also known
as the Cecil and Ida Green Building or Building 54) -- that stands as one
of the tallest buildings in Cambridge (at 21 stories) -- and dropping them
off to see what would happen -- the most "notable" IIRC was a whole bunch
of door knobs.

geoff

On Mon, Aug 25, 2025 at 10:44 AM Karl Auerbach via Internet-history <
internet-history at elists.isoc.org> wrote:

> The UCLA computer club had a kinda sub-rosa practice of taking things
> (superballs frozen in liquid nitrogen, empty lead radiation containers,
> etc) up to the top of Boelter Hall (9 floors up) and dropping them off
> to see what would happen.  (The landing/crash/bounce zone was where the
> Institute of Traffic and Transportation Engineering - my project group -
> had stacked the wrecks resulting from its early car-crash tests.)
>
> Anyway, Mark Kampe took a look at those lifting hooks on the top of IMP
> #1 and began asking whether the IMP was truly ruggedized and whether we
> ought to test that assertion by subjecting the IMP to the Boelter Hall
> drop test.
>
> We resisted that temptation.
>
> (A few years later I did end up dropping, accidentally, an early HP
> minicomputer onto a floor.  Surprisingly it worked afterwords. It was
> one tough machine.)
>
>          --karl--
>
> On 8/25/25 6:00 AM, Steve Crocker via Internet-history wrote:
> > The 516s were ruggedized but they were not "milspec."  Frank Heart was
> > concerned they might get roughed up in transit or, possibly, in the
> > research labs where they were accessible to STUDENTS!
> >
> > I believe the hooks -- really eyes, to be precise -- on top were for
> being
> > picked up by a crane for loading and unloading.  I don't know if they
> were
> > ever used.
> >
> > There was no requirement or expectation that the IMPs would survive a
> > nuclear event.  The idea that the net was designed for nuclear
> > survivability is a red herring.  At best there was the possibility that
> > some aspects of the technology, if successful, might be useful in
> > the future for designing a nuclear survivable communications system.  It
> > was NOT a design goal for the Arpanet.
> >
> > Steve
> >
> >
> >
> > On Sun, Aug 24, 2025 at 9:38 PM the keyboard of geoff goodfellow <
> > geoff at iconia.com> wrote:
> >
> >> steve, alex, ben, scott, et al. thanks so very much for the
> >> color(ful) esoterica on the "The IMP Lights" reliability history!
> >>
> >> yours truly's next esoteric 516 IMP issue is:
> >>
> >> what were the 516 ARPANET IMP cabinet lifting hooks for?
> >>
> >> in seeing "nuclear survivability" mentioned below in one of Steve's
> >> back-and-forths: it jogged a memory of past where someone once
> mentioning
> >> -- perhaps in "josh" -- eons ago that the 516 ARPANET IMP cabinet top
> >> lifting hooks were put there so that a 516 ARPANET IMP could "easily be
> >> loaded into a (nuclear) submarine" (!)
> >>
> >> sadly, yours truly can't recall where or what mailing list that was
> >> mentioned on... but it got yours truly "thinking" about the
> >> nature/origin/purpose of the 4 516 IMP cabinet top lifting hooks ➔➔
> >> https://iconia.com/516IMP.JPG ?
> >>
> >> geoff
> >>
> >> On Sun, Aug 24, 2025 at 12:38 PM Steve Crocker via Internet-history <
> >> internet-history at elists.isoc.org> wrote:
> >>
> >>> [Apparently the attachment got stripped when I sent this to the
> >>> Internet-History list.  This is a resend, with the text of the
> interview
> >>> copied directly into this message at the end.]
> >>>
> >>> Folks,
> >>>
> >>> On Monday, 18 August 2025, I described how the lights on the IMPs often
> >>> burned out and caused a noticeable amount of downtime on the IMPs.
> Geoff
> >>> Goodfellow asked for more details.  That exchange is copied below.
> >>>
> >>> I learned of the problem with the IMP lights during a virtual
> roundtable
> >>> with Ben Barker and others.  We published the roundtable in [1].
> >>>
> >>> I later interviewed Ben and Scott Bradner to learn more details. The
> >>> interview [2] is attached. appended.
> >>>
> >>> In the process of checking, Alex McKenzie sent me a more recent article
> >>> Dave Walden, he and Ben wrote which covers several incidents related to
> >>> reliability, and he sent a reference to the article to the list.  See
> [3]
> >>> below.  I also learned that Ben passed away two years ago.  I'm sad.
> He
> >>> was a delightful and always positive guy.
> >>>
> >>> After further discussion with Alex, we agreed [1] has the least detail.
> >>> [3] is best, but it's behind a paywall.  The interview [2] is a
> >>> close second.
> >>>
> >>> I think this is all the information that's available.
> >>>
> >>> Thanks to Ben for the delightful story, to Geoff for asking for the
> >>> details, to Scott for permission to use the interview, and to Alex for
> the
> >>> recent article and advice on how to proceed.
> >>>
> >>> Steve
> >>>
> >>>
> >>> [1]  "The Arpanet and Its Impact on the State of Networking," Stephen
> D.
> >>> Crocker, Shinkuro, Inc., Computer, October 2019.  This was a virtual
> >>> roundtable with Ben Barker, Vint Cerf, Bob Kan, Len Kleinrock and Jeff
> >>> Rulifson.  Ben mentioned the problem with the IMP lights.  It's only a
> >>> small portion of the overall roundtable.  The next two references have
> >>> more
> >>> detail.
> >>>
> >>> [2] "Fixing the lights on the IMPs," an unpublished interview with Ben
> >>> Barker and Scott Brader, 3 July 2020.  It's attached appended.
> >>>
> >>> [3] "Seeking High IMP Reliability in the 1970' ARPAnet" by Walden,
> >>> McKenzie, and Barker, published in Vol 44, No 2 (April - June 2022) of
> >>> IEEE
> >>> Annals of the History of Computing.
> >>>
> >>> ---------- Forwarded message ---------
> >>> From: the keyboard of geoff goodfellow <geoff at iconia.com>
> >>> Date: Mon, Aug 18, 2025 at 8:54 PM
> >>> Subject: Re: [ih] Nit-picking an origin story
> >>> To: Steve Crocker <steve at shinkuro.com>
> >>> Cc: John Day <jeanjour at comcast.net>, Dave Crocker <dhc at dcrocker.net>,
> >>> Internet-history <internet-history at elists.isoc.org>, <
> dcrocker at bbiw.net>
> >>>
> >>>
> >>> [I] am innately curious about the ARPANET "The IMPs Lights Reliability
> >>> Issue" you mention here and wonder if some additional color could be
> >>> elucidated to the colorful story as to just HOW "the lights on the IMP
> >>> panel being a major source of outages" and specifically what
> >>> "re-engineering" was effectuated to ameliorate them from crashing the
> >>> IMPs?
> >>>
> >>> On Mon, Aug 18, 2025 at 7:22 AM Steve Crocker via Internet-history <
> >>> internet-history at elists.isoc.org> wrote:
> >>>
> >>>> ... Ben Barker has a colorful
> >>>> story about the lights on the IMP panel being a major source of
> outages.
> >>>> The IMPs had a 98% percent uptime at first.  98% was astonishingly
> good
> >>>> compared to other machines of the day, but intolerably poor in terms
> of
> >>>> providing an always available service.  Ben re-engineered the lights
> and
> >>>> brought the reliability up to 99.98%.  How's that for a small thing
> >>> having
> >>>> a big effect!
> >>>>
> >>> Fixing the lights on the IMPs
> >>>
> >>>
> >>>
> >>> Below is an exchange with Ben Barker, stimulated by a comment by Scott
> >>> Bradner.  I had been talking to Scott about another project but I
> opened
> >>> by
> >>> asking a bit about his early years.  He worked at Harvard in several
> >>> capacities over many decades.  He started as a programmer in the
> >>> psychology
> >>> department.  Ben Barker had been a student at Harvard and later joined
> >>> BBN.  Ben was a hardware guy.  Scott mentioned Ben hired him to develop
> >>> circuit boards for the front panels of the IMPs.  Ben participated in a
> >>> virtual roundtable last year, and I recall his vivid story of improving
> >>> the
> >>> reliability of the IMPs.
> >>>
> >>>
> >>>
> >>> For context, the following details are alluded to but not explained.
> >>>
> >>>
> >>>
> >>> ·    The first several IMPs were built using ruggedized Honeywell 516
> >>> computers.  I believe the cost to ARPA was $100K each.  Production
> later
> >>> shifted to using regular, i.e. not ruggedized, Honeywell 316
> computers.  I
> >>> believe this dropped the cost to $50K each.  I believe the architecture
> >>> was
> >>> identical but probably slightly slower.  Apparently, speed wasn’t an
> >>> issue,
> >>> so this change saved a noticeable amount of money.  Also, as Ben makes
> >>> clear, there were some unfortunate changes to the front panel that may
> >>> have
> >>> saved some cost in the production but were problematic in operation.
> >>>
> >>> ·    The software in the IMP included the routines for receiving and
> >>> forwarding packets and retransmitting if they hadn’t been received
> >>> correctly.  It also included the distributed algorithm for computing
> the
> >>> routing tables based on periodic exchange of tables with neighboring
> IMPs.
> >>> In addition to these primary functions, there were also four background
> >>> processes that implemented “fake hosts,” i.e. processes that were
> >>> addressable over the network as if they were hosts.  Each one
> implemented
> >>> a
> >>> specific function.  One was DDT, the standard debugging technology of
> the
> >>> day.  I say “standard” because I had encountered other implementations
> of
> >>> DDT while at MIT.  I have no idea whether similar software was used in
> >>> other environments, but the concept was both simple and powerful, and I
> >>> assume it would have been copied widely.  In brief, it’s an interactive
> >>> program that has access to the memory of one or more processes.  There
> are
> >>> commands for starting and stopping the execution of the subject
> process,
> >>> examining or changing memory and setting breakpoints.  AI Memo AIM-147
> by
> >>> Tom Knight in 1968 describes DDT for the MIT AI Lab PDP-6.  An earlier
> >>> 1963
> >>> memo, Recent Improvements in DDT, by D. J. Edwards and M. L. Minsky
> makes
> >>> it clear DDT had been around for several years.
> >>>
> >>>
> >>>
> >>> See my comments after the exchange.
> >>>
> >>>
> >>>
> >>> Date: Jul 3, 2020, 2:37 PM
> >>> From: Steve Crocker <steve at shinkuro.com>
> >>> To: Scott, Ben, me
> >>>
> >>> Ben,
> >>>
> >>>
> >>>
> >>> I was chatting with Scott about his early days.  He mentioned doing the
> >>> circuit boards for the IMP front panels, and I recognized it was part
> of
> >>> the same story you told me about fixing the lights.
> >>>
> >>>
> >>>
> >>> Scott,
> >>>
> >>>
> >>>
> >>> Thanks for your time today.  You mentioned doing the circuit boards for
> >>> the
> >>> IMP front panels.  As I mentioned, I listened to Ben Barker vividly
> >>> describe how this made a big difference in reducing the number of
> service
> >>> calls and improving the uptime of the IMPs.  Ben participated in a
> virtual
> >>> roundtable last year that was published in the IEEE's Computer
> magazine.
> >>> A
> >>> copy is attached.  Ben mentions reliability in both his brief bio and
> >>> later
> >>> on page 22.  I've copied the text from that page.
> >>>
> >>>
> >>> *BARKER: *Reliability surprised me. I was surprised to find that the
> >>> reliability requirements for a network are much more extreme than the
> >>> reliability requirements for a computer, even for the computer upon
> which
> >>> the network switch is built. When we first started operating the
> Arpanet,
> >>> on average each node was down about a half hour a day. And people were
> >>> saying, the net’s always down and there was talk of canceling it,
> shutting
> >>> down the net because it just wasn’t good enough to be useful. We had a
> >>> meeting with our subcontractor for hardware to present our complaint to
> >>> them that nodes were down a half hour a day. Their reaction was,
> “You’re
> >>> down a half hour out of 24. That’s about 2%. You’re getting 98% up
> time.
> >>> We’ve never gotten 98% uptime on anything we’ve built. How are you
> doing
> >>> it?”
> >>>
> >>>
> >>> Eventually, I took over field operations of the network. They thought
> it
> >>> was strange to have a Ph.D. running a field service operation, but, you
> >>> know, we were weird guys. But little by little, we chipped away at it
> over
> >>> the course of a year and a half. We got the availability from 98 to
> >>> 99.98%,
> >>> and the user community reaction went from “the net’s always down” to
> “the
> >>> net’s never down.” But that change is something that would have been
> >>> written into the spec if they were talking about that kind of
> application,
> >>> nuclear survivability.
> >>>
> >>>
> >>> Ben doesn't mention the lights in the article but I definitely remember
> >>> him
> >>> describing this to me.  It might have been in a separate conversation
> that
> >>> wasn't recorded.
> >>>
> >>>
> >>>
> >>> Cheers,
> >>>
> >>>
> >>>
> >>> Steve
> >>>
> >>>
> >>>
> >>> hat
> >>>
> >>>
> >>>
> >>> Date: Fri, Jul 3, 4:01 PM
> >>> From: Ben Barker
> >>> To: me, sob at sobco.com
> >>>
> >>> Indeed!  And hello you old SOB.  How have you been doing the last half
> >>> century?
> >>>
> >>>
> >>>
> >>> Honeywell built the 516s and 316s using low-voltage incandescent bulbs
> for
> >>> the displays.  On the ruggedized 516s, they were in sockets with a
> >>> screw-on
> >>> clouded cover.  Not too bad.  On the 316, the bulbs were mounted inside
> >>> the
> >>> momentary contact rocker switches that were used to input data.
> >>> Unfortunately, these switches were actuated by pressing and releasing
> >>> them,
> >>> allowing the switch to pop back to the resting position.  There was a
> >>> strong spring pushing the switch back out resulting in a pretty strong
> >>> mechanical snap on release.  More unfortunately, this mechanical shock
> was
> >>> simultaneous with the bulb turning on – the inrush of maximum current
> into
> >>> the cold filament – ideal conditions and timing for burning out a bulb.
> >>> More unfortunately yet, the bulbs were mounted in the switch by
> soldering
> >>> the leads to the connector.  This meant that the bulbs burnt out very
> >>> frequently and a dead bulb required taking the IMP down, disassembling
> the
> >>> front panel, unsoldering the dead bulb, and soldering in a new one.
> >>>
> >>>
> >>>
> >>> But wait!  It gets worse!  The IMPs were fragile: once a machine was
> taken
> >>> down, it typically took hours – sometimes days – to get it back up.
> >>>
> >>>
> >>>
> >>> I asked Scott to come up with a design that would replace the switches
> and
> >>> bulbs, using red LED bulbs in their place.  Scott found a switch / LED
> >>> combo that fit just right into the holes in the 316 front panel and
> >>> designed a PC card that carried the switch / lights in just the right
> >>> places to fit in. Scott went into production and we retrofitted them in
> >>> all
> >>> the 316s in the field. Down time dropped amazingly.
> >>>
> >>>
> >>>
> >>> The other half of the strategy was dealing with the fragile IMPs that
> took
> >>> a long time to bring back up.  Scott’s switch / light panel was the
> first
> >>> big step in eliminating probably the majority of the times we had to
> take
> >>> the IMPs down.  The next was stand-up PMs – leaving the machines up and
> >>> running while performing preventative maintenance – mostly cleaning the
> >>> air
> >>> filters and checking and adjusting the power supply voltages and
> >>> performing
> >>> a visual check.  It eliminated one IMP down episode per IMP per month
> and
> >>> helped enormously in eliminating the extended effort to bring the
> machines
> >>> back up.  I have recently been informed that this is a known phenomenon
> >>> known as the “Waddington Effect” from eliminating unnecessary PM on
> world
> >>> war 2 bombers producing a 60% increase in the number of effective
> flying
> >>> hours.
> >>>
> >>>
> >>>
> >>> The third leg of the stool was using self-diagnostic and remote
> diagnostic
> >>> techniques to find problems early on before the network users were
> aware
> >>> of
> >>> a problem and scheduling a tech to go out to replace an
> already-identified
> >>> card, typically off-hours when nobody from that site was using the net.
> >>>
> >>>
> >>>
> >>> Sorry to ramble…
> >>>
> >>>
> >>>
> >>> /b
> >>>
> >>>
> >>>
> >>>
> >>>
> >>> Date: Jul 3, 2020, 4:15 PM
> >>>  From Steve Crocker
> >>> To: Ben, me, sob at sobco.com
> >>>
> >>>
> >>>
> >>>
> >>>
> >>> Very cool stuff.  Question: what sorts of things could be diagnosed
> >>> remotely?
> >>>
> >>>
> >>>
> >>> Steve
> >>>
> >>>
> >>>
> >>> Date: Jul 3, 2020, 6:49 PM
> >>>
> >>> From: Ben Barker
> >>> To: me, sob at sobco.com
> >>>
> >>> [image: https://mail.google.com/mail/u/0/images/cleardot.gif]
> >>>
> >>>
> >>>
> >>>
> >>> Mostly it was figuring out how to find problems that had brought one or
> >>> more IMPs down previously.  One was an IMP that was just running
> slow.  We
> >>> figured out to check a count of how many background loops the machine
> >>> would
> >>> complete in a given length of time – a minute? Easy to do with a
> program
> >>> on
> >>> our PDP-1 reaching out through the IMPs DDT to read the register that
> was
> >>> incremented by the background loop.  If something was keeping the
> machine
> >>> inordinately busy, it would show up as low loop counts.  If it was
> low, we
> >>> would check the return address of the interrupt routines which would
> show
> >>> us what the machine was doing the last time the interrupt happened.
> Then
> >>> there was just debugging using the DDT.
> >>>
> >>>
> >>>
> >>> We had a machine that was confused about time.  It turned out its Real
> >>> Time
> >>> Clock card was bad.  I wrote a PDP-1 routine called RTCHEK that would
> >>> trivially check an IMP’s RTC.
> >>>
> >>>
> >>>
> >>> There was the infamous Harvard crash wherein a memory failure in the
> area
> >>> used by the IMP to store its routing tables. John McQuillan modified
> the
> >>> IMP code to checksum the table before using it.  Told us instantly of
> >>> memory failures in that stack.
> >>>
> >>>
> >>>
> >>> The modem interfaces generated and checked 24-bit checksums on all
> packets
> >>> on the line. We sometimes would get packets that passed the checksum
> check
> >>> but whose contents were in error.  We started having the IMPs send such
> >>> packets to a teletype in Cambridge where we would print them out in
> octal
> >>> and I would examine them.  The most common packets were routing packets
> >>> and
> >>> they were very stylized.  Sometimes a given bit would not always get
> >>> recorded in memory properly and it would be clear which one from
> looking
> >>> at
> >>> the packet.  If it was a problem in the input or output shift
> register, it
> >>> would show up on a given bit and on the bits to its left.  More
> typically
> >>> it was a problem gating the data onto the data bus.  In any case, you
> >>> could
> >>> pretty well identify which gate was failing and schedule out service
> tech
> >>> out to replace that card at night.
> >>>
> >>>
> >>>
> >>> At times, we would patch the IMP to software checksum the packets on
> the
> >>> line to find out if the check registers were failing.  At times we
> would
> >>> turn on the software checksum and turn off the hardware checksum to see
> >>> problems in the AT&T equipment.
> >>>
> >>>
> >>>
> >>> These are random examples.  We did lots of such.  Mostly all done from
> our
> >>> PDP-1 DDT.  It was pretty cool.
> >>>
> >>>
> >>>
> >>> [image: 😊]
> >>>
> >>>
> >>>
> >>> /b
> >>>
> >>>
> >>>
> >>>
> >>>
> >>> Date: Fri, Jul 3, 7:27 PM
> >>> From: Steve Crocker
> >>> To: Ben, Scott
> >>>
> >>>
> >>>
> >>> Cool stuff. Did you guys ever write up these details? A bigger
> question is
> >>> whether the techniques you developed ever influenced or got used by
> >>> others. I
> >>> would imagine that virtually every network operator and router vendor
> >>> needed to develop similar tools.
> >>>
> >>>
> >>>
> >>> Thanks,
> >>>
> >>>
> >>>
> >>> Steve
> >>>
> >>>
> >>>
> >>> Date: Fri, Jul 3, 7:32 PM
> >>> From: Ben Barker
> >>> To: me, sob at sobco.com
> >>>
> >>> 1 – No. This thread is the most detail I know of.
> >>>
> >>>
> >>>
> >>> 2- Not to my knowledge.  I believe that I was told that DECNet later
> >>> incorporated remote diagnostics, but I don’t think they had something
> like
> >>> the remote DDT in the switch that was the basis for most of what we
> did.
> >>> But I am only speculating here.
> >>>
> >>>
> >>>
> >>> Reflective comments
> >>>
> >>>
> >>>
> >>> ·      These details bring a bit of life into the seemingly ordinary
> >>> process of fielding IMPs.
> >>>
> >>> ·      The BBN IMP group was small, smart and agile.  Most were
> software
> >>> or
> >>> hardware engineers who had been at MIT, Harvard and Lincoln Laboratory.
> >>>
> >>> ·      Barker says the improvement from 98% uptime to 99.98%, i.e. a
> >>> reduction of downtime from 2% to 0.02%, which is the hundredfold
> >>> improvement Barker refers to in his bio section of the virtual round
> >>> table,
> >>> made a qualitative difference in the perception within the user
> community
> >>> about the reliability of the network.  This speaks directly to the dual
> >>> nature of the project, i.e. a research project to some like Kleinrock
> and
> >>> Kahn, versus a durable platform for others to build applications upon
> and
> >>> get work done.
> >>>
> >>> To press this point a bit further, there were several time-sharing
> >>> projects
> >>> pursued with IPTO support.  These ranged from Multics at the high end
> down
> >>> to GENIE at Berkeley, Tenex at BBN, ITS at MIT and various others over
> the
> >>> years.  IPTO didn’t put all of its eggs into any single basket.  Some
> of
> >>> these, particularly GENIE on the SDS 940 and Tenex on the PDP-10, were
> >>> adopted by others in the community and became workhorses.  In the
> network
> >>> arena, however, IPTO did not sponsor multiple projects.  Hence, there
> was
> >>> more emphasis for the Arpanet to be usable system and not just one of
> >>> several possible systems.
> >>>
> >>> ·      The learning curve Barker describes is not a surprise.  It’s
> >>> exactly
> >>> what’s to be expected once an initial system is put into operation.
> >>> However, the fact these techniques were not documented and promulgated
> >>> suggests either or both of:
> >>>
> >>> a.     Although the BBN group published multiple papers about their
> work,
> >>> there may have been less publication than there would have been in a
> >>> university.
> >>>
> >>> b.     The remote debugging and other aspects of improving the
> reliability
> >>> might not have seemed special enough to be worth publishing.
> >>> --
> >>> Internet-history mailing list
> >>> Internet-history at elists.isoc.org
> >>> https://elists.isoc.org/mailman/listinfo/internet-history
> >>>
> >>>
> >> --
> >> Geoff.Goodfellow at iconia.com
> >> living as The Truth is True
> >>
> >>
> --
> Internet-history mailing list
> Internet-history at elists.isoc.org
> https://elists.isoc.org/mailman/listinfo/internet-history
> -
> Unsubscribe:
> https://app.smartsheet.com/b/form/9b6ef0621638436ab0a9b23cb0668b0b?The%20list%20to%20be%20unsubscribed%20from=Internet-history
>
>

-- 
Geoff.Goodfellow at iconia.com
living as The Truth is True


More information about the Internet-history mailing list