[ih] Saving IETF history

Toerless Eckert tte at cs.fau.de
Thu May 13 11:51:01 PDT 2021


Hah. Your last paragraph was the interesting aspect i completely ignored "Jury trial" ;-)

Cheers
    Toerless

On Wed, May 12, 2021 at 09:50:50PM -0700, Jack Haverty wrote:
> Well, it's complicated.  Real answer is "probably not".  The actual
> claim is written in Patentese, and translating it into "download new
> versions of system software" was not straightforward. 
> 
> FYI, here's the actual claim:
> 
> =====
> 
> 12. A reprogrammable system comprising: 
> 
>   a digital detector for 
> receiving information of a transmission and detecting digital signals in
>  said transmission, said digital signals including new operating 
> instructions; 
> 
>   a processor operatively connected to said digital
>  detector for receiving and processing information of some of said 
> digital signals, said processor identifying those of said operating 
> instructions addressed to said processor, said processor instructing 
> said detector to detect and pass specified signals; 
> 
>   a memory 
> device operatively connected to said processor for holding operating 
> instructions addressed to said processor, said operating instructions 
> controlling the operation of said processor; and 
> 
>   said processor
>  loading said operating instructions that are addressed to said 
> processor into said memory device to thereby reprogram said processor, 
> said operating instructions including instructions to cause said 
> processor to cause said detector to detect different signals.
> 
> =====
> 
> That's probably not what you would consider a description of what an IMP
> does.   It was a bit of a leap to go from that description to the IMP
> software, and unlikely that any old documentation would make the argument.
> 
> Every word matters, and every word is a subject of argument, sometimes
> for years.
> E.g., what is "memory"?  Is it just RAM?  If a device has any ROM, does
> that mean it doesn't qualify since it can't be reprogrammed?   What is
> the difference between "programming" and "reprogramming"?
> 
> Also, a jury trial was expected.  Imagine 12 people, plus a gaggle of
> lawyers, and a judge, none of whom understand much about computers at
> all.  Yet they make the decisions.   That's why a demo was a goal - it's
> much easier to show something happening than to convince someone that
> some old document proves it happened long ago.
> 
> /Jack
> 
> 
> 
> On 5/12/21 8:21 PM, Toerless Eckert wrote:
> > Thanks for the story, Jack.
> >
> > Given how much of a believer in good public documentation i am, however curious:
> >
> > Was the functionality in question well enough publically documented with
> > according early dates ? I suspect not befcause i would be surprised if the documentation
> > would not have been good enough, if it existed. After all, most patents are also graanted
> > without evidence that they work, so its patenting of concept, not evidence
> > thereof (which might have been different in decades before my time though..).
> >
> > Cheers
> >     Toerless
> >
> > On Wed, May 12, 2021 at 05:44:19PM -0700, Jack Haverty via Internet-history wrote:
> >> In 2012-2013, I got involved as an Expert Witness in one of those huge
> >> patent lawsuits, which had by then been going through the legal process
> >> for about 30 years (!!).  One of the lessons learned was that although
> >> documentation is useful, the most valuable evidence is physical. 
> >>
> >> In this case, I noticed (and amazingly remembered) that the ARPANET IMP
> >> program was a viable example of relevant "prior art", which is the holy
> >> grail in patent battles.   The ARPANET did what the patent covered, and
> >> did it well before the patent was filed.  But proving that such an idea
> >> was actually implemented and used almost 40 years earlier was the
> >> necessary task.
> >>
> >> A huge effort ensued, driven by the lawyers and their clients, trying to
> >> find an old IMP and make it operable again, so that the "prior art"
> >> behavior could be demonstrated at trial before a jury.   No actual IMP
> >> was unearthed (at least one that could still operate), but an early-70s
> >> listing of the IMP software had survived in a basement.   After a lot
> >> more effort, that old software was brought back to life, running on an
> >> emulator of the old Honeywell 316 computer that was used as an IMP.
> >>
> >> The particular behavior of the IMP that demonstrated the prior art was
> >> apparently never described in the documentation.   It only could be
> >> proven by looking at the code itself.  The old saying "The documentation
> >> is in the code" was exactly correct.
> >>
> >> Bottom line - don't just think about documents, which can be
> >> insufficient to prove that something actually existed.   Software, and
> >> even hardware, is even better, at least for some patent fights.   It
> >> captures parts of history that never got into words.
> >>
> >> For the curious, there's a good description of the "ARPANET
> >> Resurrection" here:
> >>
> >> https://walden-family.com/impcode/imp-code.pdf
> >>
> >> Look around page 33.   Also see https://walden-family.com/impcode/
> >>
> >> /Jack Haverty
> >>
> >>
> >> On 5/12/21 4:22 PM, Toerless Eckert via Internet-history wrote:
> >>> To put Karls conerns into a maybe easier understood (but theoretical) example for those on the list that have not been involved in practical instances of the problem:
> >>>
> >>> - printed public/user product documentation from 2000 gets thrown out 15 years later because
> >>>   of "we need to get rid of all this old junk", maybe because of refurnishing offices.
> >>> - half a year later, a lawsuit with such a "bogus" patent that was filed in 2002 ensues.
> >>> - Obviously, the 2000 public/user product documentation would exactly show the patent
> >>>   claim to be "bogus" because the public documentation from 2000 explains exactly the same
> >>>   thing the patent filed in 2002 claimed to be novel.
> >>> - Online web page of the prior art product of course did not keep old version information reaching
> >>>   that far back, and even if it would have, it would not have date information on it, but only
> >>>   version numbers.
> >>>
> >>> These type of things easily happen in multi-million dollar lawsuits over and over.
> >>>
> >>> Going forwarding, IMHO, the best solution for e.g.: IETF documentation would be:
> >>>
> >>> a) have all data such as all of datatracker and IETF mailing list archive in an easy mirrored access form,
> >>>    which i think we do not have, at least i have not found it, only for some subset of our data.
> >>>
> >>> b) Have multiple, independent of each other mirrors around the world that would create
> >>>    signed/dated certificates for the hashes of each mirrored document - and keep old
> >>>    (versions of) documents and their signatures even when they would be deleted/changed on the origin site.
> >>>
> >>>    Maybe those mirrors cost money, but IMHO worth it. especially for stuff like IETF whose overall
> >>>    volume on disk is laughable small. And this becam standard tooling, folks like CHM should be
> >>>    ideal places for such mirroring.
> >>>
> >>> Without something equivalent to a/b i fear it is way too easy to create fake evidence for anything,
> >>> and the "evidence" may not hold up as well court as  the "good old printed evidence".
> >>>
> >>> This "creation time" tracking in a more trustworthy fashion will of course not
> >>> work retroactively, which is why it would be even more important to understand the value of
> >>> doing this now, so someone starts doing it for the benefit of future bogus lawsuits for
> >>> stuff we start working on now. Especially given how paper already has disappeared as more
> >>> reliable evidence.
> >>>
> >>> Cheers
> >>>     Toerless
> >>>
> >>> On Wed, May 12, 2021 at 03:12:43PM -0700, Karl Auerbach via Internet-history wrote:
> >>>> I have also been highly concerned about the tendency of modern tech history
> >>>> to erase its own records.
> >>>>
> >>>> My concern may, however, be in a different direction.
> >>>>
> >>>> I am concerned about the growth of specious patents.  There are a lot of
> >>>> patent trolls out there who buy-up weak patents that got past the relatively
> >>>> lax patent examiners in the US and elsewere, examiners who often have no
> >>>> notion of ideas in networking or computer systems, whether embodied in
> >>>> software or hardware.
> >>>>
> >>>> By erasing our past we make it difficult to rebut these bad patents - we
> >>>> have discarded the evidence that the claims of those patents are neither
> >>>> novel nor non-obvious.
> >>>>
> >>>> I think that over the last few years the IETF has done a spectacular job of
> >>>> organizing and tracking the RFC series.
> >>>>
> >>>> However, we still have a tendency to forget the old when the newer, shinier
> >>>> thing comes along.
> >>>>
> >>>> We should strive to make sure that our past is recorded.  And we ought to
> >>>> consider legal evidentiary requirements so that one who is challenging
> >>>> specious patents is not blocked by the complexities of the rules of
> >>>> evidence.
> >>>>
> >>>> 	--karl--
> >>>>
> >>>>
> >>>> On 5/9/21 1:23 AM, John Gilmore via Internet-history wrote:
> >>>>> Dave Crocker wrote:
> >>>>>> Saving the RFCs is obvious.  What appears to be less obvious and, IMO,
> >>>>>> is just as important in historical terms, is /all/ of the IETF-related
> >>>>>> work materials.  Drafts.  Mailing list archives.  Session notes.
> >>>>>> Everything.
> >>>>> John Day wrote:
> >>>>>> Agreed. It should go somewhere. Same for all of the other standards
> >>>>>> groups, forums, consortia, etc.
> >>>>> Re the IETF, look in:
> >>>>>
> >>>>>    https://archive-it.org/collections/11034
> >>>>>
> >>>>> A few years ago, I set up an Archive-It.org job to monitor the IETF's
> >>>>> web presence.  I was disturbed at the deliberate ephemerality of the
> >>>>> Internet-Draft ecosystem.  I had been looking back at a 10-year-old
> >>>>> effort to eliminate some ridiculous restrictions on the IPv4 address
> >>>>> space, and IETF had thrown away most of the relevant documents (though I
> >>>>> found copies elsewhere once I knew their names).
> >>>>>
> >>>>> Archive-It is a service of the nonprofit Internet Archive (archive.org).
> >>>>> So, the Internet Archive's robots are now crawling (various parts of)
> >>>>> the IETF websites every week, month, and quarter, under my direction.
> >>>>> And saving the results forever, or as long as the Internet Archive and
> >>>>> the Wayback Machine exist.  Between 1998 and now it's pulled in about
> >>>>> 1.8 TB of documents, which are accessible and searchable either from the
> >>>>> above URL, or from the main Wayback Machine at web.archive.org.
> >>>>>
> >>>>> The IETF websites aren't organized for archiving.  I frankly don't
> >>>>> understand their structure, so am probably missing some important
> >>>>> things, and overcollecting other things.  But at least I tried.
> >>>>> Suggestions are welcome.
> >>>>>
> >>>>> Just be glad the IETF is copying-friendly.  Imagine trying to archive
> >>>>> the IEEE or OSI standards development process.  Then imagine big
> >>>>> copyright lawsuits from self-serving people who tied their income
> >>>>> stream to restricting who can access the standards and the
> >>>>> standardization process.
> >>>>>
> >>>>> 	John
> >>>>>
> >>>>> PS: Anyone or any institution can get an Archive-It account for roughly
> >>>>> $10K/year.  The service automates the collecting of *anything* you want
> >>>>> from the web for posterity.  (If you want them to, the Internet Archive
> >>>>> will also write copies of it on new hard drives and send them to you for
> >>>>> your own archival collection.)  About 800 institutions are customers today.
> >>>>> You can also get a low-support low-volume Archive-It Basic account for
> >>>>> $500/year.  Or get custom Digital Preservation services to improve the
> >>>>> likelihood that your own curated digital assets will survive into the
> >>>>> distant future.  See https://Archive-It.org .
> >>>>>
> >>>>> PPS: The Internet Archive's long term survival is, of course, not
> >>>>> guaranteed.  In particular, it will go through a tough transition when
> >>>>> its founder eventually dies.  What is guaranteed is that they have built
> >>>>> a corpus of useful information: Millions of books, billions of web
> >>>>> pages, hundreds of thousands of concerts, decades of saved television
> >>>>> channels, etc.  They are absorbing a lot of archival microfilm, too,
> >>>>> including genealogical and census records, magazines, etc.  This corpus
> >>>>> will likely motivate people to preserve and replicate it into being
> >>>>> useful in the distant future.  They have tried to design the technical
> >>>>> storage to encourage that result.  Does anyone here know anybody who has
> >>>>> both the money and the motivation to make a complete and ongoing copy in
> >>>>> a separately administered, separately owned organization?  That would
> >>>>> significantly mitigate the long term risk of having all the replicated
> >>>>> copies of the corpus owned by a single US nonprofit.  It would probably
> >>>>> take a bare minimum staff of 10 people to run and manage such an
> >>>>> operation, with dozens of petabytes of rotating storage in multiple data
> >>>>> centers and a large collection of (mostly free) software keeping it all
> >>>>> organized and accessible.
> >>>>>
> >>>> -- 
> >>>> Internet-history mailing list
> >>>> Internet-history at elists.isoc.org
> >>>> https://elists.isoc.org/mailman/listinfo/internet-history
> >> -- 
> >> Internet-history mailing list
> >> Internet-history at elists.isoc.org
> >> https://elists.isoc.org/mailman/listinfo/internet-history
> 

-- 
---
tte at cs.fau.de



More information about the Internet-history mailing list