This seems COFF, not TUHS, and mostly not digital...
I have many 4mm DAT cartridges from 20-30 years ago. Every now and then
I will access one. So far I've yet to see evidence of the media degrading.
On 1/28/2023 4:12 AM, Steve Simon wrote:
> baking old, badly stored magnetic tapes prior to reading them is a common practice.
For the last year+ I have been digitizing selected audio tapes made in
the 70s at AWHQ
(https://en.wikipedia.org/wiki/Armadillo_World_Headquarters) The ones
I've been working with are 1/4" on 10.5" reels. A printed inventory I
was given says "bake" next to almost all of the items, but so far, after
processing roughly 40 reels, I've yet to find one that seemed to need
"baking" (actually, "baking" is a bit overstated, in that best practice
is to raise temperature to roughly 150F --
https://www.radioworld.com/industry/baking-magnetic-recording-tape)
For now, I'm not able to share those AWHQ recordings, but I can share
other recordings I made in the 60s and 70s at
https://technologists.com/60sN70s/. In all those reels, many of which
are cheap, unbranded tape, I didn't find any that seemed to me to need
baking.
Charlie
--
voice: +1.512.784.7526 e-mail: sauer(a)technologists.com
fax: +1.512.346.5240 Web: https://technologists.com/sauer/
Facebook/Google/LinkedIn/Twitter: CharlesHSauer
On 28 Jan 2023 11:29 -0800, from frew(a)ucsb.edu (James Frew):
> As I was leaving the lab late one evening during this mini-crisis, I had to
> walk around a custodian who was busy giving the linoleum floor in the
> hallway its annual deep cleaning / polishing. This involved a dingus with a
> large (~18" diameter) horizontal buffing wheel, atop which sat an enormous
> (like, a cylinder about as big around as a soccer ball) electric motor,
> sparking commutator clearly visible through the vents in the metal housing.
This is probably more COFF than TUHS, but I recall a story from almost
certainly much later where someone (I think it was a secretary; for
now, let's pretend it was) had been told to change backup tapes daily
and set the freshly taken backup aside for safekeeping. Then one day
the storage failed and the backups were needed, only it turned out
when trying to restore the backups that _every_ _single_ _tape_ was
blank. Nobody, least of all the secretary, could explain how that
could have happened, and eventually, the secretary was asked to
demonstrate exactly what had been done every day. Turned out that
while getting the replacement tape, the secretary put the freshly
taken backup tape on a UPS, which apparently generated a strong
magnetic field, before setting that tape aside. So the freshly taken
backup tape was dutifully well and thoroughly erased. Nobody had
mentioned the little detail of not putting the tape near the UPS.
Oops.
--
Michael Kjörling 🏡 https://michael.kjorling.se
“Remember when, on the Internet, nobody cared that you were a dog?”
On 2023-01-27 14:23, Stuff Received wrote:
> On 2023-01-27 12:42, Henry Mensch wrote (in part):
>> I'd like to find solid Android and Windows clients so I could once
>> again use USENET.
>
> I read USENET (at Newsdemon -- USD3 monthly) with Firefox (on MacOS) but
> presumably will also work on Windows.
Oops -- Thunderbird, not Firefox.
>
> N.
>
> (We seem to have strayed into COFF territory...)
> I have yet to look at the oral history things from Corby, etc which may
> answer this in passing.
This page:
https://multicians.org/project-mac.html
links to oral history transcripts from Fano, Corby and Dennis. Only Corby:
https://conservancy.umn.edu/handle/11299/107230http://archive.computerhistory.org/resources/access/text/2013/05/102658041-…
was involved when Bell Labs came on board; he says:
"Also Bell Labs was interested in acquiring a new computer system. They were
quite intrigued and sympathetic to the notions of what we were doing. Ed
David down there was a key figure and an old friend of Fano. They decided to
become partners (pg. 16, CHM interview)
"Simultaneously, Bell Labs had been looking for a new computer acquisition
for their laboratories, and they had been scouting out GE. There was some
synergism: because they knew we were interested they got interested. I think
they first began to look independently of us. But they saw the possibility
of our developing a new operating system together." (pg. 43, CBI interview)
So it sounds like it was kind of a mutual thing, aided by the connection
between Fano and David (who left Bell after Bell bailed on multics).
Noel
> From: Dan Cross
> From Acceptable Name <metta.crawler(a)gmail.com>:
Gmail has decided this machine is a source of spam, and is rejecting all email
from it, and I have yet to sort out what's going on, so someone might want to
forward anything I turn up to this person.
>> Did Bell Labs approach MIT or was it the other way around?
I looked around the Multics site, but all I could find is this: "Bell
Laboratories (BTL) joined the Multics software development effort in November
of 1964"
https://multicians.org/history.html
I did look through IEEE Annals of the History of Computing, vol. 14, no. 2,
listed there, but it's mostly about the roots of CTSS. It does have a footnote
referring to Doug, about the timing, but no detail of how Bell Labs got
involved.
I have yet to look at the oral history things from Corby, etc which may answer
this in passing.
I'll ask Jerry Saltzer if he remembers; he's about the only person left from
MIT who was around at that point.
>> Did participating in Project MAC come from researchers requesting
>> management at Bell Labs/MIT
At MIT, the 'managers' and the researchers were the same people, pretty much.
If you read the panel transcript in V14N2, Fano was the closest thing to a
manager (although he was really a professor) there was at MIT, and he talks
about not wanting to be involved in managing the thing!
Noel
[Bcc: to TUHS as it's not strictly Unix related, but relevant to the
pre-history]
This came from USENET, specifically, alt.os.multics. Since it's
unlikely anyone in a position to answer is going to see it there, I'm
reposting here:
From Acceptable Name <metta.crawler(a)gmail.com>:
>Did Bell Labs approach MIT or was it the other way around?
>Did participating in Project MAC come from researchers requesting
>management at Bell Labs/MIT or did management make the
>decision due to dealing with other managers in each of the two
>organizations? Did it grow out of an informal arrangement into
>a format one?"
These are interesting questions. Perhaps Doug may be in the know?
- Dan C.
(Move to COFF)
> On 22 Jan 2023, at 05:43, Luther Johnson <luther(a)makerlisp.com> wrote:
>
> Yes, I know, but some of that SW development is being automated ... I'm not saying it will totally go away, but the numbers will become smaller, and the number of people who know how to do it will become smaller, and the quality will continue to deteriorate. The number of people who can detect quality problems before the failures they cause, will also get smaller. Not extinct, but endangered, and we are all endangered by the quality problems.
Is it possible that this concern mirrors that of skilled programmers seeing the introduction of high-level languages?
I’ve played with ChatGPT, and the first 10 minutes were a bit confronting. But the remainder of the hour or so I played overcame my initial concerns.
It’s amazing what can be done, especially with Javascript or Python, when you ask for something that’s fairly simple to define, and in a common application area. You can get reasonable code, refine it, ask for an altered approach, etc. It’s probably quicker than writing it yourself, especially if you’re not intimately familiar with the library being used (or even the language).
But … it pretty quickly becomes clear that there’s no semantic understanding of what’s being done behind it. And unless you specify what you want in pretty minute detail, the output is unlikely to be what you want. And, as always, going from a roughed-out implementation of the core functionality to a production-ready program is a lot of work.
In the end, it’s like having an intern with a bit of experience, Stack Overflow, and a decent aptitude driving the keyboard: you still have to break down the spec into small, detailed chunks, and while sometimes they come back with the right thing, more often, you need to walk through it line by line to correct the little mistakes.
I’m looking forward to seeing a generative model combined with static analysis, incremental compilation, unit test output: I think it will be possible for a good programmer to multiply their productivity by a few times (maybe even 10x, but I don’t see 100x). There’ll still be times when it’s simpler to just write the code yourself, rather than trying to rephrase the request.
All of which makes me think of the assembly language to high-level language transition ...
d
COFF'd
I wonder if we'll see events around 2038 that renew interest in conventional computing. There are going to be more public eyes on vintage computers and aging computational infrastructure the closer we get to that date methinks, if even just in the form of Ric Romero-esque curiosity pieces.
Hopefully the cohort of folks that dive into Fortran and Cobol for the first time to pick up some of the slack on bringing 2038-averse software and systems forward will continue to explore around the margins of their newfound skills. I know starting in assembly and C influenced me to then come to understand the bigger picture in which those languages and their paradigms developed, so hopefully the same is true of a general programming community finding itself Fortran-and-Cobol-ish for a time.
- Matt G.
------- Original Message -------
On Saturday, January 21st, 2023 at 10:43 AM, Luther Johnson <luther(a)makerlisp.com> wrote:
> Yes, I know, but some of that SW development is being automated ... I'm
> not saying it will totally go away, but the numbers will become smaller,
> and the number of people who know how to do it will become smaller, and
> the quality will continue to deteriorate. The number of people who can
> detect quality problems before the failures they cause, will also get
> smaller. Not extinct, but endangered, and we are all endangered by the
> quality problems.
>
> On 01/21/2023 11:12 AM, arnold(a)skeeve.com wrote:
>
> > Real computers with keyboards etc won't go away; think about
> > all those servers running the backends of the apps and the
> > databases for the cool stuff on the phones. Someone is still
> > going to have to write those bits.
> >
> > Arnold
> >
> > Luther Johnson luther(a)makerlisp.com wrote:
> >
> > > Well, that's a comforting thought, I hope it goes that way.
> > >
> > > On 01/19/2023 06:10 PM, John Cowan wrote:
> > >
> > > > On Thu, Jan 19, 2023 at 5:23 PM Luther Johnson <luther(a)makerlisp.com
> > > > mailto:luther@makerlisp.com> wrote:
> > > >
> > > > Computers that are not smart phone-like are definitely on the
> > > > endangered
> > > > species list. You know, the kind on a desk, with a keyboard ...
> > > >
> > > > I don't have statistics for this, but I doubt it. Consider amateur
> > > > radio, which has been around for a century now. Amateur stations are
> > > > an ever-shrinking fraction of all transmitters, to say nothing of
> > > > receivers, but in absolute terms there are now more than 2 million
> > > > hams in the world, which is almost certainly more than ever.
"G. Branden Robinson" <g.branden.robinson(a)gmail.com> writes:
[migrating to COFF]
[snip]
>> In that time frame there was a number of microkernel designs. One
>> that has not been mentioned was OS-9 for the 6809/68000 processor. I
>> used it pretty extensively. OS-9 was very unix like from the userland
>> POV, when you consider something like V5 unix, however it didn't share
>> any of the same command names, just many of the same concepts.
>
> This is emphatically true. I used this system as a kid on a 64KiB
> machine, and I don't remember even a mention of Unix in the doorstop of
> a manual by Dale Puckett and Peter Dibble (who gave you something like 6
> chapters of architectural background before introducing the shell
> prompt). Maybe they did mention Unix , but since it had no meaning to
> me at the time, it didn't sink in. I think it is also possible they
> avoided any names that they thought might draw legal ire from AT&T.
That is more or less me too... However, in later years when I was
familiar with Unix I looked at some of the OS-9 books and the block
diagrams in the books about the OS-9 OS could have described early Unix
pretty well.
>> It was close enough that if you had the C compiler, a very basic K&R
>> compiler, you could get some of the unix command to compile without
>> too much trouble.
>
> Years later I went to college, landed on Sun IPC workstations, and
> quickly recognized OS-9's "T/S Edit" as a vi clone, and its "T/S Word"
> as a version of nroff. There was also a "T/S Spell" product but I don't
> recall it clearly enough to venture whether it was a clone of ispell.
Ya, I think I even had a patch that turned T/S Edit into a much closer
vi clone. But, I think by then I had another vi clone already on hand.
One of the other things I did with OS-9/6809 was worked on UUCP. I
didn't write the original OS-9 UUCP code, but I did modify it quite a
bit and had it talking UUCP g protocol to UUnet via a phone line. I did
write a 'rn' Usenet news reader clone and was pulling down a few news
groups as well as email every day. In the last days of that system, I
also logged into the system via a serial port complete with Username and
password prompts. This was all on a Color Computer 3 with 512K.
[snip]
>> and nothing like Mach or even Minix.
>
> With the source of all three available, a technical paper analyzing and
> contrasting them would be a worthwhile thing to have. (It's unclear to
> me if even a historical version of QNX is available for study.)
The source to OS-9/6809 would have been released by Microware a long
time ago had it not been for a particular person in the user community.
Got mucked up. I fell out of following it after the BSD Unixs became
available.
>> It was also very much positioned to real time OS needs of the time and
>> was not really marketed generally and unless you happened to have a
>> Color Computer from Radio Shack
>
> Lucky me! How I yearned for a 128KiB Color Computer 3 so I could
> upgrade to OS-9 Level 2 and the windowing system. (512KiB was
> preferred, but there had been a spike in RAM prices right about the time
> the machine was released. Not that greater market success would have
> kept Tandy from under-promoting and eventually killing the machine.[1])
Level II was nice. It was able to use bank switching and would allow a
set of random 8k memory blocks out of the 128k or 512k present in the
CC3 system to be mapped into the 6809 64k address space. The Color
Computer didn't support memory protection, so no paging or any real
process protection, but this banking allowed for a lot of possibilities.
I know that there was other OS-9 systems around that ran Level II but I
don't really know how they managed memory. I would suspect it to be
simular to the CC3, but that is just a guess on my part.
[snip]
> Regards,
> Branden
>
> [1] Here's a story you may have to sit down for from Frank Durda IV (now
> deceased) about how the same company knifed their m68k-based
> line--which ran XENIX--in the gut repeatedly. It's hard to find
> this story via Web search so I've made a Facebook post
> temporarily(?) public. I'd simply include it, but it's pretty long.
>
> https://www.facebook.com/g.branden.robinson/posts/pfbid0F8MrvauQ6KPQ1tytme9…
>
> [2] https://www.cnn.com/2000/TECH/computing/03/21/os9.suit.idg/index.html
> [3] https://appleinsider.com/articles/10/06/08/cisco_licenses_ios_name_to_apple…
> [4] https://sourceforge.net/projects/nitros9/
--
Brad Spencer - brad(a)anduin.eldar.org - KC8VKS - http://anduin.eldar.org
> On Jan 2, 2023, at 1:36 PM, Larry McVoy <lm(a)mcvoy.com> wrote:
>
> The /bin/sh stuff made me think of an interview question I had for engineers,
> that a surprisingly few could pass:
>
> "Tell me about something you wrote that was entirely you, the docs, the
> tests, the source, the installer, everything. It doesn't have to be a
> big thing, but it has to have been successfully used by at least 10
> people who had no contact with you (other than to say thanks)."
>
> Most people fail this. I think the people who pass might look
> positively on the v7 sh stuff. But who knows?
Huh. That is a surprisingly tricky question, depending on how you want to construe "entirely you".
v1 of https://atariage.com/software_page.php?SoftwareLabelID=2023 (before Thomas Jentzsch optimized the display engine) was ... stuff I did, but obviously neither the idea nor the execution was all that original, since I used Greg Troutman's Dark Mage source, which in turn was derived from Stellar Track.
There's a certain very large text adventure I once did, which I would certainly not bring up at a real job interview since it's riotously pornographic, but it is 200,000 words of source text, got surprisingly good reviews from many people (Emily Short loved it; Jimmy Maher hated it), and I put it all together myself, but the whole thing is a hodgepodge of T.S. Eliot and The Aeneid and then a few dozen other smaller sources, all tossed in a blender. Not going to directly link it but it's not hard to find with a little Googling. The arrangement is original, sure, but its charm--such as it is--may be that it is in some ways a love letter to early D&D and its "what if Gandalf and Conan teamed up to fight Cthulhu" sort of ethos. (Jimmy Maher found the intertextuality very dense and unappetizing, whereas Emily Short really enjoyed the playfulness.)
There's https://github.com/athornton/uCA which fits the criteria but really is a very small wrapper around OpenSSL to automate SAN generation, which is a huge PITA with plain old OpenSSL. Now, of course, you wouldn't bother with this, you'd just use Let's Encrypt, but that wasn't a thing yet. Such as it is it's all me but it is entirely useless without a functional OpenSSL under it.
I'm not sure that ten other people ever used https://github.com/athornton/nerdle-solver because there may have been fewer than ten people other than me that found Nerdle all that fascinating. It was fun talking with that community and finding out that the other solver I'm aware of was completely lexical, rather than actually doing the math. But again: it's a thing that makes no sense without someone else having invented Nerdle first.
Or there's https://github.com/athornton/tmenu; probably also not actually used by ten other people, but it's the front-end of https://mvsevm.fsf.net (which certainly has been enjoyed by...uh...let's go with "at least a dozen" people). It's original work, insofar as it goes, but it (like uCA) is really just glue between other things: a web server front end, a Javascript terminal emulator, and telnet/tn3270 clients.
Which of these, if any, do you count?