[Info-vax] queue errors

Stephen Hoffman seaohveh at hoffmanlabs.invalid
Wed Oct 17 10:44:30 EDT 2012


On 2012-10-17 12:58:57 +0000, Tom Adams said:

> The files are log files from a system that don't make it to the
> logger.  Not printed by an end-user.

You're probably not going to go for it, but syslog or syslogng or 
analogous would be more typical in recent application designs and 
updates; centralized logging.

No, syslog and syslogng are not part of TCP/IP Services (and probably 
never will be), but open-source versions of the necessary clients are 
available.

> But we do have it all in a disk log that loses nothing.  So the losses
> are not real important, but they have gotten on a todo list prepared
> by QA.
> 
> I am looking into this because the QA people tend to focus on the
> paper log. A bit hidebound and not totally rational I know.

I might wonder anout the organizational focus and funding here.  That 
this "minor issue" of lost files is near the top of the to-do list, yet 
efforts such as ensuring that comparatively stringent compiler 
diagnostics (/CHECK, et al), better and centralized logging, 
diagnostics from failing processes ("because of legacy code issues and 
the fact it might crash some detached processes for unimportant 
reasons"), and the maintenance of current VMS versions and patches, 
aren't.

Legacy code and down-revision versions can be and often are excellent 
sources of latent and subtle bugs.

In aggregate, this can be inferred to mean the organization's QA 
efforts might benefit from a somewhat broader view; toward the options 
and alternatives and trade-offs that are available, and where the bugs 
are located in the current application code base.

I'd wonder what's getting missed, too.  A lack of encryption on 
critical data is a fairly common omission in these legacy-code 
environments, for instance.

> If I had the files, I could tell exactly what is missing from the
> paper log with little effort.

Try the ANALYZE /DISK /REPAIR command, and then go look in [SYSLOST]

And look at the code that's generating this logging data; that code 
contains either design bugs, or run-time error-handling bugs, or 
(likely) both.  It's assuming that the printers always work, and that's 
never been the case.  (And this also suggests use of capabilities such 
as syslog / syslogng or some other form of centralized logging.)

> As things stand, we go to the disk log if someone notices that
> something seems to be missing from the paper log.  That has only
> happened once in more than a decade.

One case that you know of.

DECserver and network devices and printers can all get flaky.

Old computers and old code can sometimes be wonderfully pernicious, too.

And FWIW, design bugs and error-handling bugs can be latent from the 
first line of code ever included into an application; for "more than a 
decade".  I've encountered day-one bugs in central parts of VMS, and 
there have been documented cases of bugs latent from twenty and thirty 
years and more.

More succinctly, "runs" doesn't mean "correct", nor "bug free".

-- 
Pure Personal Opinion | HoffmanLabs LLC




More information about the Info-vax mailing list