[Info-vax] VMS QuickSpecs
Stephen Hoffman
seaohveh at hoffmanlabs.invalid
Thu Aug 13 15:48:27 EDT 2015
On 2015-08-13 15:41:17 +0000, Dirk Munk said:
> The Quickspecs don't mention SCS (used to be called LAVC?)
> specifically. For clarity, the ethernet part could be enhanced by
> specifically mentioning SCS and IP.
That's nothing new. The HP OpenVMS SPD doesn't mention SCS, either:
<http://h18000.www1.hp.com/info/SP4229/SP4229PF.PDF>
It's arguably below the level of what's covered in the SPD or QuckSpecs.
SCS is System Communications Services, it's the underpinnings of
clustering. Among other supported paths, clustering has operated via
CI, NI (Network Interconnect; Ethernet) and DSSI as a communications
interconnect, and operates via SCSI as a storage interconnect.
LAVc Local Area VAXcluster support introduced the VAXport Port Emulator
driver (PEDRIVER), which allowed clustering over Ethernet. "VAXport
drivers: In a VAXcluster environment, device drivers that control the
communication paths between local and remote ports. (Examples are
PADRIVER for the CI, PEDRIVER for the LAN, and PIDRIVER for the DSSI.)"
> Fibrechannel clustering would be nice too, after all Fibrechannel is in
> a way comparible with ethernet, and it can also be used for IP.
That likely won't happen. There used to be — haven't looked recently —
too much variation among the different FC HBA interfaces to make
supporting that approach be reasonably efficient.
More recently, FCIP has become one of the typical approaches for folks
using FC — so you can tunnel FC over an IP network underneath, and
since you have an IP network, then there's no need for IP over FC.
(FCIP follows the same reasoning behind IPCI, too: consolidation onto
IP infrastructure.)
Toss into this discussion the arrival of server boards with embedded 10
GbE NICs, of the availability of 10 GbE and 40 Gbps NICs, or of the
recently-approved 100 Gbps 100GBASE stuff, why bother with FC? If
you're truly going high-end here, then that's usually InfiniBand.
FWIW, SuperMicro is selling a PCIe board that does both 40 Gbps
InfiniBand and 10 GbE, and also selling 1U server boxes with dual 40
GbE NICs.
http://www.supermicro.com/manuals/other/datasheet-AOC-UIBQ-M2.pdf
http://www.supermicro.com/white_paper/white_paper_ultra_servers.pdf
ps: there's also the question of whether going replicated host-to-host
starts to makes sense over server to shared storage controller, for
some environments. One HGST prototype recently showed 2.5 microsecond
latency for server to remote server to remote NVMe SSD storage access
across InfiniBand, after all. For other "more traditional"
configurations, use fast host-to-host for most stuff and use FC or
10GbE, 40 GbE or 100 GbE (as it arrives) iSCSI as the I/O bus.
> In fact it was used by IP before it was used by SCSI (for storage SANs).
Not on VMS it wasn't. SCSI is also not usable for host-to-host
connections on OpenVMS, either.
Now where VSI might go with any of this, we shall learn. But it
wouldn't surprise me to see little more than incremental new hardware
support on OpenVMS I64 — comparatively little beyond getting Poulson
and Kittson servers to boot — and dedicating most of the available
engineering and hardware efforts to the x86-64 port and to the hardware
and buses that are available in the target x86-64 environment.
--
Pure Personal Opinion | HoffmanLabs LLC
More information about the Info-vax
mailing list