[Info-vax] OpenVMS, and Memory and Storage Addressing (was: Re: VMS Software needs to port VAX DIBOL to OpenVMS X86 platform
Stephen Hoffman
seaohveh at hoffmanlabs.invalid
Sat Dec 26 16:21:20 EST 2020
On 2020-12-26 18:00:03 +0000, <kemain.nospam at gmail.com> said:
>> -----Original Message-----
>> From: Info-vax <info-vax-bounces at rbnsn.com> On Behalf Of Arne Vajhøj
>> via Info-vax
>> Sent: December-24-20 2:38 PM
>>
>> So they [IBM POWER] can do 2 PB and x86-64 can only do 256 TB. ...
>>
>
> The market is changing very rapidly.
>
> Seagate and Western Digital now offer 18TB disks. (google "18TB disk")
>
> Large enterprise class VMware servers hosting large numbers of VM's are
> configured today with 1-2TB memory per server.
>
> How many would have predicted these capabilities even 2 or 3 years ago?
>
> Now, fast forward 5 years - its not hard to see what appear to be far
> away limits being pushed.
Currently-available Intel x86-64 processors support 57-bit physical
addressing. That permits 128 PiB of physical memory.
https://software.intel.com/sites/default/files/managed/2b/80/5-level_paging_white_paper.pdf
Linux added support for that, and likely too most other virtual
machines either have or are adding 57-bit support. I expect that VSI is
aware of Intel 57-bit physical addressing support, as well.
Interestingly for the "just add more memory" approach, particularly as
the I/O speeds increase, the need for larger physical memory can be
lessened—benchmarks from one recent commercially-available Arm-based
system are showing that less can RAM work well, when the memory and I/O
paths are reworked. Big memory is a good cache when main storage access
is HDD-speed glacial. When main storage is NVMe speeds, paging and
swapping can (again) be a viable trade-off. And with byte-addressable
non-volatile storage becoming available, these same server design
trade-offs trade-offs shift again.
It'll be a while before we're using 18 TB HDDs with OpenVMS, though.
That's all dependent on the 64-bit I/O system updates (queued for the
unreleased V8.5, prolly arriving in production at V9.2), on a new file
system, and/or on app updates.
And as for big memory and big storage and OpenVMS as most will probably
see it, the VM can present a four-level page table to the OpenVMS
guest, as well as the 2 TiB and smaller disks expected by existing apps
and the ODS-2 and ODS-5 file systems. Those folks that need or want
native boot will be picking their configurations to operate within
whatever limits are then applicable to OpenVMS—or finding a person or a
vendor to configure and test their servers for them.
Most (all?) vendors with 64-bit are running a subset of bits for
physical addressing and the unused or unimplemented address bits all 0
or all 1 bits, which means the servers will be available as the demand
warrants. And I can do pretty well with SSDs and less than 48 bits of
memory for the apps that I'm dealing with and know about, and NVMe
storage will speed that. Folks that are stuck by 48- or 50-bit
addressing, or planning for OpenVMS supercomputer-scale configurations,
have a chat with VSI. And have a look at your I/O speeds and feeds and
related app designs, too.
--
Pure Personal Opinion | HoffmanLabs LLC
More information about the Info-vax
mailing list