[Info-vax] BL870c shared memory?
Kerry Main
kemain.nospam at gmail.com
Mon Oct 17 11:22:17 EDT 2016
> -----Original Message-----
> From: Info-vax [mailto:info-vax-bounces at rbnsn.com] On Behalf
> Of Stephen Hoffman via Info-vax
> Sent: 17-Oct-16 9:24 AM
> To: info-vax at rbnsn.com
> Cc: Stephen Hoffman <seaohveh at hoffmanlabs.invalid>
> Subject: Re: [Info-vax] BL870c shared memory?
>
> On 2016-10-16 13:58:57 +0000, Snowshoe said:
>
> > Is there a way for a BL870c or a BL890c, divided into two (or
> more)
> > systems, to have special memory set aside that can be shared
> between
> > them, while most of the memory is per-system specific as
> usual? If
> > so, how to configure them to do so? Kind of like a Alpha Galaxy
> feature.
> >
> > Does VMS support this, if it is possible?
>
> Nope.
>
> Galaxy (what HP/HPE later called vPar) is a feature dependent on
> the console firmware and is not available on Itanium, nor would I
> expect it
> on x86-64. EFI and UEFI do not provide the necessary
> mechanisms, and
> would require custom work to support differential configuration
> presentations; Galaxy.
> http://labs.hoffmanlabs.com/node/813
> http://h41379.www4.hpe.com/availability/galaxy.html
>
While the dependency on console FW is certainly true for previous flavors of Galaxy, we should also mention that HW virtualization was not part of the previous Alpha/IA64 architectures.
HW virtualization is part of the X86-64 architecture standard.
http://www.cs.columbia.edu/~cdall/candidacy/pdf/Uhlig2005.pdf
Will this mean a future version of Galaxy (post V9?) is in the works via VSI?
I have no idea, but sure hope so.
To be clear, let's remember what Galaxy provides with specific flavors of Alpha HW:
- dynamic cpu sharing between different OS instances (VMware does not do this) based on workloads via customizable business rules or drag-n-drop. Huge capability for those who want to maximize server utilization based on varying workloads;
- shared memory access for very high speed, low latency TCPIP, Cluster DLM and inter-process communications between different OS instances;
- RAM disk in shared memory;
- NUMA (RAD) aware (remember all X86-64 arch's are now NUMA based)
- GUI configurator;
- improved DR capabilities;
- custom API's to write local Galaxy resource management services;
> As for shared or remote memory access into other servers...
>
> Memory Channel was an Alpha feature providing reflective
> memory, and the hardware involved didn't sell in large volumes.
> AFAIK, there is no Itanium support for Memory Channel.
> https://people.eecs.berkeley.edu/~culler/cs258-
> s99/project/memchannel/memchannel.pdf
>
Technology standards are changing rapidly.
Check this link out: (posted Oct 17, 2016)
http://bit.ly/2enpgNF
"Opening Up The Server Bus For Coherent Acceleration"
>
> VSI hasn't gotten around to adding support for RDMA adapters or
> analogous.
>
To clarify - the difference between RDMA (original spec) and RDMA V2 (RoCEv2 is current spec) is huge. The new spec allows a much simpler migration from current drivers.
The current technology that has already been adopted by Windows/Linux and available via very high speed, very low latency cards from companies like Mellanox is based RoCEv2.
> Most folks didn't head toward Galaxy-style processor and
> memory
> sharing. Console-level virtualization didn't (hasn't?) caught on.
By those who are still promoting distributed computing.
The "enterprise" world is rapidly changing to fewer, much larger systems with TB scale memories, very high speed, low latency interconnects.
These types of system require much better levels of workload management and server utilization techniques than old technologies like VMware which has minimal workload mgmt.
> For most folks, virtualization can happen at the system hardware
> level — this is what most virtual machines present, pretending to
> be a descendent of the 1981-vintage box that came from Boca
> Raton, or a para-virtualization of that box — or happens at the
> software and particularly at the operating system level — and this
> is what containers provide, particularly when sandboxing and
> related are available to keep the apps from accidentally or
> intentionally stomping
> on other apps.
I am not convinced containers are not the answer for most enterprise deployments.
http://bit.ly/2dZ9c6C
http://bit.ly/2ebdjwS
In addition, my understanding, but willing to be corrected, is that sandbox's are basically processes with TMPMBX only priv's and perhaps some additional monitoring/auditing like what one can already do with products like System Detective from PointSecure .
> VSI has stated that they will be providing
> OpenVMS
> support for both native x86-64 boot and for x86-64 virtualization
> in specific virtual machines, and that they're pondering adding
> support for host-level virtualization — containers — as part of
> some more distant future work.
>
Again, I really do not see any value with industry containers in the OpenVMS environment.
What is the issue containers are trying to address?
Containers on Linux/Windows are making headway because these platforms recognize VM sprawl is a huge issue and they do not have good App stacking capabilities like what OpenVMS Cust's have been doing for decades. Running 15+ business apps from different groups or BU's on 1 system or cluster is considered normal in most OpenVMS environments.
See links above for security and network issues with containers.
Imho, enhancing native things like LD technologies (boot support?), additional security features, class scheduler, batch / workload management ect. would be far more beneficial than trying to resolve issues that do not apply to OpenVMS.
> Maybe some hardware vendor that's pondering qualifying
> OpenVMS x86-64 support on their iron might decide to create a
> customized UEFI with
> customizable ACPI reports? But then that's ~2020-vintage
> discussion
> and roll-out, at the earliest. There's also that UEFU and ACPI are
> not exactly the most user-friendly bits in an Integrity or x86-64
> box, and adding Galaxy support atop that could well cause UI
> nausea.
>
Imho, VSI will need to address the "Why OpenVMS" as feature differentiators on X86-64 over Windows/Linux.
Besides just being a more stable platform, Galaxy-like virtualization shared resource features, clustering (improved scalability of shared disk/everything), improved server utilization and workload management are going to be critical components of "Why OpenVMS" in the V9/V10+ era on X86-64 and other future HW platforms.
(yes, the OpenVMS license model and marketing needs to improve with V9 + as well)
Regards,
Kerry Main
Kerry dot main at starkgaming dot com
More information about the Info-vax
mailing list