HP OpenVMS Systems Documentation
HP OpenVMS Cluster Systems
Order Number: BA554-90021
This manual describes cluster concepts, procedures, and guidelines for configuring and managing OpenVMS Cluster systems. Except where noted, the procedures and guidelines apply equally to Integrity servers and Alpha systems. This manual also includes information for providing high availability, building-block growth, and unified system management across coupled systems.
Revision/Update Information: This manual supersedes HP OpenVMS Cluster Systems, OpenVMS Alpha Version 7.3--1 and OpenVMS VAX Version 7.3.
OpenVMS Version 8.4 for Integrity servers
© Copyright 2010 Hewlett-Packard Development Company, L.P.
Confidential computer software. Valid license from HP required for possession, use or copying. Consistent with FAR 12.211 and 12.212, Commercial Computer Software, Computer Software Documentation, and Technical Data for Commercial Items are licensed to the U.S. Government under vendors standard commercial license.
The information contained herein is subject to change without notice. The only warranties for HP products and services are set forth in the express warranty statements accompanying such products and services. Nothing herein should be construed as constituting an additional warranty. HP shall not be liable for technical or editorial errors or omissions contained herein.
Intel and Itanium are trademarks or registered trademarks of Intel Corporation or its subsidiaries in the United States and other countries.
Microsoft and Windows are US registered trademarks of Microsoft Corporation.
Printed in the US
The HP OpenVMS documentation set is available on CD-ROM.
HP OpenVMS Cluster Systems describe system management for OpenVMS Cluster systems. Although the OpenVMS Cluster software for Integrity servers and Alpha computers is separately purchased, licensed, and installed, the difference between the two architectures lies mainly in the hardware used. Essentially, system management for Integrity servers and Alpha computers in an OpenVMS Cluster is identical. Exceptions are pointed out.
Who Should Use This Manual
This document is intended for anyone responsible for setting up and managing OpenVMS Cluster systems. To use the document as a guide to cluster management, you must have a thorough understanding of system management concepts and procedures, as described in the HP OpenVMS System Manager's Manual.
How This Manual Is Organized
HP OpenVMS Cluster Systems contains ten chapters and seven appendixes.
Chapter 1 introduces OpenVMS Cluster systems.
Chapter 2 presents the software concepts integral to maintaining OpenVMS Cluster membership and integrity.
Chapter 3 describes various OpenVMS Cluster configurations and the ways they are interconnected.
Chapter 4 explains how to set up an OpenVMS Cluster system and coordinate system files.
Chapter 5 explains how to set up an environment in which resources can be shared across nodes in the OpenVMS Cluster system.
Chapter 6 discusses disk and tape management concepts and procedures and how to use Volume Shadowing for OpenVMS to prevent data unavailability.
Chapter 7 discusses queue management concepts and procedures.
Chapter 8 explains how to build an OpenVMS Cluster system once the necessary preparations are made, and how to reconfigure and maintain the cluster.
Chapter 9 provides guidelines for configuring and building large OpenVMS Cluster systems, booting satellite nodes, and cross-architecture booting.
Chapter 10 describes ongoing OpenVMS Cluster system maintenance.
Appendix A lists and defines OpenVMS Cluster system parameters.
Appendix B provides guidelines for building a cluster common user authorization file.
Appendix C provides troubleshooting information.
Appendix D presents three sample programs for LAN control and explains how to use the Local Area OpenVMS Cluster Network Failure Analysis Program.
Appendix E describes the subroutine package used with local area OpenVMS Cluster sample programs.
Appendix F provides techniques for troubleshooting network problems related to the NISCA transport protocol.
Appendix G describes how the interactions of workload distribution and network topology affect OpenVMS Cluster system performance, and discusses transmit channel selection by PEDRIVER.
This document is not a one-volume reference manual. The utilities and commands are described in detail in the HP OpenVMS System Manager's Manual, the HP OpenVMS System Management Utilities Reference Manual, and the HP OpenVMS DCL Dictionary.
For additional information on the topics covered in this manual, see the following documents:
For additional information about HP OpenVMS products and services, see:
HP welcomes your comments on this manual. Please send your comments or suggestions to:
How To Order Additional Documentation
For information about how to order additional documentation, see:
The following conventions are used in this manual:
|Resource sharing||OpenVMS Cluster software automatically synchronizes and load balances batch and print queues, storage devices, and other resources among all cluster members.|
|Flexibility||Application programmers do not have to change their application code, and users do not have to know anything about the OpenVMS Cluster environment to take advantage of common resources.|
|High availability||System designers can configure redundant hardware components to create highly available systems that eliminate or withstand single points of failure.|
|Nonstop processing||The OpenVMS operating system, which runs on each node in an OpenVMS Cluster, facilitates dynamic adjustments to changes in the configuration.|
|Scalability||Organizations can dynamically expand computing and storage resources as business needs grow or change without shutting down the system or applications running on the system.|
|Performance||An OpenVMS Cluster system can provide high performance.|
|Management||Rather than repeating the same system management operation on multiple OpenVMS systems, management tasks can be performed concurrently for one or more nodes.|
|Security||Computers in an OpenVMS Cluster share a single security database that can be accessed by all nodes in a cluster.|
|Load balancing||OpenVMS Cluster systems distribute work across cluster members based on the current load of each member.|
OpenVMS Cluster system configurations consist of hardware components from the following general groups:
References: Detailed OpenVMS Cluster configuration
guidelines can be found in the OpenVMS Cluster Software Product
Description (SPD) and in Guidelines for OpenVMS Cluster Configurations.
Up to 96 computers, ranging from desktop to mainframe systems, can be members of an OpenVMS Cluster system. Active members that run the OpenVMS Alpha or OpenVMS Integrity server operating system and participate fully in OpenVMS Cluster negotiations can include:
An interconnect is a physical path that connects computers to other computers and to storage subsystems. OpenVMS Cluster systems support a variety of interconnects (also referred to as buses) so that members can communicate using the most appropriate and effective method possible:
The CI, DSSI, and FDDI interconnects are supported on Alpha and VAX systems. Memory Channel and ATM interconnects are supported only on Alpha systems. For more information about these interconnects, see the previous version of the manual.
|IP: UDP||Integrity servers and Alpha||Supports Fast Ethernet/Gigabit Ethernet/10 Gb Ethernet. 10 Gb Ethernet is supported on Integrity servers only.|
|Fibre Channel||Integrity servers and Alpha||Shared storage only|
|SCSI||Integrity servers and Alpha||Limited shared storage configurations only|
|LAN: Ethernet, Fast Ethernet, Gigabit Ethernet, and 10 Gb Ethernet||Integrity servers and Alpha||10 Gb Ethernet is supported on Integrity servers only|
|MEMORY CHANNEL||Alpha||Node-to-node communications only|
For the most recent list of supported interconnects and speeds, see the HP OpenVMS Cluster Software Software Product Description (SPD 29.78.xx):
In addition to the physical interconnects listed in Section 1.2.2, another type of interconnect, a shared memory CI (SMCI) for OpenVMS Galaxy instances, is available. SMCI supports cluster communications between Galaxy instances.
For more information about SMCI and Galaxy configurations, see the
HP OpenVMS Alpha Partitioning and Galaxy Guide.
1.2.4 Storage Devices
A shared storage device is a disk or tape that is accessed by multiple computers in the cluster. Nodes access remote disks and tapes by means of the MSCP and TMSCP server software (described in Section 1.3.1).
Systems within an OpenVMS Cluster support a wide range of storage devices:
|HSG and HSV||FC|
|LSI 1068 and LSI Logic 1068e||SAS|
For the most recent list of supported storage devices, see the HP OpenVMS Version 8.4 Software Software Product Description (SPD 29.78.xx).
Also see the AlphaServer Supported Options Lists that can be found at the individual AlphaServer Web pages:
See the Integrity servers Supported Options Lists that can be found at the individual Integrity server Web pages at: