![]() |
Software > OpenVMS Systems > Documentation > 84final > 4477 ![]() HP OpenVMS Systems Documentation |
![]() |
HP OpenVMS Cluster Systems
Chapter 2
|
Not all interconnects are supported on all three architectures of OpenVMS. The CI, DSSI, and FDDI interconnects are supported on Alpha and VAX systems. Memory Channel and ATM interconnects are supported only on Alpha systems. |
This lowest level of the architecture provides connections, in the form of communication ports and physical paths, between devices. The port layer can contain any of the following interconnects:
Each interconnect is accessed by a port (also referred to as an adapter) that connects to the processor node. For example, the Fibre Channel interconnect is accessed by way of a Fibre Channel port.
The SCS layer provides basic connection management and communications services in the form of datagrams, messages, and block transfers over each logical path. Table 2-1 describes these services.
Service | Delivery Guarantees | Usage |
---|---|---|
Datagrams | ||
Information units that fit in 1 packet or less. | Delivery of datagrams is not guaranteed. Datagrams can be lost, duplicated, or delivered out of order. |
Status and information messages whose loss is not critical.
Applications that have their own reliability protocols such as DECnet or TCP/IP. |
Messages | ||
Information units that fit in 1 packet or less. | Messages are guaranteed to be delivered and to arrive in order. Virtual circuit sequence numbers are used on the individual packets. | Disk read and write requests. |
Block data transfers | ||
Copying (that is, reading or writing) any contiguous data between a local process or system virtual address space and an address on another node. Individual transfers are limited to the lesser of 2 32-1 bytes, or the physical memory constraints of the host. Block data is a form of remote DMA transfer. | Delivery of block data is guaranteed. The sending and receiving ports and the port emulators cooperate in breaking the transfer into data packets and ensuring that all packets are correctly transmitted, received, and placed in the appropriate destination buffer. Block data transfers differ from messages in the size of the transfer. | Disk subsystems and disk servers to move data associated with disk read and write requests. Fast remastering of large lock trees. Transferring large ICC messages. |
The SCS layer is implemented as a combination of hardware and software, or software only, depending upon the type of port. SCS manages connections in an OpenVMS Cluster and multiplexes messages between system applications over a common transport called a virtual circuit. A virtual circuit exists between each pair of SCS ports and a set of SCS connections that are multiplexed on that virtual circuit.
The next higher layer in the OpenVMS Cluster architecture consists of the SYSAPs layer. This layer consists of multiple system applications that provide, for example, access to disks and tapes and cluster membership control. SYSAPs can include:
These components are described in detail later in this chapter.
A wide range of OpenVMS components layer on top of the OpenVMS Cluster system architecture, including:
These components, except for volume shadowing, are described in detail later in this chapter. Volume Shadowing for OpenVMS is described in Section 6.6.
The OpenVMS Cluster software components that implement OpenVMS Cluster communication and resource-sharing functions always run on every computer in the OpenVMS Cluster. If one computer fails, the OpenVMS Cluster system continues operating, because the components still run on the remaining computers.
The following table summarizes the OpenVMS Cluster communication and resource-sharing functions and the components that perform them.
Function | Performed By |
---|---|
Ensure that OpenVMS Cluster computers communicate with one another to enforce the rules of cluster membership | Connection manager |
Synchronize functions performed by other OpenVMS Cluster components, OpenVMS products, and other software components | Distributed lock manager |
Share disks and files | Distributed file system |
Make disks available to nodes that do not have direct access | MSCP server |
Make tapes available to nodes that do not have direct access | TMSCP server |
Make queues available | Distributed job controller |
The connection manager ensures that computers in an OpenVMS Cluster system communicate with one another to enforce the rules of cluster membership.
Computers in an OpenVMS Cluster system share various data and system resources, such as access to disks and files. To achieve the coordination that is necessary to maintain resource integrity, the computers must maintain a clear record of cluster membership.
The connection manager creates an OpenVMS Cluster when the first computer is booted and reconfigures the cluster when computers join or leave it during cluster state transitions. The overall responsibilities of the connection manager are to:
A primary purpose of the connection manager is to prevent cluster partitioning, a condition in which nodes in an existing OpenVMS Cluster configuration divide into two or more independent clusters.
Cluster partitioning can result in data file corruption because the distributed lock manager cannot coordinate access to shared resources for multiple OpenVMS Cluster systems. The connection manager prevents cluster partitioning using a quorum algorithm.
The quorum algorithm is a mathematical method for determining if a majority of OpenVMS Cluster members exist so that resources can be shared across an OpenVMS Cluster system. Quorum is the number of votes that must be present for the cluster to function. Quorum is a dynamic value calculated by the connection manager to prevent cluster partitioning. The connection manager allows processing to occur only if a majority of the OpenVMS Cluster members are functioning.
Two system parameters, VOTES and EXPECTED_VOTES, are key to the computations performed by the quorum algorithm. The following table describes these parameters.
Parameter | Description |
---|---|
VOTES |
Specifies a fixed number of votes that a computer contributes toward
quorum. The system manager can set the VOTES parameters on each
computer or allow the operating system to set it to the following
default values:
Each Integrity server or an Alpha computer with a nonzero value for the VOTES system parameter is considered a voting member. |
EXPECTED_VOTES | Specifies the sum of all VOTES held by OpenVMS Cluster members. The initial value is used to derive an estimate of the correct quorum value for the cluster. The system manager must set this parameter on each active Integrity server system or an Alpha system, including satellites in the cluster. |
The quorum algorithm operates as follows:
Step | Action | ||||||
---|---|---|---|---|---|---|---|
1 |
When nodes in the OpenVMS Cluster boot, the connection manager uses the
largest value for EXPECTED_VOTES of all systems present to derive an
estimated quorum value according to the following
formula:
Estimated quorum = (EXPECTED_VOTES + 2)/2 | Rounded down |
||||||
2 |
During a state transition (whenever a node enters or leaves the cluster
or when a quorum disk is recognized), the connection manager
dynamically computes the cluster quorum value to be the
maximum of the following:
Note: Quorum disks are discussed in Section 2.3.8. |
||||||
3 |
The connection manager compares the cluster votes value to the cluster
quorum value and determines what action to take based on the following
conditions:
|
Note: When a node leaves the OpenVMS Cluster system, the connection manager does not decrease the cluster quorum value. In fact, the connection manager never decreases the cluster quorum value; the connection manager only increases the value, unless the REMOVE NODE option was selected during shutdown. However, system managers can decrease the value according to the instructions in Section 10.11.2.
Consider a cluster consisting of three computers, each computer having its VOTES parameter set to 1 and its EXPECTED_VOTES parameter set to 3. The connection manager dynamically computes the cluster quorum value to be 2 (that is, (3 + 2)/2). In this example, any two of the three computers constitute a quorum and can run in the absence of the third computer. No single computer can constitute a quorum by itself. Therefore, there is no way the three OpenVMS Cluster computers can be partitioned and run as two independent clusters.
To select the optimal sub-cluster and to continue after the communication failure occurs, two possible sub-clusters are compared as follows:
A cluster system manager can designate a disk a quorum disk. The quorum disk acts as a virtual cluster member whose purpose is to add one vote to the total cluster votes. By establishing a quorum disk, you can increase the availability of a two-node cluster; such configurations can maintain quorum in the event of failure of either the quorum disk or one node, and continue operating.
Note: Setting up a quorum disk is recommended only for OpenVMS Cluster configurations with two nodes. A quorum disk is neither necessary nor recommended for configurations with more than two nodes.
For example, assume an OpenVMS Cluster configuration with many satellites (that have no votes) and two nonsatellite systems (each having one vote) that downline load the satellites. Quorum is calculated as follows:
(EXPECTED VOTES + 2)/2 = (2 + 2)/2 = 2 |
Because there is no quorum disk, if either nonsatellite system departs from the cluster, only one vote remains and cluster quorum is lost. Activity will be blocked throughout the cluster until quorum is restored.
However, if the configuration includes a quorum disk (adding one vote to the total cluster votes), and the EXPECTED_VOTES parameter is set to 3 on each node, then quorum will still be 2 even if one of the nodes leaves the cluster. Quorum is calculated as follows:
(EXPECTED VOTES + 2)/2 = (3 + 2)/2 = 2 |
Rules: Each OpenVMS Cluster system can include only one quorum disk. At least one computer must have a direct (not served) connection to the quorum disk:
Reference: For more information about enabling a quorum disk, see Section 8.2.4. Section 8.3.2 describes removing a quorum disk.
Method | Perform These Steps |
---|---|
Run the CLUSTER_CONFIG.COM procedure
(described in Chapter 8) |
Invoke the procedure and:
The procedure uses the information you provide to update the values of the DISK_QUORUM and QDSKVOTES system parameters. |
Respond YES when the OpenVMS installation procedure asks whether the
cluster will contain a quorum disk
(described in Chapter 4) |
During the installation procedure:
The procedure uses the information you provide to update the values of the DISK_QUORUM and QDSKVOTES system parameters. |
Edit the
MODPARAMS or AGEN$ files (described in Chapter 8) |
Edit the following parameters:
|
Hint: If only one quorum disk watcher has direct access to the quorum disk, then remove the disk and give its votes to the node.
For the quorum disk's votes to be counted in the total cluster votes, the following conditions must be met:
Hint: By increasing the quorum disk's votes to one less than the total votes from both systems (and by increasing the value of the EXPECTED_VOTES system parameter by the same amount), you can boot and run the cluster with only one node.
OpenVMS Cluster state transitions occur when a computer joins or leaves an OpenVMS Cluster system and when the cluster recognizes a quorum disk state change. The connection manager controls these events to ensure the preservation of data integrity throughout the cluster.
A state transition's duration and effect on users (applications) are determined by the reason for the transition, the configuration, and the applications in use.
Every transition goes through one or more phases, depending on whether its cause is the addition of a new OpenVMS Cluster member or the failure of a current member.
Table 2-2 describes the phases of a transition caused by the addition of a new member.
Phase | Description |
---|---|
New member detection |
Early in its boot sequence, a computer seeking membership in an OpenVMS
Cluster system sends messages to current members asking to join the
cluster. The first cluster member that receives the membership request
acts as the new computer's advocate and proposes reconfiguring the
cluster to include the computer in the cluster. While the new computer
is booting, no applications are affected.
Note: The connection manager will not allow a computer to join the OpenVMS Cluster system if the node's value for EXPECTED_VOTES would readjust quorum higher than calculated votes to cause the OpenVMS Cluster to suspend activity. |
Reconfiguration | During a configuration change due to a computer being added to an OpenVMS Cluster, all current OpenVMS Cluster members must establish communications with the new computer. Once communications are established, the new computer is admitted to the cluster. In some cases, the lock database is rebuilt. |
Previous | Next | Contents | Index |