1 basic information collection
1.1 View model
# model
9000/800 / L3000-7X
1.2 View CPU and memory usage
#top
System: portal1 thu jul 1 11:55:15 2004
Load averages: 0.00, 0.02, 0.03
143 Processes: 135 Sleeping, 8 Running
CPU State:
CPU Load User Nice Sys iDle Block Swait Intr Ssys
0 0.00 0.0% 0.0% 0.2% 99.8% 0.0% 0.0% 0.0% 0.0%
1 0.00 0.0% 0.0% 0.0% 100.0% 0.0% 0.0% 0.0% 0.0%
---- - - ---------------------------------
AVG 0.00 0.0% 0.0% 0.2% 99.8% 0.0% 0.0% 0.0% 0.0%
Memory: 815620K (668292K) Real, 1130440K (920468K) Virtual, 573140K Free Page # 1/5
CPU TTY PID Username Pri Ni Size RES State Time% WCPU% CPU Command
0? 23932 root -27 20 10848k 7032k Run 6:15 0.68 0.68 cmCLD
0? 35 root 152 20 1472k 1472k Run 6:19 0.33 0.33 VXFSD
0? 1197 Oracle 156 20 438M 19736K Sleep 0:00 0.19 0.19 Ora_SMON_PTL1
...
Exit Press 'Q'
1.3 How to view memory
#dmesg
Jul 1 12:32
Gate64: sysvec_vaddr = 0xc0002000 for 2 Pages
NOTOFS_LINK (): File System Was Registered At Index 3.
Notice: cachefs_link (): File System Was Registered At Index 5.
NOTICE: NFS3_LINK (): File System Was Registered At Index 6.
0 SBA
0/0 LBA
...
1.4 View the number of processors
# getconf kernel_bits
64
1.5 View Switch Space (SWAP)
# swapinfo -a
KB KB KB PCT START / KB
TYPE AVAIL Used Free Used Limit Reserve Pri Name
DEV 4194304 0 4194304 0% 0 - 1 / dev / vg00 / lvol2
RESERVE - 69224-69224
Memory 1572088 167116 1404972 11%
1.6 View the size of the hard drive
# DiskInfo / dev / rdsk / c1t2d0
SCSI DESCRIBE OF / DEV / RDSK / C1T2D0: VENDOR: HP 36.4G
Product ID: MAS3367NC
TYPE: DIRECT Access
Size: 35566480 KBYTES
BYTES Per Sector: 512 View operating system version and license
#uname -a
HP-UX SCP1 B.11.00 U 9000/800 1124961527 Unlimited-User License
1.7 View the number of hard drives
# ioscan -func disk
Class I h / w path driver s / w state h / w type description
============================================================================================================================================================================================================= ======================
Disk 0 0/0/1 / 1.2.0 SDisk Claimed Device HP 36.4Gmas3367NC
/ DEV / DSK / C1T2D0 / DEV / RDSK / C1T2D0
Disk 1 0/0/2 / 0.2.0 SDisk Claimed Device HP 36.4Gmas3367NC
/ DEV / DSK / C2T2D0 / DEV / RDSK / C2T2D0
Disk 2 0/0/2 / 1.2.0 SDISK CLAIMED Device HP DVD-ROM 305
/ DEV / DSK / C3T2D0 / DEV / RDSK / C3T2D0
Disk 3 0/8/0 / 0.8.0.110.0.0.0 SDISK CLAIMED Device HP A6189B
/ DEV / DSK / C4T0D0 / DEV / RDSK / C4T0D0
Disk 4 0/8/0 / 0.8.0.110.0.0.1 SDISK CLAIMED Device HP A6189B
/ DEV / DSK / C4T0D1 / DEV / RDSK / C4T0D1
1.8 View file system
# bdf
FileSystem Kbytes Used Avail% Used Mounted ON
/ DEV / VG00 / LVOL3 204800 43440 160120 21% /
/ DEV / VG00 / LVOL1 298928 51960 217072 19% / stand
/ DEV / VG00 / LVOL8 4710400 431352 4246240 9% / var
/ DEV / VG00 / LVOL6 2097152 1097128 992240 53% / usr
/ DEV / VG00 / LVOL5 1310720 633376 672696 48% / TMP
/ DEV / VG00 / LVOL4 8388608 4207536 4155184 50% / OPT
/ DEV / VG00 / LVOL7 10485760 4115968 6320080 39% / home exists in two files: / etc / fstab / etc / mnttab
1.9 View PV
# pvdisplay / dev / dsk / c4t0d1
--- Physical Volumes ---
PV Name / DEV / DSK / C4T0D1
VG Name / dev / vg_data
PV Status Available
Allocatable Yes
VGDA 2
Cur lv 24
PE size (mbytes) 4
Total PE 10238
Free PE 3631
Allocated PE 6607
Stale PE 0
IO Timeout (Seconds) Default
Autoswitch on
1.10 View VG
# vgdisplay vg_data
--- Volume Groups ---
VG Name / dev / vg_data
VG WRITE Access Read / Write
VG Status Available, Shared, Server
Max LV 255
Cur lv 24
Open LV 24
MAX PV 16
Cur Pv 1
ACT PV 1
Max pee PV 10239
VGDA 2
PE size (mbytes) 4
Total PE 10238
Alloc PE 6607
Free PE 3631
Total PVG 0
Total SPARE PVS 0
Total Spare PVS in Use 0
1.11 View LV
# lvdisplay / dev / vg_data / Ring_Data
--- Logical Volumes ---
LV Name / dev / vg_data / ring_data
VG Name / dev / vg_data
LV Permission Read / Write
LV Status Available / Syncd
Mirror COPIES 0
Consistency Recovery MWC
Schedule Parallel
LV size (mbytes) 1000
Current Le 2500
Allocated PE 2500Stripes 0
Stripe size (kbytes) 0
Bad Block on
ALLOCATION STRICT
IO Timeout (Seconds) Default
1.12 View machine name
# Hostname
Portal1
1.13 View Network Resources
# lanscan
Hardware Station CRD HDW Net-Interface NM Mac HP-DLPI DLPI
Path Address In # state nameppa ID Type Support MJR #
0/0/0/0 0x00306EC32158 0 UP LAN0 SNAP0 1 Ether Yes 119
0/10/0/0 0x00306E4AA9FA 1 UP LAN1 SNAP1 2 Ether Yes 119
0/12/0/0 0x00306E4AA9D9 2 UP LAN2 SNAP2 3 Ether Yes 119
1.14 View a NIC
# impconfig lan0
LAN0: Flags = 843
INET 10.71.111.171 Netmask ffffff80 Broadcast 10.71.111.255
1.15 View Floating IP
# netstat -in
Name MTU Network Address IPKTS IERRS OPKTS OERRS COLL
LAN0: 1 1500 10.71.111.128 10.71.111.170 3 0 0 0 0
LAN2 * 1500 None None 0 0 0 0 0
LAN1 1500 192.0.166.0 192.0.166.1 2099845 0 2760 0 0
LAN0 1500 10.71.111.128 10.71.111.171 7876597 0 4918169 0 0
LO0 4136 127.0.0.0 127.0.0.1 535887 0 535887 0 0
View VG
2 disk and file system configuration
2.1 Creating a PV (Format Disk)
# pvcreate -f / dev / rdsk / cxtydz ------ (x = instance, y = target, and z = unit)
2.2 Creating VG
l Create a VG directory and feature file Group
#MKDIR / DEV / VG_DATA
# MKNOD / DEV / VG_DATA / GROUP C 64 0x060000
Note: The auxiliary number of Group files must be unique in the VG on all nodes of the system.
l Create a PV-LINKS
# vgcreate / dev / vg_data / dev / dsk / c0t1d0 / dev / dsk / c1t0d0
l extended VG
# VGEXTEND / DEV / VG_DATA / DEV / DSK / C1T0D1 / DEV / DSK / C0T1D1 Continue to extend the VG, know that all required disks are expanded in.
2.3 Creating LV
#LVCREATE -L 100-NAME / DEV / VG_DATA
The size of the LV created by default is PE, each PE is 4m.
2.4 Creating a File System
l Initialization file system
#NewFS -F HFS / DEV / VGLOCK / RLV_Oracle
l Create a directory of the file system mount
#mkdir / oracle
l Mount the file system
#mount / dev / vglock / lv_oracle / oracle
2.5 activation VG
l monograph mode activation
1) deactivate VG
#vgchange -a n / dev / vg_data
2) Modify the properties of VG on all nodes
# vgchange-c n / dev / vg_data
3) Activate VG
#vgchange -a y / dev / vg_data
l sharing mode activation
1) deactivate VG
#vgchange -a n / dev / vg_data
2) Modify the properties of VG on all nodes
#vgchange -s y -c y / dev / vg_data
3) Activate all nodes VG
#vgchange -a s / dev / vg_data
2.6 deactivation VG
#vgchange -a n / dev / vg_data
2.7 Export VG
l Deactivate VG:
#vgchange -a n / dev / vg_data
l Create a VG mapping file:
#vgexport -v -p -s -m vgdataMap / dev / vg_data
l All nodes in the copy map of the file collection:
# RCP VGDataMap Node2: / TMP / VGDATAMAP
2.8 Import VG
l Create a VG directory containing the feature file Group on the import node:
# MKDIR / DEV / VG_DATA
# MKNOD / DEV / VG_DATA / GROUP C 64 0x060000
Note: The auxiliary number must be the same as the export node.
l Import VG:
# vgimport -v -s -m / tmp / vgdataMap / dev / vg_data
l Check if the device has been imported:
# strings / etc / lvmtab
2.9 expand LV
l Take all the nodes to activate VG
#vgchange -a n / dev / vg_data
l Deleted VG on the second node
#VGEXPORT -P -V -M /TMP/vgdata.map / dev / vgdata
#vgexport -s -m /tmp/vgdata.map / dev / vgdata
l The master node monographically activated VG
# vgchange-c n / dev / vg_data
#vgchange -a y / dev / vg_data
l Primary node expands LV
#LVEXTEND -L 1000 / dev / vglock / ring_data (unit is PE)
l Export VG at the primary node
#vgexport -v -p -s -m vgdataMap / dev / vg_data
l Copy the secondary node in the mapping group:
# RCP VGDataMap Node2: / TMP / VGDATAMAP
l Create a VG directory containing the feature file group on the second node:
# MKDIR / DEV / VG_DATA
# MKNOD / DEV / VG_DATA / GROUP C 64 0x060000
Note: The auxiliary number must be the same as the main node.
l Import VG on the second node:
# vgimport -v -s -m / tmp / vgdataMap / dev / vg_data
2.10 Expansion File System
l unmount file system
#umount / oracle
l Extended LV:
#LVEXTEND -L 50 / dev / vg00 / lv_oracle
l extended file system
#extenDFS / DEV / VG00 / LV_ORACLE
(If it is a VXFS file system, use #extendfs -f vxfs / dev / vg00 / lv_oracle)
l mount file system
#mount / dev / vg00 / lv_oracle / oracle
2.11 Expansion System Directory File System
l Enter HP-UX in single users
1) Restart,
#reboot
2) After the self-test is completed, this line "to Discontinue, Press Any Key in 10 Seconds", at this time, the system is started, and the system is started, enters Main Menu "
3) Type "Bo", when the system is inquiry "Interact with IPL (Y / N?)?", Enter "Y".
4) After the prompt isl>, enter "HPUX - IS", the system enters the single user status (ie maintenance mode)
l extended LV
#extenDFS / DEV / VG00 / LV_ORACLE
(If it is a VXFS file system, use #extendfs -f vxfs / dev / vg00 / lv_oracle)
l Switch to multiple user status
2.12 Delete LV
#LVRemove / dev / vglock / lv_informix
2.13 Delete VG
l Deactivate VG
#vgchange -a n / dev / vglock (if you are not activated, you can use the following command to force to activate, vgchange -c n / dev / vglock)
l Pre-deleted VG
#vgexport -p -s -m /tmp/vglock.map / dev / vglock
l Delete VG
#vgexport -s -m /tmp/vglock.map / dev / vglock
3 network configuration
3.1 Configuring an IP address
l Modify NetConf files
#vi /etc/rc.config.d/netconf
# NetConf: Configuration Values for Core Networking Subsystems
#
# @ (#) B.11.11_lr $ revision: 1.6.119.6 $ date: 97/09/10 15:56:01 $
#
# Hostname: Name of Your System for uname -s and hostname
#
# OPERATING_SYSTEM: NAME OF OPERATING SYSTEM RETURNED by uname -s
# ---- Do Not Change this Value ----
#
# Loopback_address: loopback address # ---- do not change this value ----
#
# Important: for 9.x-to-10.0 Transition, Do Not Put Blank Lines Between
# The next set of statements
# Host name, operating system
Hostname = "portal1"
Operating_system = HP-UX
Loopback_address = 127.0.0.1
# Internet configure parameters. See ifconfig (1M), Autopush (1M)
#
# Interface_name: Network interface name (see lanscan (1m))
#
# Ip_address: Hostname (In / etc / hosts) or ip address in decimal-dot
# NOTATION (E.G., 192.1.2.3)
#
# Subnet_mask: SubnetWork Mask in Decimal-Dot Notassion, IF DIFFERENT
# From default
#
# Broadcast_address: Broadcast Address in Decimal-Dot Notation, IF
# DIFFERENT from DEFAULT
#
# Interface_State: Desired Interface State At Boot Time.
# Either Up or Down, default is up.
#
# DHCP_ENABLE DETERMINES WHETER OR NOT DHCP Client FunctionAlity
# Will Be enabled on the network interface (see
# Auto_parms (1M), DHCPCLIENT (1M)). DHCP Clients Get
# Their IP Address Assignments from DHCP Servers.
DHCP Client Functionality; 0 Disables IT.
#
# For Each Additional Network Interfaces, Add A Set of Variable Assignments
# Like the ones below, Changing the index to "[1]", "[2]" ET Cetera.
#
# Important: for 9.x-to-10.0 Transition, Do Not Put Blank Lines Between
# The next set of statements
# 一网 网卡 i 配置
Interface_name [0] = lan0
IP_ADDRESS [0] = 10.71.111.171
SUBNET_MASK [0] = 255.255.255.128
Broadcast_address [0] = ""
Interface_state [0] = ""
DHCP_ENABLE [0] = 0
# Internet routing configuration. See Route (1M), Routing (7)
#
# Route_Destination: Destination Hostname (in / etc / hosts) or host or network
# Ip address in decimal-dot notation, preceded by the word
# "Host" or "net"; or simply the word "default".
#
# Route_mask: SubnetWork Mask in Decimal-Dot Notation, Or C Language
# Hxadecimal notation. This is an optional field.
# A ip address, SUBNET MASK PAIR UNIQUELY IDENTIFIES
# A subnet to be reached. If a subnet mask is not given,
# THEN THE SYSTEM WILL Assign THE LONGEST SUBNET MASK
# Of the configured network interfaces to this route.
# I i n no matching subnet mask, THEN THE SYSTEM
# Will Assign The Default Network Mask as The Route'S
# Subnet Mask.
#
# Route_gateway: Gateway Hostname (In / etc / hosts) OR IP Address in
# Decimal-dot notation. If local interface, Must Use the
# Same form as buy for ip_address above (Hostname OR)
# Decimal-dot notation). If loopback interface, i.e.,
# 127.0.0.1, The route_count must be set to zero.
#
# Route_count: an INTEGER That INDICES WHETER THE GATEWAY IS A
# Remote Interface (One) or the local interface (zero)
# Or loopback interface (E.G., 127. *).
#
# Route_Args: Route Command Arguments and Options. This Variable
# May Contain A Combination of The Following Arguments: # "-f", "-n" and "-p pmtu".
#
# For Each Additional Route, Add A Set of Variable Assignments Like the ones
# Below, Changing the index to "[1]", "[2]" ET Cetera.
#
# Important: for 9.x-to-10.0 Transition, Do Not Put Blank Lines Between
# The next set of statements
# 路 路
Route_Destination [0] = default
ROUTE_MASK [0] = ""
Route_gateway [0] = "10.71.111.129"
ROUTE_COUNT [0] = "1"
Route_args [0] = ""
# Dynamic Routing Daemon Configuration. See Gated (1M)
#
# Gated: set to 1 to Start Gated Daemon.
# Gaated_args: arguments to the gaical daemon.
Gated = 0
Gaate_args = ""
#
# Router Discover Protocol Daemon Configuration. See Rdpd (1M)
#
# R r 1 1 to start rdpd daemon
#
RDPD = 0
#
# Reverse Arp Daemon Configuration. See Rarpd (1M)
#
# RARP: SET TO 1 To Start Rarpd Daemon
#
# 网关
RARP = 0
Route_gateway [0] = 10.71.111.129
ROUTE_COUNT [0] = 1
Route_Destination [0] = default
#
# 第二 网
Interface_name [1] = lan1
IP_ADDRESS [1] = 192.0.166.1
SUBNET_MASK [1] = 255.255.255.0
Broadcast_address [1] = ""
Interface_state [1] = ""
DHCP_ENABLE [1] = 0
After modifying the file, you need to restart the network in order to make the changes take effect.
l Configure IP for a block of NICs separately with the following command.
#ifconfig lan0 110.71.111.171 255.255.255.128
#ifconfig lant 192.0.166.1 255.255.255.0
3.2 IP, host name mapping
#vi / etc / hosts
# @ (#) B.11.11_lrhosts $ revision: 1.9.214.1 $ date: 96/108 13:20:01 $
#
# The Form for Each Entry IS:
#
#
# For example:
# 192.1.2.34 HPFCRM LOGHOST #
# See the hosts (4) Manual page for more information.
# Note: The Entries Cannot Be Preceded by A Space.
# The Format Described in this file is the correct format.
# The Original Berkeley Manual Page Contains An Error In
# The format description.
#
10.71.111.171 Portal1
127.0.0.1 Localhost loopback
10.71.111.172 Portal2
10.71.111.142 AIP1
10.71.111.140 AIP2
#
192.0.166.1 Portal1HT # Heat Beat of Portal1
192.0.166.2 Portal2ht # heat beat of portal2
3.3 Trust relationship
To build a mutual trust house in multi-node, edit the.rhosts file in the home directory of the account on each node.
#vi /Root/.rhosts
Portal2 root
Portal1 root
3.4 Stopping Network
# / sbin / rc2.d / s340net stop
3.5 Startup Network
# /sbin/rc2.d/s340NET START
3.6 activation network card
#ifconfig lan0 Up // Activate NIC LAN0
3.7 destination
#ifconfig lan0 down // to activate the network card LAN0
4 cluster command
4.1 View MC / Serviceguard Version
#cmversion
A.11.15.00
4.2 View cluster status
#cmviewcl -v
Cluster status
Portal_Cluster Up
Node Status State
Portal1 Up Running
Network_parameters:
Interface Status path name
Primary up 0/0/0/0 lan0
PRIMARY UP 0/10/0 LAN1
Standby up 0/12/0/0 lan2
Node Status State
Portal2 Up Running
Network_parameters:
Interface Status path name
Primary up 0/0/0/0 lan0
Standby up 0/12/0/0 lan2
PRIMARY UP 0/10/0 LAN1Package Status State Auto_Run Node
TestPKG UP Running Enabled Portal2
Policy_Parameters:
Policy_Name Configured_Value
Failover configured_node
Failback Manual
Script_parameters:
Item Status Max_Restarts Restarts Name
Service Up 3 0 ORAMONITOR
Subnet Up 10.71.111.128
Node_Switching_Parameters:
Node_type status switching name
PRIMARY UP ENABED Portal1
Alternate Up Enabled Portal2 (CURRENT)
4.3 Configuring Clusters
1. Create a cluster configuration template:
#cmquerycl -n portal1 -n portal2 -v -c /etc/cmcluster/rac.asc
2. Edit the cluster configuration file (rac.asc).
#vi /etc/cmculster/rac.asc
#************************************************************ ********************
# ****** HIGH AVAILABILITY Cluster Configuration File *************
# ***** for Complete Details About Cluster Parameters and how to *******
# ***** SET THEM, CONSULT The Serviceguard Manual. *****************************
#************************************************************ ********************
# Enter a name for this cluster. This name will be used to ide Identify the
# Cluster when Viewing or Manipulating IT.
# 集群 name
Cluster_name portal_cluster
# Cluster Lock Parameters
#
# The Cluster Lock Is Used as a TIE-BREAKER for Situations
# in Which a Running Cluster Fails, and the Two Equal-Sized
# Sub-Clusters Are Both Trying to Form A New Cluster. The
# Cluster Lock May Be Configured Using Either a Lock Disk
# or a quorum server.
#
# You can use Either the quorum server or the lock disk as
# a Cluster Lock But Not Both in The Same Cluster. #
# Consider The Following When Configuring a Cluster.
# For a two-node cluster, You Must Use a cluster lock. For
# a Cluster of Three Or Four Nodes, a Cluster Lock Is Strongly
# Recommended. For a cluster of more tour nodes, a
# Cluster Lock Is Recommended. if you decide to configure
# a Lock for a Cluster of More Than Four Nodes, IT Must Be
# a quorum server.
# Lock Disk parameters. Use the first_cluster_lock_vg and
#First_cluster_lock_pv parameters to define a lock disk.
# THE FIRST_CLUSTER_LOCK_VG IS The LVM VOLUME GROUP THAT
# Holds The Cluster Lock. THIS VOLUME GROUP SHOULD NOT BE
# Used by Any Other Cluster As a Cluster Lock Device.
# Quorum Server Parameters. Use the qs_host, qs_polling_interval,
# and qs_timeout_extension parameters to define a quorum server.
# THE qs_host is the host name or ip address of the system
# That Is Running The Quorum Server Process. THE
# Qs_polling_interval (microseconds) is the interval at Which
# Serviceguard Checks to make Sure The Quorum Server is running.
# THE OPTIONAL QS_TIMEOUT_EXTENSION (Microseconds) Is Used to Increase
# The Time Interval at Which The Quorum Server is Marked Down.
#
# The default quorum server timeout is cagculated from the
# Serviceguard Cluster Parameters, INCLUDING NODE_TIMEOUT AND
# Hartbeat_interval. If you are experiencing quorum server
# Timeouts, You Can Adjust these Parameters, or you can include
# The qs_timeout_extension parameter.
#
# For example, to configure a quorum server running on node
# "qshost" with 120 seconds for the qs_polling_interval and to
# Add 2 seconds to the system assigned value for the quorum server
# TIMEOUT, ENTER:
#
# Qs_host qshost
# Qs_polling_interval 120000000
# Qs_timeout_extension 2000000
# 盘
First_Cluster_lock_vg / dev / vglock
# Definition of nodes in the cluster.
# Repeat node definitions as Necessary for additional nodes.
# Node_name is The specified nodename in the cluster.
# It must match the hostname and both Cannot Contain Full Domain name.
# Each Network_Interface, if configured with ipv4 address,
# Must Have ONLY ONE IPV4 Address Entry with it Which Could
# Be Either HeartBeat_ip or stationary_ip.
# Each Network_Interface, IF Configured with IPv6 Address (ES)
# Can Have Multiple IPv6 Address Entries (Up to a Maximum of 2,
# Only One IPv6 Address Entry Belonging to Site-Local Scope
# and only one belonging to global scope) Which MUST BE ALL
# Stationary_ip. They cannot behartbeat_ip.
# Node 1 configuration
Node_name portal1
Network_interface lan0
Heartbeat_ip 10.71.111.171
Network_interface lan2
Network_interface lan1
Stationary_ip 192.0.166.1
First_cluster_lock_pv / dev / dsk / c4t0d0
# List of serial device file names
# For example:
# Serial_Device_file / dev / tty0p0
# Possible Standby Network Interfaces for lan0, lan1: lan2.
# Node 2 configuration
Node_name portal2
Network_interface lan0
Heartbeat_ip 10.71.111.172
Network_interface lan2
Network_interface lan1
Stationary_ip 192.0.166.2
First_cluster_lock_pv / dev / dsk / c4t0d0
# List of serial device file names
# For example:
# Serial_Device_file / dev / tty0p0
# Possible Standby Network Interfaces for lan0, lan1: lan2.
# Cluster Timing Parameters (Microseconds).
# The node_timeout parameter Defaults to 2000000 (2 Seconds).
# This default setting yields the failest cluster reformations. # However, the use of the default value increas the potential
# for spurious reformations due to momentary system hangs or
# NetWork loading spikes.
# For a significant portion of installations, a setting of
# 5000000 to 8000000 (5 to 8 seconds) is more appropriate.
# THE MAXIMUM VALUE Recommended for Node_timeout IS 30000000
# (30 seconds).
# 心 间, timeout setting
Heartbeat_Interval 2000000
Node_timeout 6000000
# Configuration / Reconfiguration Timing Parameters (Microseconds).
# Automatic restart time
Auto_Start_Timeout 600000000
Network_polling_interval 2000000
# Package Configuration Parameters.
# Enter The Maximum Number of Packages Which Will Be configured in the cluster.
# You can not add packages beyond this limit.
# This parameter is required.
# 集 集 中 数
MAX_CONFIGURED_PACKAGES 5
# List of Cluster Aware LVM Volume Groups. Thase Volume Groups Will
# be used by package applications via the vgchange -a e commnd.
#Neither cvm or vxvm disk groups surund be used here.
# For example:
# Volume_group / dev / vgdatabase
# Volume_group / dev / vg02
#LVM management VG,
# 包 应用 应用 a e a 将 生效 进行 进行
VOLUME_GROUP / DEV / VG_DATA
# List of ops volume groups.
# Formerly Known As DLM Volume Groups, Thase Volume GROUPS
# Will Be Used by Ops or Rac Cluster Applications VIA
# The vgchange -a s command. (NOTE: THE NAME DLM_VOLUME_GROUP
# is Also Still Supported for Compatibility with Earlier Versions.
# For example:
# OPS_VOLUME_GROUP / DEV / VGDATABASE
# OPS_VOLUME_GROUP / DEV / VG02
# Oracle RAC used VG
OPS_VOLUME_GROUP / DEV / VG_DATA3. Check the cluster configuration:
# cmcheckconf -v -c /etc/cmcluster/rac.asc
4. Distribute all nodes in the district distribution group. Enable the configuration takes effect.
# cmapplyconf -v -c /etc/cmcluster/rac.asc
4.4 Start Cluster
l Start the cluster from any node:
# CMRuncl -V
l or start a node:
# CMRunnode -v portal1
4.5 Stop Cluster
l If the 9i RAC instance has started, stop them first
l On all nodes, deactivate the volume group:
#vgchange -a n vg_data
l Medium stop cluster at any node:
# cmhaltcl -v
4.6 launch package
# CMRUNPKG [-V] [-node_name] package_name
4.7 stop package
# cmhaltpkg [-v] [-n node_name] package_name
4.8 Modify the switching attribute of the package
# cmmodpkg [-v] {-e | -d} [-n node_name] ... package_name ...
or:
# cmmodpkg [-v] -r -s service_name package_name
4.9 Add IP to the cluster
# cmmodnet [-v] {-A | -r} -i
4.10 launch service
# RunServ [-v] [-r Restart number | -r] Service Name Service Command Line
4.11 Stop Service
# cmhaltserv [-v] service_name
5 appendix
5.1 Creating a LV of Oracle Rac
CreateServerLVS.sh:
Pvcreate -f / dev / rdsk / c4t0d1
Vgcreate vg_data / dev / dsk / c4t0d1
Vgchange -a n vg_data
Vgchange -c n vg_data
Vgchange -a y vg_data
Lvcreate -l 250 -n ora9_system vg_data
Lvcreate -l 250-N Ora9_Temp VG_DATA
Lvcreate -l 125 -n ora9_rbs1 vg_data
Lvcreate -l 125-N Ora9_RBS2 VG_DATA
Lvcreate -l 20 -n ora9_user vg_data
Lvcreate -l 20 -n ora9_index vg_data
Lvcreate -l 15 -n ora9_tools vg_data
Lvcreate -l 25 -n ora9_drsys vg_data
Lvcreate -l 15 -n ora9_spfile vg_data
Lvcreate -l 25-N Ora9_XDB VG_DATA
Lvcreate -l 75 -n ora9_ctl1 vg_data
Lvcreate -l 75 -n ora9_ctl2 VG_DATA
Lvcreate -l 75 -n ora9_ctl3 vg_data
Lvcreate -l 75 -n ora9_redo111 VG_DATA
Lvcreate -l 75 -n ora9_redo122 VG_DATA
Lvcreate -l 75 -n ora9_redo121 vg_datalvcreate -l 75 -n ora9_redo122 VG_DATA
Lvcreate -l 75 -n ora9_redo211 VG_DATA
Lvcreate -l 75 -n ora9_redo212 VG_DATA
Lvcreate -l 75 -n ora9_redo221 vg_data
Lvcreate -l 75 -n ora9_redo222 VG_DATA
Lvcreate -l 7500-N Ring_Data VG_DATA
Lvcreate -l 7500-N Ring_index vg_data
Lvcreate -l 30 -n ora9_srvmconfig vg_data
Vgchange -a n vg_data
VGEXPORT -V -P -S -M / TMP / VGDATAMAP VG_DATA
RCP / TMP / VGDATAMAP Portal2: / TMP / VGDATAMAP
CHMOD 777 / DEV / VG_DATA
CHMOD 660 / DEV / VG_DATA / R *
Chown Oracle: DBA / DEV / VG_DATA / R *
Vgchange -a n vg_data
VGEXPORT -P -S -M / TMP / TEMPMAP VG_DATA
VGEXPORT -S -M / TMP / TEMPMAP VG_DATA
MKDIR / DEV / VG_DATA
MKNOD / DEV / VG_DATA / GROUP C 64 0x060000
Vgimport -V -S -M / TMP / VGDATAMAP VG_DATA
CHMOD 777 / DEV / VG_DATA
CHMOD 660 / DEV / VG_DATA / R *
Chown Oracle: DBA / DEV / VG_DATA / R *
CreateClientlvs.sh:
$ Echo "Create Logic Volumes for Oracle 9i RAC"
Vgchange -a n vg_data
VGEXPORT -P -S -M / TMP / TEMPMAP VG_DATA
VGEXPORT -S -M / TMP / TEMPMAP VG_DATA
MKDIR / DEV / VG_DATA
MKNOD / DEV / VG_DATA / GROUP C 64 0x060000
How to use: Modify the node name, disk path, and LV size in the script createserverlvs.sh. Run the script on the main node; run the script in the second node CreateClientlvs.sh
5.2 Cluster Profile CMCluster.ascii