Oracle10203RAC環境新增新節點(二)

yangtingkun發表於2009-11-16

簡單描述一下,在Oracle 10203 for Solaris sparcRAC雙節點環境中,新增一個節點的過程。共享儲存已經在第三個節點上配置完成,這裡主要介紹作業系統上和Oracle上的配置。

這一篇描述CLUSTER軟體的建立。

Oracle10203RAC環境新增新節點(一):http://yangtingkun.itpub.net/post/468/493738

 

 

在新增節點racnode3上將共享儲存上的裸裝置授權給Oracle使用者:

bash-3.00# chown oracle:oinstall /dev/rdsk/c1t500601603022E66Ad*

建立和節點1、節點2上一致的/dev/rac/vot/dev/rac/ocr連結,執行相同的裸裝置檔案:

bash-3.00# mkdir /dev/rac
bash-3.00# ln -s /dev/rdsk/c1t500601603022E66Ad2s1 /dev/rac/ocr
bash-3.00# ln -s /dev/rdsk/c1t500601603022E66Ad2s3 /dev/rac/vot

在安裝節點執行$ORA_CLS_HOME/out/bin/addNode.sh來啟動圖形化介面新增新的節點:

bash-2.03$ ./addNode.sh
Starting Oracle Universal Installer...

Checking installer requirements...

Checking operating system version: must be 5.8, 5.9 or 5.10.    Actual 5.8
                                      Passed

Checking Temp space: must be greater than 150 MB.   Actual 8715 MB    Passed
Checking monitor: must be configured to display at least 256 colors.    Actual 16777216    Passed

All installer requirements met.

Oracle Universal Installer, Version 10.2.0.1.0 Production
Copyright (C) 1999, 2005, Oracle. All rights reserved.

在圖形介面中,填入新增節點的PUBLICPRIVATEVIP對應的名稱。

安裝過程Oracle跳過了一些遠端執行碰到的錯誤,需要手工更正,在racnode3上執行下面的指令碼:

bash-3.00$ /data/oracle/product/10.2/crs/oui/bin/runInstaller -attachHome -noClusterEnabled ORACLE_HOME=/data/oracle/product/10.2/crs ORACLE_HOME_NAME=OraCrs10g_home1 CLUSTER_NODES=racnode1,racnode2,racnode3 CRS=true "INVENTORY_LOCATION=/data/oracle/oraInventory" LOCAL_NODE=racnode3
Starting Oracle Universal Installer...

Checking installer requirements...

Checking operating system version: must be 5.8, 5.9 or 5.10.    Actual 5.10
                                      Passed

Checking Temp space: must be greater than 250 MB.   Actual 10167 MB    Passed
Checking swap space: must be greater than 500 MB.   Actual 7415 MB    Passed

All installer requirements met.

'AttachHome' was successful.

在節點2上執行下面的指令碼:

bash-2.03$ /data/oracle/product/10.2/crs/oui/bin/runInstaller -updateNodeList -noClusterEnabled ORACLE_HOME=/data/oracle/product/10.2/crs CLUSTER_NODES=racnode1,racnode2,racnode3 CRS=true "INVENTROY_LOCATION=/data/oracle/oraInventory" LOCAL_NODE=racnode2
Starting Oracle Universal Installer...

Checking installer requirements...

Checking operating system version: must be 5.8, 5.9 or 5.10.    Actual 5.8
                                      Passed

Checking Temp space: must be greater than 150 MB.   Actual 8233 MB    Passed
Checking monitor: must be configured to display at least 256 colors.    Actual 16777216    Passed

All installer requirements met.

'UpdateNodeList' was successful.

節點3上執行類似的指令碼:

bash-3.00$ /data/oracle/product/10.2/crs/oui/bin/runInstaller -updateNodeList -noClusterEnabled ORACLE_HOME=/data/oracle/product/10.2/crs CLUSTER_NODES=racnode1,racnode2,racnode3 CRS=true "INVENTROY_LOCATION=/data/oracle/oraInventory" LOCAL_NODE=racnode3
Starting Oracle Universal Installer...

Checking installer requirements...

Checking operating system version: must be 5.8, 5.9 or 5.10.    Actual 5.10
                                      Passed

Checking Temp space: must be greater than 250 MB.   Actual 10167 MB    Passed
Checking swap space: must be greater than 500 MB.   Actual 7415 MB    Passed
Checking monitor: must be configured to display at least 256 colors.    Actual 16777216    Passed

All installer requirements met.

'UpdateNodeList' was successful.

隨後需要依次執行下面的指令碼:

1.節點3上的/data/oracle/oraInventory/orainstRoot.sh

bash-3.00# ./data/oracle/oraInventory/orainstRoot.sh
Creating the Oracle inventory pointer file (/var/opt/oracle/oraInst.loc)
Changing permissions of /data/oracle/oraInventory to 770.
Changing groupname of /data/oracle/oraInventory to oinstall.
The execution of the script. is complete

2.節點1上的/data/oracle/product/10.2/crs/install/rootaddnode.sh

# ./data/oracle/product/10.2/crs/install/rootaddnode.sh
clscfg: EXISTING configuration version 3 detected.
clscfg: version 3 is 10G Release 2.
Attempting to add 1 new nodes to the configuration
Using ports: CSS=49895 CRS=49896 EVMC=49898 and EVMR=49897.
node :
node 3: racnode3 racnode3-priv racnode3
Creating OCR keys for user 'root', privgrp 'other'..
Operation successful.
/data/oracle/product/10.2/crs/bin/srvctl add nodeapps -n racnode3 -A racnode3-vip/255.255.255.0/ce0 -o /data/oracle/product/10.2/crs

3.節點3上的/data/oracle/product/10.2/crs/root.sh

bash-3.00# ./data/oracle/product/10.2/crs/root.sh
WARNING: directory '/data/oracle/product/10.2' is not owned by root
WARNING: directory '/data/oracle/product' is not owned by root
WARNING: directory '/data/oracle' is not owned by root
WARNING: directory '/data' is not owned by root
Checking to see if Oracle CRS stack is already configured
OCR LOCATIONS =  /dev/rac/ocr
OCR backup directory '/data/oracle/product/10.2/crs/cdata/crs' does not exist. Creating now
Setting the permissions on OCR backup directory
Setting up NS directories
Oracle Cluster Registry configuration upgraded successfully
WARNING: directory '/data/oracle/product/10.2' is not owned by root
WARNING: directory '/data/oracle/product' is not owned by root
WARNING: directory '/data/oracle' is not owned by root
WARNING: directory '/data' is not owned by root
clscfg: EXISTING configuration version 3 detected.
clscfg: version 3 is 10G Release 2.
Successfully accumulated necessary OCR keys.
Using ports: CSS=49895 CRS=49896 EVMC=49898 and EVMR=49897.
node :
node 1: racnode1 racnode1-priv racnode1
node 2: racnode2 racnode2-priv racnode2
clscfg: Arguments check out successfully.

NO KEYS WERE WRITTEN. Supply -force parameter to override.
-force is destructive and will destroy any previous cluster
configuration.
Oracle Cluster Registry for cluster has already been initialized
Startup will be queued to init within 30 seconds.
Adding daemons to inittab
Expecting the CRS daemons to be up within 600 seconds.
CSS is active on these nodes.
        racnode1
        racnode2
        racnode3
CSS is active on all nodes.
Waiting for the Oracle CRSD and EVMD to start
Oracle CRS stack installed and running under init(1M)
Running vipca(silent) for configuring nodeapps
IP address "racnode1-vip" has already been used. Enter an unused IP address.

由於vip的配置出現了錯誤,需要手工啟動vipca來進行配置,在圖形化介面用root執行/data/oracle/product/10.2/crs/bin/vipca,輸入racnode3對應的VIP名稱和ip地址:racnode3-vip172.25.198.227,然後執行確定。

成功配置後,在新增節點的圖形化工具addNode中點選繼續。

至此,cluster node節點新增成功。

 

 

來自 “ ITPUB部落格 ” ,連結:http://blog.itpub.net/4227/viewspace-619786/,如需轉載,請註明出處,否則將追究法律責任。

相關文章