RedHat 5.8 安裝Oracle 11gR2_Grid叢集
RedHat 5.8 安裝Oracle 11gR2_Grid叢集
1、 透過rpm安裝相應軟體包:
rpm -ivh binutils-2*.i386.rpm
rpm -ivh binutils-2*.x86_64.rpm
rpm -ivh compat-libstdc++-33-*i*
rpm -ivh compat-libstdc++-33-x*
rpm -ivh elfutils-libelf-0*.i386.rpm
rpm -ivh elfutils-libelf-0*.x*
rpm -ivh elfutils-libelf-devel-0*.i386.rpm
rpm -ivh elfutils-libelf-devel-0.137-3.el5.x*
rpm -ivh compat-libstdc++-2*.i*.rpm
rpm -ivh compat-libstdc++-3*.i*.rpm
rpm -ivh compat-libstdc++-3*.x*.rpm
rpm -ivh gcc-4*.i*
rpm -ivh gcc-4*.x*
rpm -ivh gcc-4.1.2-52.el5.i386.rpm
rpm -ivh gcc-c++-*.x*.rpm
rpm -ivh gcc-c++-*.i*.rpm
rpm -ivh glibc-2.5-81.i*.rpm
rpm -ivh glibc-2.5-81.x*.rpm
rpm -ivh glibc-common-*.i*.rpm
rpm -ivh glibc-common-*.x*.rpm
rpm -ivh glibc-devel*.i*.rpm
rpm -ivh glibc-devel*.x*.rpm
rpm -ivh glibc-headers*.i*.rpm
rpm -ivh glibc-headers*.x*.rpm
rpm -ivh ksh-*.x*.rpm
rpm -ivh ksh-*.i*.rpm
rpm -ivh pdksh-*.i*.rpm
rpm -ivh pdksh-*.x*.rpm
rpm -ivh libaio-0*.i*.rpm
rpm -ivh libaio-0*.x*.rpm
rpm -ivh libaio-devel*.i*.rpm
rpm -ivh libaio-devel*.x*.rpm
rpm -ivh libgcc-4*.i*.rpm
rpm -ivh libgcc-4*.x*.rpm
rpm -ivh libstdc++-4*.i*.rpm
rpm -ivh libstdc++-4*.x*.rpm
rpm -ivh libstdc++-devel*.i*.rpm
rpm -ivh libstdc++-devel*.x*.rpm
rpm -ivh make-3*.i*.rpm
rpm -ivh make-3*.x*.rpm
rpm -ivh sysstat-*.i*.rpm
rpm -ivh sysstat-*.x*.rpm
rpm -ivh unixODBC-libs-*.i*.rpm
rpm -ivh unixODBC-libs-*.x*.rpm
rpm -ivh unixODBC-2*.i*.rpm
rpm -ivh unixODBC-2*.x*.rpm
rpm -ivh unixODBC-devel*.i*.rpm
rpm -ivh unixODBC-devel*.x*.rpm
--DNS相關軟體包
rpm -ivh bind-9*.i*.rpm
rpm -ivh bind-9*.x*.rpm
rpm -ivh bind-utils*.i*.rpm
rpm -ivh bind-utils*.x*.rpm
rpm -ivh caching-nameserver*.i*.rp
rpm -ivh caching-nameserver*.x*.rp
rpm -ivh system-config-bind*.i*.rpm
rpm -ivh system-config-bind*.rpm
rpm -ivh bind-chroot*.i*.rpm
rpm -ivh bind-chroot*.x*.rpm
--ASM包安裝 ---需下載
rpm -ivh oracleasm-support-2.1.7-1.el5.i386.rpm
rpm -ivh oracleasm-2.6.18-308.el5-2.0.5-1.el5.i686.rpm
rpm -ivh oracleasmlib-2.0.4-1.el5.i386.rpm
2、修改主機名
[root@localhost Server]# vi /etc/sysconfig/network
NETWORKING=yes
NETWORKING_IPV6=no
HOSTNAME=grid01/grid02
3、配置host檔案
[root@grid01 ~]# cat /etc/hosts
# Do not remove the following line, or variousprograms
# that require network functionality will fail.
127.0.0.1 localhost.localdomain localhost
#public
172.168.9.15 grid01.prudentwoo.com grid01
172.168.9.16 grid02.prudentwoo.com grid02
#private
10.10.10.1 pri01.prudentwoo.com pri01
10.10.10.2 pri02.prudentwoo.com pri02
#virtual
172.168.9.21 vip01.prudentwoo.com vip01
172.168.9.22 vip02.prudentwoo.com vip02
#scan
172.168.9.17 scan.prudentwoo.com scan
172.168.9.18 scan.prudentwoo.com scan
172.168.9.19 scan.prudentwoo.com scan
172.168.9.20 scan.prudentwoo.com scan
4、配置DNS:
4.1、進入/var/named/chroot/etc目錄中
[root@rac01 chroot]# cd /var/named/chroot/etc
檢視當前目錄下的檔案,將named.caching-nameserver.conf複製一份為named.conf
[root@rac01 etc]# ls
localtime named.caching-nameserver.conf named.rfc1912.zones rndc.key
[root@rac01 etc]# cp –p named.caching-nameserver.confnamed.conf
[root@rac01 etc]# ls
localtime named.caching-nameserver.conf named.conf named.rfc1912.zones rndc.key
建立快捷方式到/etc/目錄下
[root@localhost etc]# ln -s/var/named/chroot/etc/named.conf /etc/named.conf
4.2編輯named.conf檔案,並進進行修改,陰影部分為修改部分:
[root@grid01 ~]# ll /etc/named.conf
lrwxrwxrwx 1 root root 32 Dec 2422:12 /etc/named.conf -> /var/named/chroot/etc/named.conf
[root@rac01etc]# vi /etc/name.conf
options{
listen-on port 53 { any; };
listen-on-v6 port 53 { ::1; };
directory "/var/named";
dump-file "/var/named/data/cache_dump.db";
statistics-file"/var/named/data/named_stats.txt";
memstatistics-file "/var/named/data/named_mem_stats.txt";
// Those options should be usedcarefully because they disable port
// randomization
// query-source port 53;
// query-source-v6 port 53;
allow-query { any; };
allow-query-cache { any; };
};
logging{
channel default_debug {
file"data/named.run";
severity dynamic;
};
};
viewlocalhost_resolver {
match-clients { any; };
match-destinations { any; };
recursion yes;
include "/etc/named.rfc1912.zones";
};
4.3 編輯/etc/named.rfc1912.zones新增如下內容
zone"prudentwoo.com" IN {
type master;
file "prudentwoo.com.zone";
allow-update { none; };
};
zone"9.168.172.in-addr.arpa" IN {
type master;
file "172.168.9.zone";
allow-update { none; };
allow-transfer {
172.168.9.17;
172.168.9.18;
172.168.9.19;
172.168.9.20;
};
};
4.4 編輯正向解析檔案:
[root@grid01named]# cd /var/named/chroot/var/named
[root@rac01 named]# cp–p localdomain.zone prudentwoo.com.zone
[root@grid01named]# cp –p named.local 172.168.9.zone
[root@node1named]# vi prudentwoo.zone
$TTL 86400
@ IN SOA prudentwoo.com. root.prudentwoo.com. (
42 ; serial (d. adams)
3H ; refresh
15M ; retry
1W ; expiry
1D) ; minimum
IN NS prudentwoo.com
scan IN A 172.168.9.17
scan IN A 172.168.9.18
scan IN A 172.168.9.19
scan IN A 172.168.9.20
4.5編輯反向解析檔案:
[root@node1 named]# more 172.168.10.zone
$TTL 86400
@ IN SOA 9.168.172.in-addr.arpa. root.prudentwoo.com. (
1997022700 ; Serial
28800 ; Refresh
14400 ; Retry
3600000 ; Expire
86400) ; Minimum
@ IN NS prudentwoo.com.
17 IN PTR scan.prudentwoo.com.
18 IN PTR scan.prudentwoo.com.
19 IN PTR scan.prudentwoo.com.
20 IN PTR scan.prudentwoo.com.
4.6 新增DNS伺服器地址,兩個節點都要配置,節點二nameserver 反過來配置
[root@rac01 ~]# vi /etc/resolv.conf
#search localdomain
search prudentwoo.com
nameserver 172.168.9.15
nameserver 172.168.9.16
4.7 建立連線檔案
[root@grid01 named]# ln -s/var/named/chroot/var/named/172.168.9.zone /var/named/172.168.9.zone
[root@grid01 named]# ln -s/var/named/chroot/var/named/prudentwoo.comn.zone /var/named/prudentwoo.comn.zone
[root@grid01 named]# ll /var/named/
total 16
lrwxrwxrwx 1 root root 42 Dec 24 21:19172.168.9.zone -> /var/named/chroot/var/named/172.168.9.zone
drwxr-x--- 6 root named 4096 Dec 24 19:38 chroot
drwxrwx--- 2 named named 4096 Dec 2 2011data
lrwxrwxrwx 1 root named 44 Dec 24 19:19localdomain.zone -> /var/named/chroot/var/named/localdomain.zone
lrwxrwxrwx 1 root named 42 Dec 24 19:19localhost.zone -> /var/named/chroot/var/named/localhost.zone
lrwxrwxrwx 1 root named 43 Dec 24 19:19named.broadcast -> /var/named/chroot/var/named/named.broadcast
lrwxrwxrwx 1 root named 36 Dec 24 19:19 named.ca-> /var/named/chroot/var/named/named.ca
-rw-r----- 1 root root 1206 Dec 24 20:55 named.conf
lrwxrwxrwx 1 root named 43 Dec 24 19:19named.ip6.local -> /var/named/chroot/var/named/named.ip6.local
lrwxrwxrwx 1 root named 39 Dec 24 19:19named.local -> /var/named/chroot/var/named/named.local
lrwxrwxrwx 1 root named 38 Dec 24 19:19 named.zero-> /var/named/chroot/var/named/named.zero
lrwxrwxrwx1 root root 47 Dec 24 21:20 prudentwoo.com.zone ->/var/named/chroot/var/named/prudentwoo.com.zone
drwxrwx---2 named named 4096 Dec 2 2011 slaves
-----4.8重啟DNS服務
1、DNS出現如下資訊無法起來,檢查message日誌:
[root@rac01 log]# service named restart
Stopping named: [ OK ]
Starting named: [FAILED]
[root@rac01 log]# cd /var/log
[root@rac01 log]# pwd
/var/log
[root@rac01 log]# tail -f messages
Nov 100:46:01 localhost named[29489]: loading configuration: permission denied
Nov 100:46:01 localhost named[29489]: exiting (due to fatal error)
Nov 100:48:58 localhost named[29564]: starting BIND 9.3.6-P1-RedHat-9.3.6-20.P1.el5-u named -t /var/named/chroot
Nov 100:48:58 localhost named[29564]: adjusted limit on open files from 1024 to1048576
Nov 1 00:48:58localhost named[29564]: found 1 CPU, using 1 worker thread
Nov 100:48:58 localhost named[29564]: using up to 4096 sockets
Nov 100:48:58 localhost named[29564]: loading configuration from '/etc/named.conf'
Nov 100:48:58 localhost named[29564]: none:0: open: /etc/named.conf: permissiondenied
Nov 100:48:58 localhost named[29564]: loading configuration: permission denied
Nov 100:48:58 localhost named[29564]: exiting (due to fatal error)
修改named.conf許可權,然後再重啟:
[root@rac01 log]# ls -al/var/named/chroot/etc/named.conf
-rw-r----- 1 root root 1206 Nov 1 00:07 /var/named/chroot/etc/named.conf
[root@rac01 log]# chown root:named/var/named/chroot/etc/named.c
named.caching-nameserver.conf named.conf
[root@rac01 log]# chown root:named/var/named/chroot/etc/named.conf
4.8重啟:
[root@rac01 log]# service named restart
Stopping named: [ OK ]
Starting named: [ OK ]
[root@rac01 log]#
4.9 驗證DNS,至此發現DNS出現輪詢就OK了。
[root@rac01 named]# nslookupnode-scan.prudentwoo.com
Server: 172.168.9.15
Address: 172.168.9.15#53
Name: node-scan.prudentwoo.com
Address: 172.168.9.21
Name: node-scan.prudentwoo.com
Address: 172.168.9.22
Name: node-scan.prudentwoo.com
Address: 172.168.9.23
Name: node-scan.prudentwoo.com
Address: 172.168.9.24
Name: node-scan.prudentwoo.com
Address: 172.168.9.20
4.10 反向解析驗證DNS:
[root@grid01 named]# nslookup 172.168.9.15
Server: 172.168.9.15
Address: 172.168.9.15#53
** server can't find 15.9.168.172.in-addr.arpa.:NXDOMAIN
You have new mail in /var/spool/mail/root
[root@grid01 named]# nslookup 172.168.9.16
Server: 172.168.9.15
Address: 172.168.9.15#53
** server can't find 16.9.168.172.in-addr.arpa.:NXDOMAIN
[root@grid01 named]# nslookup 172.168.9.17
Server: 172.168.9.15
Address: 172.168.9.15#53
17.9.168.172.in-addr.arpa name = scan.prudentwoo.com.
[root@grid01 named]# nslookup 172.168.9.18
Server: 172.168.9.15
Address: 172.168.9.15#53
18.9.168.172.in-addr.arpa name = scan.prudentwoo.com.
[root@grid01 named]# nslookup 172.168.9.19
Server: 172.168.9.15
Address: 172.168.9.15#53
19.9.168.172.in-addr.arpa name = scan.prudentwoo.com.
[root@grid01 named]# nslookup 172.168.9.20
5、新增NTP和DNS服務自啟動
Ntsysv
或 chkconfig namedon chkchonfig ntpd on
檢查自啟動服務狀態
chkconfig --list
6、建立使用者組;
/usr/sbin/groupadd -g 502 dba
/usr/sbin/groupadd -g 501 oinstall
/usr/sbin/groupadd -g 504 asmadmin
/usr/sbin/groupadd -g 506 asmdba
/usr/sbin/groupadd -g 507 asmoper
/usr/sbin/useradd -u 501 -g oinstall -Gasmadmin,asmdba,asmoper grid
/usr/sbin/useradd -u 502 -g oinstall -G dba,asmdbaoracle
[root@rac01 named]# passwd grid
[root@rac01 named]# passwd oracle
[root@rac01 named]#
7、建立目錄:
Oracle清單目錄
mkdir -p /app/oraInventory
chown -R grid:oinstall /app/oraInventory
chmod -R 775 /app/oraInventory
Grid安裝目錄
mkdir -p /app/grid
mkdir -p /app/11.2.0/grid
chown -R grid:oinstall /app/11.2.0/grid
chmod -R 775 /app/11.2.0/grid
建立Oracle安裝目錄
mkdir -p /app/oracle/product/11.2.0/db_1
mkdir -p/app/oracle/cfgtoollogs
chown -R oracle:oinstall /app/oracle
chmod -R 775 /app/oracle
8、修改系統引數,在末端新增如下內容
8.1 修改記憶體引數:
vi /etc/sysctl.conf
kernel.shmall = 2097152
kernel.sem = 250 32000 100 128
kernel.shmmni = 4096
#kernel.shmax = 8589934592
fs.file-max = 6815744
net.ipv4.ip_local_port_range = 9000 65500
net.core.rmem_default = 262144
net.core.rmem_max = 4194304
net.core.wmem_default = 1048576
net.core.wmem_max = 1048576
fs.aio-max-nr = 1048576
8.2 修改限制引數:
編輯/etc/security/limits.conf
* soft nproc 2047
* hard nproc 16384
* soft nofile 1024
* hard nofile 65536
* soft memlock 3145728
* hard memlock 3145728
8.3 修改Oracle使用者環境變數,新增如下資訊
[root@rac01 ~]# su - oracle
[oracle@rac01 ~]$ vi .bash_profile
#.bash_profile
# Getthe aliases and functions
if [ -f~/.bashrc ]; then
. ~/.bashrc
fi
# Userspecific environment and startup programs
PATH=$PATH:$HOME/bin
exportPATH
#OracleSettings
TMP=/tmp;export TMP
exportTMPDIR=$TMP;
exportORACLE_HOSTNAME=grid01;
exportORACLE_UNQNAME=woo;
exportORACLE_BASE=/app/oracle;
exportORACLE_HOME=$ORACLE_BASE/product/11.2.0/db_1;
exportORACLE_SID=woo1;
exportORACLE_TERM=xterm;
exportPATH=/usr/sbin:$PATH;
exportPATH=$ORACLE_HOME/bin:$PATH;
exportLD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib;
exportCLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib;
umask022
8.4 修改grid使用者環境變數,新增如下資訊:
[root@rac01 ~]# su - grid
[grid@rac01 ~]$ vi .bash_profile
#.bash_profile
# Getthe aliases and functions
if [ -f~/.bashrc ]; then
. ~/.bashrc
fi
# Userspecific environment and startup programs
PATH=$PATH:$HOME/bin
exportPATH
exportTMP=/tmp;
exportTMPDIR=$TMP;
exportGRID_SID=+ASM1;
export ORACLE_BASE=/app/grid;
export ORACLE_HOME=/app/11.2.0/grid;
exportPATH=$GRID_HOME/bin:$PATH;
umask022
8.5 編輯/etc/pam.d/login,新增如下內容:
Session required pam_limits.so
編輯/etc/profile,新增如下內容:
[root@rac01 ~]# vi /etc/profile
if [ $USER = "oracle" ]||[ $USER = "grid" ] ;then
if [ $SHELL ="/bin/ksh" ]; then
ulimit -p16384
ulimit -n65536
else
ulimit -u16384 -n 65536
fi
umask 022
fi
9、停ntp時間同步(11G新增檢查選項)
/etc/init.d/ntpd stop
chkconfig ntpd off
mv /etc/ntp.conf /etc/ntp.conf.org
rm /etc/ntp.conf
如果要使用NTP,需要對NTP配置檔案引數進行修改:
編輯/etc/sysconfig/ntpd,修改如下:
OPTIONS=”-x–u ntp:ntp–p /var/run/ntpd.pid”
然後重啟ntp服務:
修改DNS主機搜尋順序,將DNS放在優先位置:
[root@rac01 named]# vi /etc/nsswitch.conf
hosts: dns files nis
修改後重啟服務:
[root@rac01 named]# service nscd restart
Stopping nscd: [FAILED]
Starting nscd: [ OK ]
建立磁碟分割槽
OCR 8G
DATA 20G
FRA 20G
[root@rac01 ~]# fdisk -l
Disk /dev/sda: 21.4 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
DeviceBoot Start End Blocks Id System
/dev/sda1 * 1 38 305203+ 83 Linux
/dev/sda2 39 1201 9341797+ 83 Linux
/dev/sda3 1202 2245 8385930 83 Linux
/dev/sda4 2246 2610 2931862+ 5 Extended
/dev/sda5 2246 2506 2096451 82 Linux swap / Solaris
Disk /dev/sdb: 8589 MB, 8589934592 bytes
255 heads, 63 sectors/track, 1044 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk /dev/sdb doesn't contain a valid partitiontable
Disk /dev/sdc: 21.4 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk /dev/sdc doesn't contain a valid partitiontable
Disk /dev/sdd: 21.4 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk /dev/sdd doesn't contain a valid partitiontable
[root@rac01~]# fdisk /dev/sdb
Device contains neither a valid DOS partitiontable, nor Sun, SGI or OSF disklabel
Building a new DOS disklabel. Changes will remainin memory only,
until you decide to write them. After that, ofcourse, the previous
content won't be recoverable.
Command (m for help): n
Command action
e extended
p primary partition (1-4)
p
Partition number (1-4): 1
First cylinder (1-1044, default 1):
Using default value 1
Last cylinder or +size or +sizeM or +sizeK (1-1044,default 1044):
Using default value 1044
Command (m for help): w
The partition table has been altered!
Calling ioctl() to re-read partition table.
Syncing disks.
[root@rac01~]# fdisk /dev/sdc
Device contains neither a valid DOS partitiontable, nor Sun, SGI or OSF disklabel
Building a new DOS disklabel. Changes will remainin memory only,
until you decide to write them. After that, ofcourse, the previous
content won't be recoverable.
The number of cylinders for this disk is set to2610.
There is nothing wrong with that, but this islarger than 1024,
and could in certain setups cause problems with:
1) software that runs at boot time (e.g., oldversions of LILO)
2) booting and partitioning software from other OSs
(e.g., DOSFDISK, OS/2 FDISK)
Warning: invalid flag 0x0000 of partition table 4will be corrected by w(rite)
Command (m for help): n
Command action
e extended
p primary partition (1-4)
p
Partition number (1-4): 1
First cylinder (1-2610, default 1):
Using default value 1
Last cylinder or +size or +sizeM or +sizeK (1-2610,default 2610):
Using default value 2610
Command (m for help): w
The partition table has been altered!
Calling ioctl() to re-read partition table.
Syncing disks.
[root@rac01 ~]# fdisk /dev/sdd
Device contains neither a valid DOS partitiontable, nor Sun, SGI or OSF disklabel
Building a new DOS disklabel. Changes will remainin memory only,
until you decide to write them. After that, ofcourse, the previous
content won't be recoverable.
The number of cylinders for this disk is set to2610.
There is nothing wrong with that, but this islarger than 1024,
and could in certain setups cause problems with:
1) software that runs at boot time (e.g., oldversions of LILO)
2) booting and partitioning software from other OSs
(e.g., DOSFDISK, OS/2 FDISK)
Warning: invalid flag 0x0000 of partition table 4will be corrected by w(rite)
Command (m for help): n
Command action
e extended
p primary partition (1-4)
p
Partition number (1-4): 1
First cylinder (1-2610, default 1):
Using default value 1
Last cylinder or +size or +sizeM or +sizeK (1-2610,default 2610):
Using default value 2610
Command (m for help): w
The partition table has been altered!
Calling ioctl() to re-read partition table.
Syncing disks.
[root@rac01 ~]# fdisk -l
Disk /dev/sda: 21.4 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
DeviceBoot Start End Blocks Id System
/dev/sda1 * 1 38 305203+ 83 Linux
/dev/sda2 39 1201 9341797+ 83 Linux
/dev/sda3 1202 2245 8385930 83 Linux
/dev/sda4 2246 2610 2931862+ 5 Extended
/dev/sda5 2246 2506 2096451 82 Linux swap / Solaris
Disk /dev/sdb: 8589 MB, 8589934592 bytes
255 heads, 63 sectors/track, 1044 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
DeviceBoot Start End Blocks Id System
/dev/sdb1 1 1044 8385898+ 83 Linux
Disk /dev/sdc: 21.4 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
DeviceBoot Start End Blocks Id System
/dev/sdc1 1 2610 20964793+ 83 Linux
Disk /dev/sdd: 21.4 GB, 21474836480 bytes
255 heads, 63 sectors/track, 2610 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
DeviceBoot Start End Blocks Id System
/dev/sdd1 1 2610 20964793+ 83 Linux
[root@rac01 ~]#
安裝ASM
[root@grid02 asm]# ls
oracleasm-2.6.18-308.el5-2.0.5-1.el5.i686.rpm oracleasmlib-2.0.4-1.el5.i386.rpm oracleasm-support-2.1.7-1.el5.i386.rpm
[root@grid02 asm]# rpm -ivhoracleasm-support-2.1.7-1.el5.i386.rpm
warning: oracleasm-support-2.1.7-1.el5.i386.rpm:Header V3 DSA signature: NOKEY, key ID 1e5e0159
Preparing... ########################################### [100%]
1:oracleasm-support ########################################### [100%]
[root@grid02 asm]# rpm -ivh oracleasm-2.6.18-308.el5-2.0.5-1.el5.i686.rpm
warning:oracleasm-2.6.18-308.el5-2.0.5-1.el5.i686.rpm: Header V3 DSA signature: NOKEY,key ID 1e5e0159
Preparing... ########################################### [100%]
1:oracleasm-2.6.18-308.el###########################################[100%]
[root@grid02 asm]# rpm -ivhoracleasmlib-2.0.4-1.el5.i386.rpm
warning: oracleasmlib-2.0.4-1.el5.i386.rpm: HeaderV3 DSA signature: NOKEY, key ID 1e5e0159
Preparing... ###########################################[100%]
1:oracleasmlib ########################################### [100%]
配置ASMLib
[root@grid01 grid]# /etc/init.d/oracleasm configure
Configuring the Oracle ASM library driver.
This will configure the on-boot properties of theOracle ASM library
driver. Thefollowing questions will determine whether the driver is
loaded on boot and what permissions it willhave. The current values
will be shown in brackets ('[]'). Hitting
answer will keep that current value. Ctrl-C will abort.
Default user to own the driver interface []: grid
Default group to own the driver interface []: asmadmin
Start Oracle ASM library driver on boot (y/n) [n]: y
Scan for Oracle ASM disks on boot (y/n) [y]: y
Writing Oracle ASM library driver configuration:done
Initializing the Oracle ASMLib driver: [ OK ]
Scanning the system for Oracle ASMLib disks: [ OK ]
建立ASM磁碟
/usr/sbin/oracleasm createdisk OCR_VOTE1 /dev/sdb1
/usr/sbin/oracleasm createdisk OCR_VOTE2 /dev/sdc1
/usr/sbin/oracleasm createdisk OCR_VOTE3 /dev/sdd1
/usr/sbin/oracleasm createdisk DATA001 /dev/sde1
/usr/sbin/oracleasm createdisk ARC001 /dev/sdf1
/usr/sbin/oracleasm createdisk FRA001 /dev/sdg1
另一節點執行發現磁碟
[root@grid02 ~]# /etc/init.d/oracleasm scandisks
Scanning the system for Oracle ASMLib disks: [ OK ]
[root@grid02 ~]# /etc/init.d/oracleasm listdisks
ARC001
DATA001
FRA001
OCR_VOTE1
OCR_VOTE2
OCR_VOTE3
驗證安裝前的檢查:
./runcluvfy.sh stage -pre crsinst -n rac01,rac02-fixup–verbose
./runcluvfy.shcomp nodecon -n rac01,rac02 –verbose
安裝完成後執行root指令碼
[root@grid01 /]#/app/oraInventory/orainstRoot.sh
Changing permissions of /app/oraInventory.
Adding read,write permissions for group.
Removing read,write,execute permissions for world.
Changing groupname of /app/oraInventory tooinstall.
The execution of the script. is complete.
[root@grid01 /]# /app/11.2.0/grid/root.sh
Running Oracle 11g root.sh script...
The following environment variables are set as:
ORACLE_OWNER= grid
ORACLE_HOME= /app/11.2.0/grid
Enter the full pathname of the local bin directory:[/usr/local/bin]:
Copyingdbhome to /usr/local/bin ...
Copyingoraenv to /usr/local/bin ...
Copyingcoraenv to /usr/local/bin ...
Creating /etc/oratab file...
Entries will be added to the /etc/oratab file asneeded by
Database Configuration Assistant when a database iscreated
Finished running generic part of root.sh script.
Now product-specific root actions will beperformed.
2012-12-26 23:38:51: Parsing the host name
2012-12-26 23:38:51: Checking for super userprivileges
2012-12-26 23:38:51: User has super user privileges
Using configuration parameter file:/app/11.2.0/grid/crs/install/crsconfig_params
Creating trace directory
LOCAL ADD MODE
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
root wallet
root walletcert
root certexport
peer wallet
profilereader wallet
pa wallet
peer walletkeys
pa walletkeys
peer certrequest
pa certrequest
peer cert
pa cert
peer rootcert TP
profilereader root cert TP
pa rootcert TP
peer pacert TP
pa peercert TP
profilereader pa cert TP
profile readerpeer cert TP
peer usercert
pa usercert
Adding daemon to inittab
CRS-4123: Oracle High Availability Services hasbeen started.
ohasd is starting
CRS-2672: Attempting to start 'ora.gipcd' on'grid01'
CRS-2672: Attempting to start 'ora.mdnsd' on 'grid01'
CRS-2676: Start of 'ora.mdnsd' on 'grid01'succeeded
CRS-2676: Start of 'ora.gipcd' on 'grid01'succeeded
CRS-2672: Attempting to start 'ora.gpnpd' on'grid01'
CRS-2676: Start of 'ora.gpnpd' on 'grid01'succeeded
CRS-2672: Attempting to start 'ora.cssdmonitor' on'grid01'
CRS-2676: Start of 'ora.cssdmonitor' on 'grid01'succeeded
CRS-2672: Attempting to start 'ora.cssd' on'grid01'
CRS-2672: Attempting to start 'ora.diskmon' on'grid01'
CRS-2676: Start of 'ora.diskmon' on 'grid01'succeeded
CRS-2676: Start of 'ora.cssd' on 'grid01' succeeded
CRS-2672: Attempting to start 'ora.ctssd' on'grid01'
CRS-2676: Start of 'ora.ctssd' on 'grid01'succeeded
ASM created and started successfully.
DiskGroup OCR_VOTE created successfully.
clscfg: -install mode specified
Successfully accumulated necessary OCR keys.
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
CRS-2672: Attempting to start 'ora.crsd' on'grid01'
CRS-2676: Start of 'ora.crsd' on 'grid01' succeeded
CRS-4256: Updating the profile
Successful addition of voting disk2b34943383f64f00bf2e91653dd35ff4.
Successful addition of voting disk2c1bbab1e3054f81bf528e9630333192.
Successful addition of voting disk88252b92194c4f62bfe070bb314bd171.
Successfully replaced voting disk group with+OCR_VOTE.
CRS-4256: Updating the profile
CRS-4266: Voting file(s) successfully replaced
## STATE File Universal Id File Name Disk group
-- ----- ----------------- --------- ---------
1.ONLINE 2b34943383f64f00bf2e91653dd35ff4(ORCL:OCR_VOTE3) [OCR_VOTE]
2.ONLINE 2c1bbab1e3054f81bf528e9630333192(ORCL:OCR_VOTE2) [OCR_VOTE]
3.ONLINE 88252b92194c4f62bfe070bb314bd171(ORCL:OCR_VOTE1) [OCR_VOTE]
Located 3 voting disk(s).
CRS-2673: Attempting to stop 'ora.crsd' on 'grid01'
CRS-2677: Stop of 'ora.crsd' on 'grid01' succeeded
CRS-2673: Attempting to stop 'ora.asm' on 'grid01'
CRS-2677: Stop of 'ora.asm' on 'grid01' succeeded
CRS-2673: Attempting to stop 'ora.ctssd' on'grid01'
CRS-2677: Stop of 'ora.ctssd' on 'grid01' succeeded
CRS-2673: Attempting to stop 'ora.cssdmonitor' on'grid01'
CRS-2677: Stop of 'ora.cssdmonitor' on 'grid01'succeeded
CRS-2673: Attempting to stop 'ora.cssd' on 'grid01'
CRS-2677: Stop of 'ora.cssd' on 'grid01' succeeded
CRS-2673: Attempting to stop 'ora.gpnpd' on'grid01'
CRS-2677: Stop of 'ora.gpnpd' on 'grid01' succeeded
CRS-2673: Attempting to stop 'ora.gipcd' on'grid01'
CRS-2677: Stop of 'ora.gipcd' on 'grid01' succeeded
CRS-2673: Attempting to stop 'ora.mdnsd' on'grid01'
CRS-2677: Stop of 'ora.mdnsd' on 'grid01' succeeded
CRS-2672: Attempting to start 'ora.mdnsd' on'grid01'
CRS-2676: Start of 'ora.mdnsd' on 'grid01'succeeded
CRS-2672: Attempting to start 'ora.gipcd' on'grid01'
CRS-2676: Start of 'ora.gipcd' on 'grid01'succeeded
CRS-2672: Attempting to start 'ora.gpnpd' on'grid01'
CRS-2676: Start of 'ora.gpnpd' on 'grid01'succeeded
CRS-2672: Attempting to start 'ora.cssdmonitor' on'grid01'
CRS-2676: Start of 'ora.cssdmonitor' on 'grid01'succeeded
CRS-2672: Attempting to start 'ora.cssd' on'grid01'
CRS-2672: Attempting to start 'ora.diskmon' on'grid01'
CRS-2676: Start of 'ora.diskmon' on 'grid01'succeeded
CRS-2676: Start of 'ora.cssd' on 'grid01' succeeded
CRS-2672: Attempting to start 'ora.ctssd' on'grid01'
CRS-2676: Start of 'ora.ctssd' on 'grid01'succeeded
CRS-2672: Attempting to start 'ora.asm' on 'grid01'
CRS-2676: Start of 'ora.asm' on 'grid01' succeeded
CRS-2672: Attempting to start 'ora.crsd' on'grid01'
CRS-2676: Start of 'ora.crsd' on 'grid01' succeeded
CRS-2672: Attempting to start 'ora.evmd' on'grid01'
CRS-2676: Start of 'ora.evmd' on 'grid01' succeeded
CRS-2672: Attempting to start 'ora.asm' on 'grid01'
CRS-2676: Start of 'ora.asm' on 'grid01' succeeded
CRS-2672: Attempting to start 'ora.OCR_VOTE.dg' on'grid01'
CRS-2676: Start of 'ora.OCR_VOTE.dg' on 'grid01'succeeded
CRS-2672: Attempting to start 'ora.registry.acfs'on 'grid01'
CRS-2676: Start of 'ora.registry.acfs' on 'grid01'succeeded
grid01 2012/12/26 23:49:34 /app/11.2.0/grid/cdata/grid01/backup_20121226_234934.olr
Preparing packages for installation...
cvuqdisk-1.0.7-1
Configure Oracle Grid Infrastructure for a Cluster... succeeded
Updating inventory properties for clusterware
Starting Oracle Universal Installer...
Checking swap space: must be greater than 500MB. Actual 2047 MB Passed
The inventory pointer is located at/etc/oraInst.loc
The inventory is located at /app/oraInventory
'UpdateNodeList' wassuccessful.
節點二執行如下:
[root@grid02 ~]#/app/oraInventory/orainstRoot.sh
Changing permissions of /app/oraInventory.
Adding read,write permissions for group.
Removing read,write,execute permissions for world.
Changing groupname of /app/oraInventory tooinstall.
The execution of the script. is complete.
[root@grid02 ~]#/app/11.2.0/grid/root.sh
Running Oracle 11g root.sh script...
The following environment variables are set as:
ORACLE_OWNER= grid
ORACLE_HOME= /app/11.2.0/grid
Enter the full pathname of the local bin directory:[/usr/local/bin]:
Copyingdbhome to /usr/local/bin ...
Copyingoraenv to /usr/local/bin ...
Copyingcoraenv to /usr/local/bin ...
Creating /etc/oratab file...
Entries will be added to the /etc/oratab file asneeded by
Database Configuration Assistant when a database iscreated
Finished running generic part of root.sh script.
Now product-specific root actions will beperformed.
2012-12-26 23:52:45: Parsing the host name
2012-12-26 23:52:45: Checking for super userprivileges
2012-12-26 23:52:45: User has super user privileges
Using configuration parameter file:/app/11.2.0/grid/crs/install/crsconfig_params
Creating trace directory
LOCAL ADD MODE
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
Adding daemon to inittab
CRS-4123: Oracle High Availability Services hasbeen started.
ohasd is starting
CRS-4402: The CSS daemon was started in exclusivemode but found an active CSS daemon on node grid01, number 1, and isterminating
An active cluster was found during exclusivestartup, restarting to join the cluster
CRS-2672: Attempting to start 'ora.mdnsd' on'grid02'
CRS-2676: Start of 'ora.mdnsd' on 'grid02'succeeded
CRS-2672: Attempting to start 'ora.gipcd' on'grid02'
CRS-2676: Start of 'ora.gipcd' on 'grid02'succeeded
CRS-2672: Attempting to start 'ora.gpnpd' on'grid02'
CRS-2676: Start of 'ora.gpnpd' on 'grid02'succeeded
CRS-2672: Attempting to start 'ora.cssdmonitor' on'grid02'
CRS-2676: Start of 'ora.cssdmonitor' on 'grid02'succeeded
CRS-2672: Attempting to start 'ora.cssd' on'grid02'
CRS-2672: Attempting to start 'ora.diskmon' on'grid02'
CRS-2676: Start of 'ora.diskmon' on 'grid02'succeeded
CRS-2676: Start of 'ora.cssd' on 'grid02' succeeded
CRS-2672: Attempting to start 'ora.ctssd' on'grid02'
CRS-2676: Start of 'ora.ctssd' on 'grid02'succeeded
CRS-2672: Attempting to start 'ora.drivers.acfs' on'grid02'
CRS-2676: Start of 'ora.drivers.acfs' on 'grid02'succeeded
CRS-2672: Attempting to start 'ora.asm' on 'grid02'
CRS-2676: Start of 'ora.asm' on 'grid02' succeeded
CRS-2672: Attempting to start 'ora.crsd' on'grid02'
CRS-2676: Start of 'ora.crsd' on 'grid02' succeeded
CRS-2672: Attempting to start 'ora.evmd' on'grid02'
CRS-2676: Start of 'ora.evmd' on 'grid02' succeeded
grid02 2012/12/26 23:58:32 /app/11.2.0/grid/cdata/grid02/backup_20121226_235832.olr
Preparing packages for installation...
cvuqdisk-1.0.7-1
Configure Oracle Grid Infrastructure for a Cluster... succeeded
Updating inventory properties for clusterware
Starting Oracle Universal Installer...
Checking swap space: must be greater than 500MB. Actual 2047 MB Passed
The inventory pointer is located at/etc/oraInst.loc
The inventory is located at /app/oraInventory
'UpdateNodeList' was successful.
[root@grid02 ~]#
檢查狀態,如果有問題,按照如下操作:
[grid@grid01 ~]$ crsctl check cluster -all
在第二個節點執行刪除操作,重新執行root指令碼
/app/11.2.0/grid/crs/install/rootcrs.pl -verbose-deconfig –force
/app/11.2.0/grid/root.sh
檢查狀態:
[grid@grid01 ~]$ crsctl check cluster -all
**************************************************************
grid01:
CRS-4537: Cluster Ready Services is online
CRS-4529: Cluster Synchronization Services isonline
CRS-4533: Event Manager is online
**************************************************************
grid02:
CRS-4537: Cluster Ready Services is online
CRS-4529: Cluster Synchronization Services isonline
CRS-4533: Event Manager isonline
[grid@grid02 ~]$ crs_stat -t
Name Type Target State Host
------------------------------------------------------------
ora....N1.lsnr ora....er.type ONLINE ONLINE grid02
ora....N2.lsnr ora....er.type ONLINE ONLINE grid01
ora....N3.lsnr ora....er.type ONLINE ONLINE grid01
ora....N4.lsnr ora....er.type ONLINE ONLINE grid01
ora....VOTE.dg ora....up.type ONLINE ONLINE grid01
ora.asm ora.asm.type ONLINE ONLINE grid01
ora.eons ora.eons.type ONLINE ONLINE grid01
ora....SM1.asm application ONLINE ONLINE grid01
ora.grid01.gsd application OFFLINE OFFLINE
ora.grid01.ons application ONLINE ONLINE grid01
ora.grid01.vip ora....t1.type ONLINE ONLINE grid01
ora....SM2.asm application ONLINE ONLINE grid02
ora.grid02.gsd application OFFLINE OFFLINE
ora.grid02.ons application ONLINE ONLINE grid02
ora.grid02.vip ora....t1.type ONLINE ONLINE grid02
ora.gsd ora.gsd.type OFFLINE OFFLINE
ora....network ora....rk.type ONLINE ONLINE grid01
ora.oc4j ora.oc4j.type OFFLINE OFFLINE
ora.ons ora.ons.type ONLINE ONLINE grid01
ora....ry.acfs ora....fs.type ONLINE ONLINE grid01
ora.scan1.vip ora....ip.type ONLINE ONLINE grid02
ora.scan2.vip ora....ip.type ONLINE ONLINE grid01
ora.scan3.vip ora....ip.type ONLINE ONLINE grid01
ora.scan4.vip ora....ip.type ONLINE ONLINE grid01
[grid@grid02 ~]$
錯誤檢查:
檢查votdisk 狀態
crsctlquery css votedisk
檢查ocr狀態
Ocrcheck
檢查叢集節點:
olsnodes –n
olsnodes -n -p -l -s -t -v
檢查監聽狀態:
srvctl status listener
檢查asm
srvctl status asm –a
建立資料盤
Asmca
啟動GSD和OC4J服務
檢視狀態
Srvctl status oc4j
Srvctl status nodeapps
開啟自動啟動
srvctl enable oc4j
srvctl enable nodeapps
手工帶起服務
srvctl start nodeapps -n grid01
srvctl start nodeapps –n grid02
可以看到服務已經全部起來:
[grid@grid01 ~]$ crs_stat -t
Name Type Target State Host
------------------------------------------------------------
ora....ER.lsnr ora....er.type ONLINE ONLINE grid01
ora....N1.lsnr ora....er.type ONLINE ONLINE grid02
ora....N2.lsnr ora....er.type ONLINE ONLINE grid01
ora....N3.lsnr ora....er.type ONLINE ONLINE grid01
ora....N4.lsnr ora....er.type ONLINE ONLINE grid01
ora....VOTE.dg ora....up.type ONLINE ONLINE grid01
ora.asm ora.asm.type ONLINE ONLINE grid01
ora.eons ora.eons.type ONLINE ONLINE grid01
ora....SM1.asm application ONLINE ONLINE grid01
ora....01.lsnr application ONLINE ONLINE grid01
ora.grid01.gsd application ONLINE ONLINE grid01
ora.grid01.ons application ONLINE ONLINE grid01
ora.grid01.vip ora....t1.type ONLINE ONLINE grid01
ora....SM2.asm application ONLINE ONLINE grid02
ora....02.lsnr application ONLINE ONLINE grid02
ora.grid02.gsd application ONLINE ONLINE grid02
ora.grid02.ons application ONLINE ONLINE grid02
ora.grid02.vip ora....t1.type ONLINE ONLINE grid02
ora.gsd ora.gsd.type ONLINE ONLINE grid01
ora....network ora....rk.type ONLINE ONLINE grid01
ora.oc4j ora.oc4j.type ONLINE ONLINE grid02
ora.ons ora.ons.type ONLINE ONLINE grid01
ora....ry.acfs ora....fs.type ONLINE ONLINE grid01
ora.scan1.vip ora....ip.type ONLINE ONLINE grid02
ora.scan2.vip ora....ip.type ONLINE ONLINE grid01
ora.scan3.vip ora....ip.type ONLINE ONLINE grid01
ora.scan4.vip ora....ip.type ONLINE ONLINE grid01
[grid@grid01 ~]$
開始安裝資料庫軟體:
安裝完成之後,在兩個節點分別執行root指令碼
[root@grid01 ~]#/app/oracle/product/11.2.0/db_1/root.sh
Running Oracle 11g root.sh script...
The following environment variables are set as:
ORACLE_OWNER= oracle
ORACLE_HOME= /app/oracle/product/11.2.0/db_1
Enter the full pathname of the local bin directory:[/usr/local/bin]:
The file "dbhome" already exists in/usr/local/bin. Overwrite it? (y/n)
[n]: y
Copyingdbhome to /usr/local/bin ...
The file "oraenv" already exists in/usr/local/bin. Overwrite it? (y/n)
[n]: y
Copyingoraenv to /usr/local/bin ...
The file "coraenv" already exists in/usr/local/bin. Overwrite it? (y/n)
[n]: y
Copyingcoraenv to /usr/local/bin ...
Entries will be added to the /etc/oratab file asneeded by
Database Configuration Assistant when a database iscreated
Finished running generic part of root.sh script.
Now product-specific root actions will beperformed.
Finished product-specific root actions.
節點二、
[root@grid02 db_1]#/app/oracle/product/11.2.0/db_1/root.sh
Running Oracle 11g root.sh script...
The following environment variables are set as:
ORACLE_OWNER= oracle
ORACLE_HOME= /app/oracle/product/11.2.0/db_1
Enter the full pathname of the local bin directory:[/usr/local/bin]:
The file "dbhome" already exists in/usr/local/bin. Overwrite it? (y/n)
[n]: y
Copyingdbhome to /usr/local/bin ...
The file "oraenv" already exists in/usr/local/bin. Overwrite it? (y/n)
[n]: y
Copying oraenvto /usr/local/bin ...
The file "coraenv" already exists in/usr/local/bin. Overwrite it? (y/n)
[n]: y
Copyingcoraenv to /usr/local/bin ...
Entries will be added to the /etc/oratab file asneeded by
Database Configuration Assistant when a database iscreated
Finished running generic part of root.sh script.
Now product-specific root actions will beperformed.
Finished product-specific root actions.
[root@grid02 db_1]#
開始建立資料庫:
[grid@grid01 ~]$ crs_stat -t
Name Type Target State Host
------------------------------------------------------------
ora.ARC001.dg ora....up.type ONLINE ONLINE grid01
ora.DATA001.dg ora....up.type ONLINE ONLINE grid01
ora.FRA001.dg ora....up.type ONLINE ONLINE grid01
ora....ER.lsnr ora....er.type ONLINE ONLINE grid01
ora....N1.lsnr ora....er.type ONLINE ONLINE grid01
ora....N2.lsnr ora....er.type ONLINE ONLINE grid02
ora....N3.lsnr ora....er.type ONLINE ONLINE grid02
ora....N4.lsnr ora....er.type ONLINE ONLINE grid02
ora....VOTE.dg ora....up.type ONLINE ONLINE grid01
ora.asm ora.asm.type ONLINE ONLINE grid01
ora.eons ora.eons.type ONLINE ONLINE grid01
ora....SM1.asm application ONLINE ONLINE grid01
ora....01.lsnr application ONLINE ONLINE grid01
ora.grid01.gsd application ONLINE ONLINE grid01
ora.grid01.ons application ONLINE ONLINE grid01
ora.grid01.vip ora....t1.type ONLINE ONLINE grid01
ora....SM2.asm application ONLINE ONLINE grid02
ora....02.lsnr application ONLINE ONLINE grid02
ora.grid02.gsd application ONLINE ONLINE grid02
ora.grid02.ons application ONLINE ONLINE grid02
ora.grid02.vip ora....t1.type ONLINE ONLINE grid02
ora.gsd ora.gsd.type ONLINE ONLINE grid01
ora....network ora....rk.type ONLINE ONLINE grid01
ora.oc4j ora.oc4j.type ONLINE ONLINE grid02
ora.ons ora.ons.type ONLINE ONLINE grid01
ora....ry.acfs ora....fs.type ONLINE ONLINE grid01
ora.scan1.vip ora....ip.type ONLINE ONLINE grid01
ora.scan2.vip ora....ip.type ONLINE ONLINE grid02
ora.scan3.vip ora....ip.type ONLINE ONLINE grid02
ora.scan4.vip ora....ip.type ONLINE ONLINE grid02
ora.woo.db ora....se.type ONLINE ONLINE grid01
[grid@grid01 ~]$ exit
資料庫安裝常見錯誤處理
1.在第二個節點執行root.sh報錯,資訊如下:
CRS-4000: Command Start failed, or completed with errors.
Command return code of 1 (256) from command: /app/11.2.0/grid/bin/crsctlstart resource ora.asm -init
Start of resource "ora.asm -init" failed
Failed to start ASM
Failed to start Oracle Clusterware stack
報上面的錯誤,上網查了下,自己疏忽造成的,由於我的第二個節點是從第一個節點複製的,/etc/hosts檔案配置不對,本地的主機名配置不正確。
解決辦法:
修改/etc/hosts檔案配置
#that require network functionality will fail.
127.0.0.1 node2localhost.localdomain localhost
執行:
[root@node2 install]# /app/11.2.0/grid/crs/install/roothas.pl -delete-force –verbose
重新執行指令碼:
[root@node2 install]# /app/11.2.0/grid/root.sh
執行成功。
2.在第二個節點執行root.sh報錯,報錯資訊如下:
INFO: ERROR:
INFO: PRVF-4664 : Found inconsistent name resolution entries for SCAN name"node-scan"
INFO: ERROR:
INFO: PRVF-4657 : Name resolution setup check for "node-scan"(IP address: 192.168.10.25) failed
INFO: ERROR:
INFO: PRVF-4664 : Found inconsistent name resolution entries for SCAN name"node-scan"
INFO: Verification of SCAN VIP and Listener setup failed
上oracle網站查了下,這種情況的原因有幾種,我的原因是由於使用了/etc/hosts檔案來解析scan-ip引起的,如果能ping通scan-ip就可以直接忽略掉。
參考metalink檔案:ID 887471.1
3.安裝oracle軟體時報錯,報錯資訊如下:
INS-35354
上metalink查了下,是由於oracle目錄清單配置檔案裡的叢集配置不對,修改即可:
$cat /home/grid/oraInventory/ContentsXML/inventory.xml
將CRS=true新增到兩個節點的配置檔案裡即可。
參考檔案:ID 1053393.1
來自 “ ITPUB部落格 ” ,連結:http://blog.itpub.net/20674423/viewspace-753480/,如需轉載,請註明出處,否則將追究法律責任。
相關文章
- redhat安裝雙節點cassandra叢集Redhat
- 3 安裝配置oracle叢集和RACOracle
- 安裝 REDIS 叢集Redis
- 安裝Kafka叢集Kafka
- CentOS 5.8上安裝Oracle 10gCentOSOracle 10g
- 完整安裝always on叢集
- 安裝Consul叢集
- 快速安裝 kafka 叢集Kafka
- FastDFS 叢集 安裝 配置AST
- Redhat5.8 X64建立裸裝置Redhat
- redis-3.0.6 安裝叢集Redis
- redis 5.0 叢集的安裝Redis
- 安裝Zookeeper和Kafka叢集Kafka
- 安裝Greenplum 5.2 叢集實戰
- Cloudera Manager安裝 & 搭建CDH叢集Cloud
- Redis安裝+叢集+效能監控Redis
- CDH安裝大資料叢集大資料
- CentOS7 安裝PG叢集CentOS
- CentOS6.5 安裝Zookeeper叢集CentOS
- CentOS6.5安裝Twemproxy叢集CentOS
- CentOS6.5 安裝Storm叢集CentOSORM
- CentOS6.5 安裝Kafka叢集CentOSKafka
- CentOS6.5 安裝Spark叢集CentOSSpark
- Redis叢集環境安裝指南Redis
- centos安裝redis3.0.0叢集CentOSRedisS3
- Redhat+Oracle+SAP Netweaver 安裝RedhatOracle
- 【安裝】RedHat 5.1 部署Oracle 10.2.0.1RedhatOracle
- redhat 6.4 安裝oracle11g RAC 安裝RDACRedhatOracle
- (轉)redhat下安裝oracle,設定redhat核心引數RedhatOracle
- 在redhat 5.8下安裝11gR2在安裝grid時報錯:INS-32026Redhat
- Laravel5.8 安裝使用 vueLaravelVue
- Cassandra安裝及分散式叢集搭建分散式
- Ubuntu 安裝k8s叢集UbuntuK8S
- centos安裝k8s叢集CentOSK8S
- CentOS上zookeeper叢集模式安裝配置CentOS模式
- Redis叢集方案,Codis安裝測試Redis
- Redhat 5 下 Oracle10g 安裝RedhatOracle
- RedHat AS3安裝Oracle 8iRedhatS3Oracle