91超碰碰碰碰久久久久久综合_超碰av人澡人澡人澡人澡人掠_国产黄大片在线观看画质优化_txt小说免费全本

溫馨提示×

溫馨提示×

您好,登錄后才能下訂單哦!

密碼登錄×
登錄注冊×
其他方式登錄
點擊 登錄注冊 即表示同意《億速云用戶服務條款》

redhat6 + 11G RAC 雙節點部署

發布時間:2020-06-23 05:41:16 來源:網絡 閱讀:1042 作者:dbapower 欄目:關系型數據庫

 

一、配置網絡環境

node1

[root@node1 ~]#vi/etc/sysconfig/network

NETWORKING=yes

NETWORKING_IPV6=no

HOSTNAME=node1

 

[root@node1 ~]# vi/etc/sysconfig/network-scripts/ifcfg-eth0

# Intel Corporation 82540EM GigabitEthernet Controller

DEVICE=eth0

BOOTPROTO=static

IPADDR=192.168.10.41

NETMASK=255.255.255.0

GATEWAY=192.168.10.1

ONBOOT=yes

 

[root@node1 ~]#vi/etc/sysconfig/network-scripts/ifcfg-eth2

# Intel Corporation 82540EM GigabitEthernet Controller

DEVICE=eth2

BOOTPROTO=static

IPADDR=10.10.10.41

NETMASK=255.255.255.0

ONBOOT=yes

 

[root@node1 ~]#vi/etc/hosts

# Do not remove the following line, orvarious programs

# that require network functionality willfail.

127.0.0.1 localhost

::1             localhost6.localdomain6 localhost6

192.168.10.41 node1

192.168.10.43 node1-vip

10.10.10.41 node1-priv

 

192.168.10.42 node2

192.168.10.44 node2-vip

10.10.10.42 node2-priv

 

192.168.10.55 rac_scan

 

[root@node1 ~]#service network restart

 

 

node2node1基本相同,IP和主機名不同)

 

二、建立用戶、組、oraclegrid文件夾

node1

[root@lq1 ~]#vimkuser.sh

groupadd-g 200 oinstall

groupadd-g 201 dba

groupadd-g 202 oper

groupadd-g 203 asmadmin

groupadd-g 204 asmoper

groupadd-g 205 asmdba

useradd-u 200 -g oinstall -G dba,asmdba,oper oracle

useradd-u 201 -g oinstall -G asmadmin,asmdba,asmoper,oper,dba grid

 

[root@node1 ~]#shmkuser.sh

 

[root@node1 ~]#vimkdir.sh

mkdir-p /u01/app/oraInventory

chown-R grid:oinstall /u01/app/oraInventory/

chmod-R 775 /u01/app/oraInventory/

mkdir-p /u01/11.2.0/grid

chown-R grid:oinstall /u01/11.2.0/grid/

chmod-R 775 /u01/11.2.0/grid/

mkdir-p /u01/app/oracle

mkdir-p /u01/app/oracle/cfgtoollogs

mkdir-p /u01/app/oracle/product/11.2.0/db_1

chown-R oracle:oinstall /u01/app/oracle

chmod-R 775 /u01/app/oracle

 

[root@node1 ~]#shmkdir.sh

 

[root@node1 ~]#passwdoracle

 

[root@node1 ~]#passwdgrid

 

[root@node1~]# id oracle

uid=200(oracle)gid=200(oinstall) groups=200(oinstall),201(dba),202(oper),205(asmdba)

 

[root@node1~]# id grid

uid=201(grid)gid=200(oinstall) groups=200(oinstall),201(dba),202(oper),203(asmadmin),204(asmoper),205(asmdba)

 

[root@node1~]# id nobody

uid=99(nobody)gid=99(nobody) groups=99(nobody)

 

node2node1相同)

 

 

三、修改/etc目錄下的4個文件

node1

[root@node1~]#vi /etc/sysctl.conf

fs.aio-max-nr= 1048576

fs.file-max= 6815744

kernel.shmall= 2097152

kernel.shmmax= 536870912

kernel.shmmni= 4096

kernel.sem= 250 32000 100 128

net.ipv4.ip_local_port_range= 9000 65500

net.core.rmem_default= 262144

net.core.rmem_max= 4194304

net.core.wmem_default= 262144

net.core.wmem_max= 1048586

 

 

[root@node1~]# sysctl –p

 

 

[root@node1~]#vi /etc/security/limits.conf

oraclesoft nproc 2047

oraclehard nproc 16384

oraclesoft nofile1024

oraclehard nofile 65536

oraclesoft stack10240

gridsoft nproc 2047

gridhard nproc 16384

gridsoft nofile 1024

gridhard nofile65536

gridsoft stack 10240

 

[root@node1~]#vi /etc/pam.d/login

sessionrequired /lib/security/pam_limits.so

 

[root@node1~]#vi /etc/profile

if [$USER = "oracle" ]||[ $USER = "grid" ]; then

if [ $SHELL = "/bin/ksh" ]; then

ulimit -p 16384

ulimit -n 65536

else

ulimit -u 16384 -n 65536

fi

fi

node2node1相同)

 

 

四、關閉ntp服務,采用oracle自帶的時間,

關閉郵件服務

node1

[root@node1~]# chkconfig ntpd off

 

[root@node1~]# chkconfig ntpd --list

 

[root@node1~]# mv /etc/ntp.conf /etc/ntp.conf.bak

 

 

 

[root@node1~]# chkconfig sendmail off

 

[root@node1~]# chkconfig sendmail --list

 

 

 

 

 

node2node1相同)

 

 

五、修改oracle和grid用戶的環境變量

 

node1

[root@node1 ~]#su - oracle

 

[oracle@node1~]$vi .bash_profile

export EDITOR=vi

exportORACLE_SID=prod1

exportORACLE_BASE=/u01/app/oracle

exportORACLE_HOME=$ORACLE_BASE/product/11.2.0/db_1

export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib

exportPATH=$ORACLE_HOME/bin:/bin:/usr/bin:/usr/sbin:/usr/local/bin:/usr/X11R6/bin

umask 022

 

[oracle@node1~]$. .bash_profile

 

[root@node1 ~]#su – grid

 

[grid@node1~]$vi .bash_profile

export EDITOR=vi

exportORACLE_SID=+ASM1

exportORACLE_BASE=/u01/app/oracle

exportORACLE_HOME=/u01/11.2.0/grid

exportGRID_HOME=/u01/11.2.0/grid

exportLD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib

exportTHREADS_FLAG=native

exportPATH=$ORACLE_HOME/bin:/bin:/usr/bin:/usr/sbin:/usr/local/bin:/usr/X11R6/bin

umask 022

 

[grid@node1~]$. .bash_profile

node2node1相同)

 

六、硬盤分區  創建ASM磁盤

node1

查看系統里所有磁盤情況

[root@node1~]# fdisk -l

Disk /dev/sda: 21.4 GB, 21474836480 bytes

255heads, 63 sectors/track, 2610 cylinders

Units= cylinders of 16065 * 512 = 8225280 bytes

DeviceBoot      Start         End      Blocks  Id  System

/dev/sda1*         1          13      104391   83  Linux

/dev/sda2          14       2610    20860402+   8e  Linux LVM

Disk/dev/sdb: 32.2 GB, 32212254720 bytes

255heads, 63 sectors/track, 3916 cylinders

Units= cylinders of 16065 * 512 = 8225280 bytes

Disk/dev/sdb doesn't contain a valid partition table

Disk/dev/sdc: 21.4 GB, 21474836480 bytes

255heads, 63 sectors/track, 2610 cylinders

Units= cylinders of 16065 * 512 = 8225280 bytes

Disk/dev/sdc doesn't contain a valid partition table

 

給/dev/sdb磁盤分區

[root@node1~]# fdisk/dev/sdb

給/dev/sdc磁盤分區

[root@node1~]# fdisk/dev/sdc

查看系統里的磁盤信息

[root@node1~]# fdisk -l

格式化/dev/sdb1磁盤

[root@node1~]# mkfs.ext3/dev/sdb1

掛載新磁盤/dev/sdb1到/u01,查看掛載情況

[root@node1~]# mount/dev/sdb1 /u01

[root@node1~]# df -h

Filesystem                       Size Used Avail Use% Mounted on

/dev/mapper/VolGroup00-LogVol00  18G 6.1G   11G  38% /

/dev/sda1                        99M  12M   82M  13% /boot

tmpfs                            782M     0 782M   0% /dev/shm

/dev/sdb1                        30G 173M   28G   1% /u01

 

查看物理內存和換頁空間swap

[root@node1~]# free -m

             total       used       free    shared    buffers     cached

Mem:          1562       1525         37          0         11       1438

-/+buffers/cache:         75       1486

Swap:         2047          0       2047

 

創建一個大文件

[root@node1soft]# ddif=/dev/zero of=/u01/swapfile1 bs=1024k count=2048

2048+0records in

2048+0records out

2147483648bytes (2.1 GB) copied, 5.66353 seconds, 379 MB/s

 

創建swap文件

[root@node1~]# mkswap -c/u01/swapfile1

Settingup swapspace version 1, size = 2147479 Kb

 

掛載swap文件

[root@node1~]# swapon/u01/swapfile1

 

查看物理內存和更改后的換頁空間swap

[root@node1~]# free -m

             total       used       free    shared    buffers     cached

Mem:          1562       1525         37          0         11       1438

-/+buffers/cache:         75       1486

Swap:         4095          0       4095

 

將掛載的新磁盤,增加的swap文件寫入到fstab文件,重啟系統后會自動掛載

[root@node1~]# vi/etc/fstab

/dev/VolGroup00/LogVol00/                      ext3    defaults        1 1

LABEL=/boot             /boot                   ext3    defaults        1 2

tmpfs                   /dev/shm                tmpfs   defaults,size=1g        0 0

devpts                  /dev/pts                devpts  gid=5,mode=620  0 0

sysfs                   /sys                    sysfs   defaults        0 0

proc                    /proc                   proc    defaults       0 0

/dev/VolGroup00/LogVol01swap                   swap    defaults        0 0

/dev/sdb1              /u01                    ext3    defaults        0 0

/u01/swapfile1          swap                   swap    defaults        0 0

 

 

[root@node2~]# mkfs.ext3/dev/sdb1

掛載新磁盤/dev/sdb1到/u01,查看掛載情況

[root@node2~]# mount/dev/sdb1 /u01

[root@node2~]# df -h

Filesystem                       Size Used Avail Use% Mounted on

/dev/mapper/VolGroup00-LogVol00  18G 6.1G   11G  38% /

/dev/sda1                        99M  12M   82M  13% /boot

tmpfs                            782M     0 782M   0% /dev/shm

/dev/sdb1                        30G 173M   28G   1% /u01

 

查看物理內存和換頁空間swap

[root@node2~]# free -m

             total       used       free    shared    buffers     cached

Mem:          1562       1525         37          0         11       1438

-/+buffers/cache:         75       1486

Swap:         2047          0       2047

 

創建一個大文件

[root@node2soft]# ddif=/dev/zero of=/u01/swapfile1 bs=1024k count=2048

2048+0records in

2048+0records out

2147483648bytes (2.1 GB) copied, 5.66353 seconds, 379 MB/s

 

創建swap文件

[root@node2~]# mkswap -c/u01/swapfile1

Settingup swapspace version 1, size = 2147479 Kb

 

掛載swap文件

[root@node2~]# swapon/u01/swapfile1

 

查看物理內存和更改后的換頁空間swap

[root@node2~]# free -m

             total       used       free    shared    buffers     cached

Mem:          1562       1525         37          0         11       1438

-/+buffers/cache:         75       1486

Swap:         4095          0       4095

 

將掛載的新磁盤,增加的swap文件寫入到fstab文件,重啟系統后會自動掛載

[root@node2~]# vi/etc/fstab

/dev/VolGroup00/LogVol00/                      ext3    defaults        1 1

LABEL=/boot             /boot                   ext3    defaults        1 2

tmpfs                   /dev/shm                tmpfs   defaults,size=1g        0 0

devpts                  /dev/pts                devpts  gid=5,mode=620  0 0

sysfs                   /sys                    sysfs   defaults        0 0

proc                    /proc                   proc    defaults       0 0

/dev/VolGroup00/LogVol01swap                   swap    defaults        0 0

/dev/sdb1              /u01                    ext3    defaults        0 0

/u01/swapfile1          swap                   swap    defaults        0 0

 

查看ASM磁盤管理軟件的位置 (從網站下載并上傳到linux系統)

[root@node1~]# cd /soft/asm

[root@node1asm]# ls

oracleasm-2.6.18-194.el5-2.0.5-1.el5.i686.rpm

oracleasmlib-2.0.4-1.el5.i386.rpm

oracleasm-support-2.1.3-1.el5.i386.rpm

 

注意與內核版本的匹配

[root@node1asm]# uname -a

Linux node1 2.6.18-194.el5 #1 SMP Tue Mar16 21:52:43 EDT 2010 i686 i686 i386 GNU/Linux

 

安裝ASM管理軟件

[root@node1asm]# rpm -ivh *.rpm

warning:oracleasm-2.6.18-194.el5-2.0.5-1.el5.i686.rpm:Header V3 DSA signature: NOKEY, key ID 1e5e0159

Preparing...              ###########################################[100%]

1:oracleasm-support      ########################################### [ 33%]

2:oracleasm-2.6.18-194.el########################################### [ 67%]

3:oracleasmlib            ###########################################[100%]

 

配置 oracleasm初始化

[root@node1soft]# serviceoracleasm configure

Configuringthe Oracle ASM library driver.

Thiswill configure the on-boot properties of the Oracle ASM library

driver.   The following questions will determinewhether the driver is

loadedon boot and what permissions it will have.  The current values

willbe shown in brackets ('[]').   Hitting<ENTER> without typing an

answerwill keep that current value.   Ctrl-Cwill abort.

Defaultuser to own the driver interface []: grid

Defaultgroup to own the driver interface []: asmadmin

StartOracle ASM library driver on boot (y/n) [n]: y

Scanfor Oracle ASM disks on boot (y/n) [y]:

WritingOracle ASM library driver configuration: done

Initializingthe Oracle ASMLib driver: [   OK   ]

Scanningthe system for Oracle ASMLib disks: [  OK   ]

 

建立 oracleasm 磁盤

[root@node1soft]# serviceoracleasm createdisk OCR_VOTE1 /dev/sdc1

Markingdisk "OCR_VOTE1" as an ASM disk: [  OK   ]

[root@node1soft]# serviceoracleasm createdisk OCR_VOTE2 /dev/sdc2

Markingdisk "OCR_VOTE2" as an ASM disk: [  OK   ]

[root@node1soft]# serviceoracleasm createdisk OCR_VOTE3 /dev/sdc3

Markingdisk "OCR_VOTE3" as an ASM disk: [  OK   ]

[root@node1soft]# serviceoracleasm createdisk ASM_DATA1 /dev/sdc5

Markingdisk "ASM_DATA1" as an ASM disk: [  OK   ]

[root@node1soft]# serviceoracleasm createdisk ASM_DATA2 /dev/sdc6

Markingdisk "ASM_DATA2" as an ASM disk: [  OK   ]

[root@node1soft]# serviceoracleasm createdisk ASM_RCY1 /dev/sdc7

Markingdisk "ASM_RCY1" as an ASM disk: [   OK   ]

[root@node1soft]# serviceoracleasm createdisk ASM_RCY2 /dev/sdc8

Markingdisk "ASM_RCY2" as an ASM disk: [  OK   ]

[root@node1soft]# serviceoracleasm listdisks

ASM_DATA1

ASM_DATA2

ASM_RCY1

ASM_RCY2

OCR_VOTE1

OCR_VOTE2      

OCR_VOTE3

 

這個時候把node1 /soft/asm的三個包拷貝到node2 /soft/asm里

 

拷貝完后查看ASM磁盤管理軟件的位置 (從網站下載并上傳到linux系統)

注意與內核版本的匹配

[root@node2asm]# uname -a

Linux node1 2.6.18-194.el5 #1 SMP Tue Mar16 21:52:43 EDT 2010 i686 i686 i386 GNU/Linux

 

安裝ASM管理軟件

[root@node2asm]# rpm -ivh *.rpm

warning:oracleasm-2.6.18-194.el5-2.0.5-1.el5.i686.rpm:Header V3 DSA signature: NOKEY, key ID 1e5e0159

Preparing...              ###########################################[100%]

1:oracleasm-support      ########################################### [ 33%]

2:oracleasm-2.6.18-194.el########################################### [ 67%]

3:oracleasmlib            ###########################################[100%]

 

Node2也需要執行oraclasm初始化

[root@node2soft]# serviceoracleasm configure

Configuringthe Oracle ASM library driver.

Thiswill configure the on-boot properties of the Oracle ASM library

driver.   The following questions will determinewhether the driver is

loadedon boot and what permissions it will have.  The current values

willbe shown in brackets ('[]').   Hitting<ENTER> without typing an

answerwill keep that current value.   Ctrl-Cwill abort.

Defaultuser to own the driver interface []: grid

Defaultgroup to own the driver interface []: asmadmin

StartOracle ASM library driver on boot (y/n) [n]: y

Scanfor Oracle ASM disks on boot (y/n) [y]:

WritingOracle ASM library driver configuration: done

Initializingthe Oracle ASMLib driver: [   OK   ]

Scanningthe system for Oracle ASMLib disks: [  OK   ]

 

然后執行asm掃描并查看

[root@node2asm]# serviceoracleasm scandisks

Scanningthe system for Oracle ASMLib disks: [  OK   ]

[root@node2asm]# serviceoracleasm listdisks

ASM_DATA1

ASM_DATA2

ASM_RCY1

ASM_RCY2

OCR_VOTE1

OCR_VOTE2

OCR_VOTE3

 

 

(node1和node2,共享磁盤/dev/sdc不用配置,其他配置相同)

 

 

 

七、建立主機間的信任關系

建立節點之間 oracle 、grid  用戶之間的信任(通過 ssh  生成成對秘鑰)

node1   --oracle用戶

[root@node1~]# su - oracle

[oracle@node1~]$ mkdir .ssh

[oracle@node1~]$ ls -a

.  .. .bash_history  .bash_logout  .bash_profile .bashrc  .emacs  .kde .mozilla  .ssh  .viminfo

[oracle@node1~]$ ssh-keygen-t rsa

 [oracle@node1 ~]$ ssh-keygen -t dsa

 

Node2   --oracle用戶

[root@node1~]# su - oracle

[oracle@node2~]$ mkdir .ssh

[oracle@node2~]$ ls -a

.  .. .bash_history  .bash_logout  .bash_profile .bashrc  .emacs  .kde .mozilla  .ssh  .viminfo

[oracle@node2~]$ ssh-keygen-t rsa

 [oracle@node2 ~]$ ssh-keygen -t dsa

配置信任關系

[oracle@node1~]$ ls .ssh

id_dsa   id_dsa.pub  id_rsa   id_rsa.pub   known_hosts

 

[oracle@node1~]$ cat.ssh/id_rsa.pub >> .ssh/authorized_keys

 

[oracle@node1~]$ cat.ssh/id_dsa.pub >> .ssh/authorized_keys

 

[oracle@node1~]$ ssh node2cat .ssh/id_rsa.pub >> .ssh/authorized_keys

 [oracle@node1 ~]$ ssh node2 cat .ssh/id_dsa.pub >>.ssh/authorized_keys

oracle@node2'spassword:

 

[oracle@node1~]$ scp.ssh/authorized_keys node2:~/.ssh

oracle@node2'spassword:

authorized_keys                                          100% 1988     1.9KB/s  00:00

 

驗證信任關系

[oracle@node1~]$ ssh node1date

 [oracle@node1 ~]$ ssh node1-priv date

 [oracle@node1 ~]$ ssh node2-priv date

 [oracle@node1 ~]$ ssh node2 date

[oracle@node1~]$ ssh node1date

WedAug 27 00:48:15 CST 2014

[oracle@node1~]$ sshnode1-priv date

WedAug 27 00:48:17 CST 2014

[oracle@node1~]$ ssh node2date

WedAug 27 00:48:18 CST 2014

[oracle@node1~]$ sshnode2-priv date

WedAug 27 00:48:21 CST 2014

[oracle@node1~]$ ssh node2date;date

WedAug 27 00:50:28 CST 2014

WedAug 27 00:50:29 CST 2014

[oracle@node1~]$ sshnode2-priv date;date

WedAug 27 00:50:38 CST 2014

WedAug 27 00:50:38 CST 2014

 

[oracle@node2~]$ ssh node2date

 [oracle@node2 ~]$ ssh node2-priv date

 [oracle@node2 ~]$ ssh node1-priv date

 [oracle@node2 ~]$ ssh node1 date

 

[oracle@node2~]$ ssh node2date

WedAug 27 00:49:09 CST 2014

[oracle@node2~]$ sshnode2-priv date

WedAug 27 00:49:11 CST 2014

[oracle@node2~]$ ssh node1date

WedAug 27 00:49:15 CST 2014

[oracle@node2~]$ sshnode1-priv date

WedAug 27 00:49:19 CST 2014

 

[oracle@node2~]$ ssh node1date;date

WedAug 27 00:51:28 CST 2014

WedAug 27 00:51:29 CST 2014

[oracle@node2~]$ ssh node1-privdate;date

WedAug 27 00:51:48 CST 2014

WedAug 27 00:51:48 CST 2014

 

 

node1   --grid用戶

[root@node1~]# su - grid

[grid@node1~]$ mkdir .ssh

[grid@node1~]$ ls -a

.  .. .bash_history  .bash_logout  .bash_profile .bashrc  .emacs  .kde .mozilla  .ssh .viminfo

 

[grid@node1~]$ ssh-keygen-t rsa

G

[grid@node1~]$ ssh-keygen-t dsa

Node2   --grid用戶

[root@node2~]# su - grid

[grid@node2~]$ mkdir .ssh

[grid@node2~]$ ls -a

.  .. .bash_history  .bash_logout  .bash_profile .bashrc  .emacs  .kde .mozilla  .ssh .viminfo

 

[grid@node2~]$ ssh-keygen-t rsa

 [grid@node2 ~]$ ssh-keygen -t dsa

配置信任關系

 

[grid@node1~]$ cat.ssh/id_rsa.pub >> .ssh/authorized_keys

 

[grid@node1~]$ cat.ssh/id_dsa.pub >> .ssh/authorized_keys

 

[grid@node1~]$ ssh node2cat .ssh/id_rsa.pub >> .ssh/authorized_keys

 [grid@node1 ~]$ ssh node2 cat .ssh/id_dsa.pub >>.ssh/authorized_keys

grid@node2'spassword:

 

[grid@node1~]$ scp.ssh/authorized_keys node2:~/.ssh

grid@node2'spassword:

authorized_keys                                          100% 1984     1.9KB/s   00:00

 

 

驗證信任關系

[grid@node1~]$ ssh node1date

 [grid@node1 ~]$ ssh node1-priv date

 [grid@node1 ~]$ ssh node2-priv date

 [grid@node1 ~]$ ssh node2 date

[grid@node1~]$ ssh node1date

WedAug 27 00:57:37 CST 2014

[grid@node1~]$ sshnode1-priv date

WedAug 27 00:57:39 CST 2014

[grid@node1~]$ ssh node2date

WedAug 27 00:57:41 CST 2014

[grid@node1~]$ sshnode2-priv date

WedAug 27 00:57:43 CST 2014

[grid@node1~]$ sshnode2-priv date;date

WedAug 27 00:57:50 CST 2014

WedAug 27 00:57:51 CST 2014

[grid@node1~]$ ssh node2date;date

WedAug 27 00:58:01 CST 2014

WedAug 27 00:58:01 CST 2014

 

[grid@node2~]$ ssh node2date

 [grid@node2 ~]$ ssh node2-priv date

 [grid@node2 ~]$ ssh node1-priv date

[grid@node2~]$ ssh node1date

 

[grid@node2~]$ ssh node2date

WedAug 27 00:59:01 CST 2014

[grid@node2~]$ sshnode2-priv date

WedAug 27 00:59:03 CST 2014

[grid@node2~]$ ssh node1date

WedAug 27 00:59:05 CST 2014

[grid@node2~]$ sshnode1-priv date

WedAug 27 00:59:08 CST 2014

 

[grid@node2~]$ sshnode1-priv date;date

WedAug 27 00:59:12 CST 2014

WedAug 27 00:59:12 CST 2014

[grid@node2~]$ ssh node1date;date

WedAug 27 00:59:25 CST 2014

WedAug 27 00:59:24 CST 2014

 

八、校驗安裝前的環境

以 grid  用戶的身份校驗安裝環境(在 grid 的安裝軟件包目錄下)

[root@node1~]# cd /soft

[root@node1soft]# ls

asm linux_11gR2_database_1of2.zip  linux_11gR2_database_2of2.zip  linux_11gR2_grid.zip

[root@node1soft]# unzip linux_11gR2_grid.zip

[root@node1soft]# ls

asm grid  linux_11gR2_database_1of2.zip  linux_11gR2_database_2of2.zip  linux_11gR2_grid.zip

[root@node1soft]# chown -Rgrid:oinstall grid/

[root@node1soft]# chmod -R775 grid/

[root@node1soft]# chown -Rgrid:oinstall /tmp/bootstrap/    沒有這個目錄就不用操作了

[root@node1soft]# chmod -R775 /tmp/bootstrap/              沒有這個目錄就不用操作了

[root@node1soft]# su - grid

[grid@node1~]$ cd/soft/grid/

[grid@node1grid]$./runcluvfy.sh stage -pre crsinst -n node1,node2 -fixup -verbose

 

注意其中“failed”的位置

對于校驗中沒有安裝的軟件包進行安裝(所有節點)

(node1和node2 相同)最后所有的節點都應該是passed自己檢查一下。

安裝Grid

 

 [root@node1 ~]# /u01/app/oraInventory/orainstRoot.sh

Changingpermissions of /u01/app/oraInventory.

Addingread,write permissions for group.

Removingread,write,execute permissions for world.

Changinggroupname of /u01/app/oraInventory to oinstall.

Theexecution of the script is complete.

 

節點2也要運行/u01/app/oraInventory/orainstRoot.sh

 

[root@node1~]# /u01/11.2.0/grid/root.sh

節點2也要運行/u01/11.2.0/grid/root.sh

 

(node2也這樣,記住node1運行完第一個腳本,node2也要運行第一個腳本,然后node1再運行第二個腳本

  node2也再運行第二個腳本,順序不能錯。)

 

完成grid安裝后,檢查crs進程是否開啟

node1

[root@node1~]# vi /etc/profile

exportPATH=$PATH:/u01/11.2.0/grid/bin

 

 

[root@node1~]# source /etc/profile

 

 

[root@node1~]# crsctl check crs

CRS-4638:Oracle High Availability Services is online

CRS-4537:Cluster Ready Services is online

CRS-4529:Cluster Synchronization Services is online

CRS-4533:Event Manager is online

 

 

[root@node1~]# crs_stat -t

Name           Type           Target    State    Host        

------------------------------------------------------------

ora....ER.lsnrora....er.type ONLINE    ONLINE    node1      

ora....N1.lsnrora....er.type ONLINE    ONLINE    node1      

ora....VOTE.dgora....up.type ONLINE    ONLINE    node1      

ora.asm        ora.asm.type   ONLINE   ONLINE    node1      

ora.eons       ora.eons.type  ONLINE   ONLINE    node1      

ora.gsd        ora.gsd.type   OFFLINE  OFFLINE              

ora....networkora....rk.type ONLINE    ONLINE    node1      

ora....SM1.asmapplication    ONLINE    ONLINE   node1      

ora....E1.lsnrapplication    ONLINE    ONLINE   node1      

ora.node1.gsd  application   OFFLINE   OFFLINE              

ora.node1.ons  application   ONLINE    ONLINE    node1      

ora.node1.vip  ora....t1.type ONLINE    ONLINE   node1      

ora....SM2.asmapplication    ONLINE    ONLINE   node2      

ora....E2.lsnrapplication    ONLINE    ONLINE   node2      

ora.node2.gsd  application   OFFLINE   OFFLINE              

ora.node2.ons  application   ONLINE    ONLINE    node2      

ora.node2.vip  ora....t1.type ONLINE    ONLINE   node2      

ora.oc4j       ora.oc4j.type  OFFLINE  OFFLINE              

ora.ons        ora.ons.type   ONLINE   ONLINE    node1      

ora....ry.acfsora....fs.type ONLINE    ONLINE    node1      

ora.scan1.vip  ora....ip.type ONLINE    ONLINE   node1      

 

 

 

 

 

 

 

 

node2

[root@node2~]# vi /etc/profile

exportPATH=$PATH:/u01/11.2.0/grid/bin

 

 

[root@node2~]# source /etc/profile

 

 

[root@node2~]# crsctl check crs

CRS-4638:Oracle High Availability Services is online

CRS-4537:Cluster Ready Services is online

CRS-4529:Cluster Synchronization Services is online

CRS-4533:Event Manager is online

 

 

 

[root@node2~]# crs_stat -t

Name           Type           Target   State     Host       

------------------------------------------------------------

ora....ER.lsnrora....er.type ONLINE    ONLINE    node1      

ora....N1.lsnrora....er.type ONLINE    ONLINE    node1      

ora....VOTE.dgora....up.type ONLINE    ONLINE    node1      

ora.asm        ora.asm.type   ONLINE   ONLINE    node1      

ora.eons       ora.eons.type  ONLINE   ONLINE    node1      

ora.gsd        ora.gsd.type   OFFLINE  OFFLINE              

ora....networkora....rk.type ONLINE    ONLINE   node1      

ora....SM1.asmapplication    ONLINE    ONLINE   node1      

ora....E1.lsnrapplication    ONLINE    ONLINE   node1      

ora.node1.gsd  application   OFFLINE   OFFLINE              

ora.node1.ons  application   ONLINE    ONLINE    node1      

ora.node1.vip  ora....t1.type ONLINE    ONLINE   node1      

ora....SM2.asmapplication    ONLINE    ONLINE   node2      

ora....E2.lsnrapplication    ONLINE    ONLINE   node2      

ora.node2.gsd  application   OFFLINE   OFFLINE              

ora.node2.ons  application   ONLINE    ONLINE    node2      

ora.node2.vip  ora....t1.type ONLINE    ONLINE   node2      

ora.oc4j       ora.oc4j.type  OFFLINE  OFFLINE              

ora.ons        ora.ons.type   ONLINE   ONLINE    node1      

ora....ry.acfsora....fs.type ONLINE    ONLINE    node1      

ora.scan1.vip  ora....ip.type ONLINE    ONLINE   node1      

 

 

 

完成grid安裝后,檢查crs進程是否開啟

node1

[root@node1~]# vi /etc/profile

exportPATH=$PATH:/u01/11.2.0/grid/bin

 

 

[root@node1~]# source /etc/profile

 

 

[root@node1~]# crsctl check crs

CRS-4638:Oracle High Availability Services is online

CRS-4537:Cluster Ready Services is online

CRS-4529:Cluster Synchronization Services is online

CRS-4533:Event Manager is online

 

 

[root@node1~]# crs_stat -t

Name           Type           Target    State    Host       

------------------------------------------------------------

ora....ER.lsnrora....er.type ONLINE    ONLINE    node1      

ora....N1.lsnrora....er.type ONLINE    ONLINE    node1      

ora....VOTE.dgora....up.type ONLINE    ONLINE    node1      

ora.asm        ora.asm.type   ONLINE   ONLINE    node1      

ora.eons       ora.eons.type  ONLINE   ONLINE    node1      

ora.gsd        ora.gsd.type   OFFLINE  OFFLINE              

ora....networkora....rk.type ONLINE    ONLINE    node1      

ora....SM1.asmapplication    ONLINE    ONLINE   node1      

ora....E1.lsnrapplication    ONLINE    ONLINE   node1      

ora.node1.gsd  application   OFFLINE   OFFLINE              

ora.node1.ons  application   ONLINE    ONLINE    node1      

ora.node1.vip  ora....t1.type ONLINE    ONLINE   node1      

ora....SM2.asmapplication    ONLINE    ONLINE   node2      

ora....E2.lsnrapplication    ONLINE    ONLINE   node2      

ora.node2.gsd  application   OFFLINE   OFFLINE              

ora.node2.ons  application   ONLINE    ONLINE    node2      

ora.node2.vip  ora....t1.type ONLINE    ONLINE   node2      

ora.oc4j       ora.oc4j.type  OFFLINE  OFFLINE              

ora.ons        ora.ons.type   ONLINE   ONLINE    node1      

ora....ry.acfsora....fs.type ONLINE    ONLINE    node1      

ora.scan1.vip  ora....ip.type ONLINE    ONLINE   node1  

 

 

 

 

九、安裝oracle軟件

 

[root@node1~]# cd /soft/

 

[root@node1soft]# ls

asmgridlinux_11gR2_database_1of2.ziplinux_11gR2_database_2of2.zip linux_11gR2_grid.zip

 

[root@node1soft]# unzip linux_11gR2_database_1of2.zip

…………

 

[root@node1soft]# unzip linux_11gR2_database_2of2.zip

…………

 

[root@node1soft]# ls

asmgrid database linux_11gR2_database_1of2.ziplinux_11gR2_database_2of2.zip linux_11gR2_grid.zip

 

[root@node1soft]# chown -R oracle:oinstall database/

 

[root@node1soft]# chmod -R 775 database/

 

[root@node1~]# su - oracle

 

[oracle@node1~]$ cd /soft/database/

 

[oracle@node1database]$ ls

doc  install response  rpm  runInstaller sshsetup  stage  welcome.html

 

 

 

 

安裝前的準備與grid方式一樣

[oracle@node1database]$ ./runInstaller

StartingOracle Universal Installer...

 

CheckingTemp space: must be greater than 80 MB.  Actual 7196 MB    Passed

Checkingswap space: must be greater than 150 MB.  Actual 4005 MB    Passed

Checkingmonitor: must be configured to display at least 256 colors.    Actual 16777216    Passed

Preparingto launch Oracle Universal Installer from /tmp/OraInstall2014-08-27_03-43-06AM.Please wait ...[oracle@node1 database]$

 

 

 

 

 

 

 

 

 

 

node1

[root@node1~]# /u01/app/oracle/product/11.2.0/db_1/root.sh

RunningOracle 11g root.sh script...

Thefollowing environment variables are set as:

    ORACLE_OWNER= oracle

    ORACLE_HOME=  /u01/app/oracle/product/11.2.0/db_1

Enterthe full pathname of the local bin directory: [/usr/local/bin]:

Thefile "dbhome" already exists in /usr/local/bin.  Overwrite it? (y/n)

[n]:y

   Copying dbhome to /usr/local/bin ...

Thefile "oraenv" already exists in /usr/local/bin.  Overwrite it? (y/n)

[n]:y

   Copying oraenv to /usr/local/bin ...

Thefile "coraenv" already exists in /usr/local/bin.  Overwrite it? (y/n)

[n]:y

   Copying coraenv to /usr/local/bin ...

Entrieswill be added to the /etc/oratab file as needed by

DatabaseConfiguration Assistant when a database is created

Finishedrunning generic part of root.sh script.

Nowproduct-specific root actions will be performed.

Finishedproduct-specific root actions.

 

 

 

node2

[root@node2~]# /u01/app/oracle/product/11.2.0/db_1/root.sh

RunningOracle 11g root.sh script...

 

Thefollowing environment variables are set as:

    ORACLE_OWNER= oracle

    ORACLE_HOME=  /u01/app/oracle/product/11.2.0/db_1

Enterthe full pathname of the local bin directory: [/usr/local/bin]:

Thefile "dbhome" already exists in /usr/local/bin.  Overwrite it? (y/n)

[n]:y

   Copying dbhome to /usr/local/bin ...

Thefile "oraenv" already exists in /usr/local/bin.  Overwrite it? (y/n)

[n]:y

   Copying oraenv to /usr/local/bin ...

Thefile "coraenv" already exists in /usr/local/bin.  Overwrite it? (y/n)

[n]:y

   Copying coraenv to /usr/local/bin ...

Entrieswill be added to the /etc/oratab file as needed by

DatabaseConfiguration Assistant when a database is created

Finishedrunning generic part of root.sh script.

Nowproduct-specific root actions will be performed.

Finishedproduct-specific root actions.

 

 

 

 

 

十、創建ASM磁盤組

[root@node1~]# su - grid

[grid@node1~]$ asmca

 

 

 

 

十一、DBCA建庫

[oracle@node1 ~]$dbca

 

完成oracle數據庫的安裝

驗證

[oracle@node1 ~]$sqlplus / as sysdba

 

SQL*Plus: Release11.2.0.1.0 Production on Wed Aug 27 04:52:36 2014

 

Copyright (c)1982, 2009, Oracle.  All rights reserved.

 

 

Connected to:

Oracle Database11g Enterprise Edition Release 11.2.0.1.0 - Production

With thePartitioning, Real Application Clusters, Automatic Storage Management, OLAP,

Data Mining andReal Application Testing options

 

SQL> select status from gv$instance;

 

STATUS

------------

OPEN

OPEN

 

SQL> show parameter name

 

NAME                                 TYPE        VALUE

----------------------------------------------- ------------------------------

db_file_name_convert                 string

db_name                              string      prod

db_unique_name                       string      prod

global_names                         boolean     FALSE

instance_name                        string      prod1

lock_name_space                      string

log_file_name_convert                string

service_names                        string      prod

 

 

 

 

 

 

 

 

 

 

 


向AI問一下細節

免責聲明:本站發布的內容(圖片、視頻和文字)以原創、轉載和分享為主,文章觀點不代表本網站立場,如果涉及侵權請聯系站長郵箱:is@yisu.com進行舉報,并提供相關證據,一經查實,將立刻刪除涉嫌侵權內容。

AI

鸡泽县| 南漳县| 西乌珠穆沁旗| 西林县| 霸州市| 甘谷县| 法库县| 融水| 绩溪县| 陕西省| 鹤山市| 南丰县| 类乌齐县| 马关县| 七台河市| 云南省| 东海县| 固始县| 蒙自县| 浮山县| 桦南县| 吉水县| 石城县| 鲁山县| 五莲县| 湖南省| 获嘉县| 胶南市| 河东区| 贡嘎县| 思南县| 松滋市| 通化市| 镇坪县| 长岛县| 社会| 恩平市| 巴林左旗| 金塔县| 来安县| 邻水|