环境介绍

OS: Oracle Enterprise Linux 6.4 (For RAC Nodes)

DB: GI and Database 12.1.0.2

所需介质

p17694377_121020_Linux-x86-64_1of8.zip

p17694377_121020_Linux-x86-64_2of8.zip

p17694377_121020_Linux-x86-64_3of8.zip

p17694377_121020_Linux-x86-64_4of8.zip

上述软件1-2介质是DB软件,3-4是GRID软件。

– 这里只给出Oracle相关的,操作系统以及其他软件请自身准备。

操作系统信息

RAC节点服务器:

(以node1节点为例)

[root@xsyy1 ~]# cat /etc/redhat-release

Red Hat Enterprise Linux Server release 6.4 (Santiago)

[root@xsyy1 ~]# uname -a

Linux xsyy1 2.6.39-400.17.1.el6uek.x86_64 #1 SMP Fri Feb 22 18:16:18 PST 2013 x86_64 x86_64 x86_64 GNU/Linux

[root@xsyy1 ~]# grep MemTotal /proc/meminfo

MemTotal:       65952032 kB

[root@xsyy1 ~]# grep SwapTotal /proc/meminfo

SwapTotal:      131070972 kB

[root@xsyy1 ~]# df -h

文件系统           容量  已用  可用 已用%% 挂载点

/dev/sdd2             1.7T   39G1.6T   3% /

tmpfs                  32G  624M31G   2% /dev/shm

/dev/sdd1             9.7G195M  9.0G   3% /boot

网络配置信息:

备注: 每个节点服务器至少需要两个网卡,eth0用于PUBLIC,而eth1用于Private

(节点1)

[root@xsyy1 ~]# ifconfig

eth0      Link encap:Ethernet  HWaddr 6C:92:BF:2A:7F:8A

inet addr:172.16.28.24Bcast:172.16.28.255Mask:255.255.255.0

inet6 addr: fe80::6e92:bfff:fe2a:7f8a/64 Scope:Link

UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1

RX packets:23832511 errors:0 dropped:93 overruns:0 frame:0

TX packets:24962108 errors:0 dropped:0 overruns:0 carrier:0

collisions:0 txqueuelen:1000

RX bytes:25937636908 (24.1 GiB)  TX bytes:29336937458 (27.3 GiB)

eth0:1    Link encap:Ethernet  HWaddr 6C:92:BF:2A:7F:8A

inet addr:172.16.28.52Bcast:172.16.28.255Mask:255.255.255.0

UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1

eth0:2    Link encap:Ethernet  HWaddr 6C:92:BF:2A:7F:8A

inet addr:172.16.28.39Bcast:172.16.28.255Mask:255.255.255.0

UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1

eth1      Link encap:Ethernet  HWaddr 6C:92:BF:2A:7F:8B

inet addr:10.1.20.24  Bcast:10.1.20.255  Mask:255.255.255.0

inet6 addr: fe80::6e92:bfff:fe2a:7f8b/64 Scope:Link

UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1

RX packets:15029998 errors:0 dropped:0 overruns:0 frame:0

TX packets:12761273 errors:0 dropped:0 overruns:0 carrier:0

collisions:0 txqueuelen:1000

RX bytes:12231984673 (11.3 GiB)  TX bytes:8970357010 (8.3 GiB)

eth1:1    Link encap:Ethernet  HWaddr 6C:92:BF:2A:7F:8B

inet addr:169.254.169.99Bcast:169.254.255.255Mask:255.255.0.0

UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1

lo        Link encap:Local Loopback

inet addr:127.0.0.1  Mask:255.0.0.0

inet6 addr: ::1/128 Scope:Host

UP LOOPBACK RUNNING  MTU:16436  Metric:1

RX packets:6453002 errors:0 dropped:0 overruns:0 frame:0

TX packets:6453002 errors:0 dropped:0 overruns:0 carrier:0

collisions:0 txqueuelen:0

RX bytes:6450405692 (6.0 GiB)  TX bytes:6450405692 (6.0 GiB)

(节点2)

[root@xsyy2 ~]# ifconfig

eth0      Link encap:Ethernet  HWaddr 6C:92:BF:2A:7C:D6

inet addr:172.16.28.28  Bcast:172.16.28.255  Mask:255.255.255.0

inet6 addr: fe80::6e92:bfff:fe2a:7cd6/64 Scope:Link

UP BROADCAST RUNNING MULTICAST  MTU:1500Metric:1

RX packets:12895198 errors:0 dropped:93 overruns:0 frame:0

TX packets:2639578 errors:0 dropped:0 overruns:0 carrier:0

collisions:0 txqueuelen:1000

RX bytes:17559461040 (16.3 GiB)  TX bytes:231832259 (221.0 MiB)

eth0:1    Link encap:Ethernet  HWaddr 6C:92:BF:2A:7C:D6

inet addr:172.16.28.53  Bcast:172.16.28.255  Mask:255.255.255.0

UP BROADCAST RUNNING MULTICAST  MTU:1500Metric:1

eth1      Link encap:Ethernet  HWaddr 6C:92:BF:2A:7C:D7

inet addr:10.1.20.28  Bcast:10.1.20.255  Mask:255.255.255.0

inet6 addr: fe80::6e92:bfff:fe2a:7cd7/64 Scope:Link

UP BROADCAST RUNNING MULTICAST  MTU:1500Metric:1

RX packets:12764698 errors:0 dropped:0 overruns:0 frame:0

TX packets:15034054 errors:0 dropped:0 overruns:0 carrier:0

collisions:0 txqueuelen:1000

RX bytes:8973128258 (8.3 GiB)  TX bytes:12234670411 (11.3 GiB)

eth1:1    Link encap:Ethernet  HWaddr 6C:92:BF:2A:7C:D7

inet addr:169.254.87.215  Bcast:169.254.255.255  Mask:255.255.0.0

UP BROADCAST RUNNING MULTICAST  MTU:1500Metric:1

lo        Link encap:Local Loopback

inet addr:127.0.0.1  Mask:255.0.0.0

inet6 addr: ::1/128 Scope:Host

UP LOOPBACK RUNNING  MTU:16436Metric:1

RX packets:2213075 errors:0 dropped:0 overruns:0 frame:0

TX packets:2213075 errors:0 dropped:0 overruns:0 carrier:0

collisions:0 txqueuelen:0

RX bytes:1213870175 (1.1 GiB)  TX bytes:1213870175 (1.1 GiB)

确认防火墙和SELinux是禁用的

(以Node1为例,两个节点相同)

[root@xsyy1 ~]# iptables -L

Chain INPUT (policy ACCEPT)

target     prot opt source               destination

Chain FORWARD (policy ACCEPT)

target     prot opt source               destination

Chain OUTPUT (policy ACCEPT)

target     prot opt source               destination

如果防火墙没禁用,那么通过如下方式修改:

[root@xsyy1 ~]# service iptables stop

[root@xsyy1 ~]# chkconfig iptables off

[root@xsyy1 ~]# getenforce

Disabled

如果SELinux没有被禁止,那就通过如下方式修改:

[root@xsyy1 ~]# cat /etc/selinux/config

-- 改成SELINUX=disabled

# This file controls the state of SELinux on the system.

# SELINUX= can take one of these three values:

#     enforcing - SELinux security policy is enforced.

#     permissive - SELinux prints warnings instead of enforcing.

#     disabled - No SELinux policy is loaded.

SELINUX=disabled

# SELINUXTYPE= can take one of these two values:

#     targeted - Targeted processes are protected,

#     mls - Multi Level Security protection.

SELINUXTYPE=targeted

临时关闭selinux

[root@xsyy1 ~]# setenforce 0

正式部署安装

1、配置/etc/hosts

修改/etc/hosts文件,前两行不动,添加hostname对应信息。

(以Node1为例,两个节点相同)

[root@xsyy1 ~]# cat /etc/hosts

127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4

::1         localhost localhost.localdomain localhost6 localhost6.localdomain6

# For Public

172.16.28.24   xsyy1

172.16.28.28   xsyy2

# For VIP

172.16.28.52    xsyyvip1

172.16.28.53    xsyyvip2

# For Private IP

10.1.20.24  xsyypriv1

10.1.20.28  xsyypriv2

# For SCAN IP

172.16.28.39 scanip

3、系统配置

【两个节点相同】

修改/etc/sysctl.conf,添加如下内容:

fs.file-max = 6815744

kernel.sem = 250 32000 100 128

kernel.shmmni = 4096

kernel.shmall = 1073741824

kernel.shmmax = 4398046511104

net.core.rmem_default = 262144

net.core.rmem_max = 4194304

net.core.wmem_default = 262144

net.core.wmem_max = 1048576

fs.aio-max-nr = 1048576

net.ipv4.ip_local_port_range = 9000 65500

生效:

[root@12c ~]# sysctl -p

 修改/etc/security/limits.conf,添加如下内容:

grid   soft   nofile    1024

grid   hard   nofile    65536

grid   soft   nproc    2047

grid   hard   nproc    16384

grid   soft   stack    10240

grid   hard   stack    32768

oracle   soft   nofile    1024

oracle   hard   nofile    65536

oracle   soft   nproc    2047

oracle   hard   nproc    16384

oracle   soft   stack    10240

oracle   hard   stack    32768

修改/etc/pam.d/login

在 /etc/pam.d/login 文件中添加或编辑下面一行内容:

session required pam_limits.so

修改/etc/profile

[root@rac1 ~]# cat /etc/profile

if [ $USER = "grid" ] || [ $USER = "oracle" ]; then

if [ $SHELL = "/bin/ksh" ]; then

ulimit -p 16384

ulimit -n 65536

else

ulimit -u 16384 -n 65536

fi

fi

4、配置YUM源并安装所需包

先将默认的yum配置文件删除或者移动,然后创建一个新的

(以Node1为例,两个节点相同)

[root@xsyy1 ~]# cd /etc/yum.repos.d

[root@xsyy1 yum.repos.d]# mkdir bk

[root@xsyy1 yum.repos.d]# mv public-yum-ol6.repo bk/

[root@xsyy1 yum.repos.d]# vi gdmsacs.repo

-- 添加如下内容

[Oracle]

name=OEL

baseurl=file:///mnt/Server

gpgcheck=0

enabled=1

将光驱挂载上

[root@xsyy1 yum.repos.d]# mount /dev/cdrom /mnt

mount: block device /dev/sr0 is write-protected, mounting read-only

下面就可以Yum方式安装所需包了

[root@xsyy1 yum.repos.d]# yum -y install binutils compat-libstdc++-33 elfutils-libelf \

elfutils-libelf-devel elfutils-libelf-devel-static gcc gcc-c++ glibc glibc-common \

glibc-devel kernel-headers ksh libaio libaio-devel libgcc libgomp libstdc++ libstdc++-devel \

make numactl-devel sysstat unixODBC unixODBC-devel pdksh compat-libcap1

5、创建用户和组

(以Node1为例,两个节点相同)

创建组:

[root@xsyy1 ~]# /usr/sbin/groupadd -g 54321 oinstall

[root@xsyy1 ~]# /usr/sbin/groupadd -g 54322 dba

[root@xsyy1 ~]# /usr/sbin/groupadd -g 54323 oper

[root@xsyy1 ~]# /usr/sbin/groupadd -g 54324 backupdba

[root@xsyy1 ~]# /usr/sbin/groupadd -g 54325 dgdba

[root@xsyy1 ~]# /usr/sbin/groupadd -g 54327 asmdba

[root@xsyy1 ~]# /usr/sbin/groupadd -g 54328 asmoper

[root@xsyy1 ~]# /usr/sbin/groupadd -g 54329 asmadmin

创建用户:

[root@xsyy1 ~]# /usr/sbin/useradd -u 54321 -g oinstall -G asmadmin,asmdba,asmoper,dba grid

[root@xsyy1 ~]# /usr/sbin/useradd -u 54322 -g oinstall -G dba,backupdba,dgdba,asmadmin oracle

设置口令:

[root@xsyy1 ~]# passwd grid

[root@xsyy1 ~]# passwd oracle

6、创建安装目录以及授权

(以Node1为例,两个节点相同)

[root@xsyy1 ~]# mkdir -p /u01/app/grid

[root@xsyy1 ~]# mkdir -p /u01/app/12.1.0/grid

[root@xsyy1 ~]# mkdir -p /u01/app/oracle/product/12.1.0/db_1

[root@xsyy1 ~]# chown -R grid.oinstall /u01

[root@xsyy1 ~]# chown -R oracle.oinstall /u01/app/oracle

[root@xsyy1 ~]# chmod -R 775 /u01

7、配置境变量

节点1: grid用户

[root@xsyy1 ~]# vi /home/grid/.bash_profile

export TMP=/tmp

export TMPDIR=$TMP

export ORACLE_HOSTNAME=xsyy1.gdmsa.gov.cn

export ORACLE_BASE=/u01/app/grid

export ORACLE_HOME=/u01/app/12.1.0/grid

export ORACLE_SID=+ASM1

export PATH=/usr/sbin:$PATH

export PATH=$ORACLE_HOME/bin:$PATH

export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib

export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib

alias sqlplus="rlwrap sqlplus"

#Oracle用户

[root@xsyy1 ~]# vi /home/oracle/.bash_profile

export PATH

export TMP=/tmp

export TMPDIR=$TMP

export ORACLE_BASE=/u01/app/oracle

export ORACLE_HOME=$ORACLE_BASE/product/12.1.0/db_1

export ORACLE_SID=gxshis1

export PATH=/usr/sbin:$PATH

export PATH=$ORACLE_HOME/bin:$PATH

export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib

export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib

节点2:

[root@xsyy2 ~]# vi /home/grid/.bash_profile

export TMP=/tmp

export TMPDIR=$TMP

export ORACLE_HOSTNAME=xsyy1.gdmsa.gov.cn

export ORACLE_BASE=/u01/app/grid

export ORACLE_HOME=/u01/app/12.1.0/grid

export ORACLE_SID=+ASM1

export PATH=/usr/sbin:$PATH

export PATH=$ORACLE_HOME/bin:$PATH

export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib

export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib

alias sqlplus="rlwrap sqlplus"

[root@xsyy2 ~]# vi /home/oracle/.bash_profile

export TMP=/tmp

export TMPDIR=$TMP

export ORACLE_BASE=/u01/app/oracle

export ORACLE_HOME=$ORACLE_BASE/product/12.1.0/db_1

export ORACLE_SID=gxshis2

export PATH=/usr/sbin:$PATH

export PATH=$ORACLE_HOME/bin:$PATH

export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib

export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib

8、Iscsi挂载磁盘并配置UDEV

[root@12cr ~]# yum -y install iscsi-initiator-utils

[root@12cr ~]# service iscsid start

[root@12cr ~]# chkconfig iscsid on

[root@12cr ~]# iscsiadm -m discovery -t sendtargets -p 192.168.80.140:3260

[  OK  ] iscsid: [OK  ]

192.168.80.140:3260,1 iqn.2006-01.com.openfiler:tsn.3a9cad78121d

[root@12cr ~]# service iscsi restart

Stopping iscsi: [  OK  ]

Starting iscsi: [  OK  ]

[root@xsyy1 ~]# fdisk -l

Disk /dev/sda: 53.7 GB, 53687091200 bytes

255 heads, 63 sectors/track, 6527 cylinders

Units = cylinders of 16065 * 512 = 8225280 bytes

Sector size (logical/physical): 512 bytes / 512 bytes

I/O size (minimum/optimal): 512 bytes / 512 bytes

Disk identifier: 0x00046ecd

Device Boot      Start         End      BlocksId  System

/dev/sda1   *           1          26      204800   83  Linux

Partition 1 does not end on cylinder boundary.

/dev/sda2              26         664     5120000   82  Linux swap / Solaris

Partition 2 does not end on cylinder boundary.

/dev/sda3             664        6528    47102976   83  Linux

Disk /dev/sdb: 2147 MB, 2147483648 bytes

67 heads, 62 sectors/track, 1009 cylinders

Units = cylinders of 4154 * 512 = 2126848 bytes

Sector size (logical/physical): 512 bytes / 512 bytes

I/O size (minimum/optimal): 512 bytes / 512 bytes

Disk identifier: 0x00000000

Disk /dev/sdc: 10.5 GB, 10502537216 bytes

64 heads, 32 sectors/track, 10016 cylinders

Units = cylinders of 2048 * 512 = 1048576 bytes

Sector size (logical/physical): 512 bytes / 512 bytes

I/O size (minimum/optimal): 512 bytes / 512 bytes

Disk identifier: 0x00000000

Disk /dev/sdd: 6610 MB, 6610223104 bytes

204 heads, 62 sectors/track, 1020 cylinders

Units = cylinders of 12648 * 512 = 6475776 bytes

Sector size (logical/physical): 512 bytes / 512 bytes

I/O size (minimum/optimal): 512 bytes / 512 bytes

Disk identifier: 0x00000000

Disk /dev/sdf: 8388 MB, 8388608000 bytes

64 heads, 32 sectors/track, 8000 cylinders

Units = cylinders of 2048 * 512 = 1048576 bytes

Sector size (logical/physical): 512 bytes / 512 bytes

I/O size (minimum/optimal): 512 bytes / 512 bytes

Disk identifier: 0x00000000

Disk /dev/sde: 8388 MB, 8388608000 bytes

64 heads, 32 sectors/track, 8000 cylinders

Units = cylinders of 2048 * 512 = 1048576 bytes

Sector size (logical/physical): 512 bytes / 512 bytes

I/O size (minimum/optimal): 512 bytes / 512 bytes

Disk identifier: 0x00000000

Disk /dev/sdg: 5335 MB, 5335154688 bytes

165 heads, 62 sectors/track, 1018 cylinders

Units = cylinders of 10230 * 512 = 5237760 bytes

Sector size (logical/physical): 512 bytes / 512 bytes

I/O size (minimum/optimal): 512 bytes / 512 bytes

Disk identifier: 0x00000000

这里只用了sde、sdf、sdg,其他的是给别的集群使用的。

//i in b c d e f g h;

[root@xsyy1 ~]# for i in e f g;do

> echo "KERNEL==\"sd*\", BUS==\"scsi\", PROGRAM==\"/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/\$name\", RESULT==\"`/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/sd$i`\", NAME=\"asm-disk$i\", OWNER=\"grid\", GROUP=\"asmadmin\", MODE=\"0660\""

> done

KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="14f504e46494c450058444273784d2d64705a6a2d544c4756", NAME="asm-diske", OWNER="grid", GROUP="asmadmin", MODE="0660"

KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="14f504e46494c45003030365263642d32714a702d6866744c", NAME="asm-diskf", OWNER="grid", GROUP="asmadmin", MODE="0660"

KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="14f504e46494c45006c58576a76452d716d50492d71436c76", NAME="asm-diskg", OWNER="grid", GROUP="asmadmin", MODE="0660"

配置UDEV: 将上述内容拷贝至节点

(以Node1为例,两个节点相同)

[root@xsyy1 ~]# vi /etc/udev/rules.d/99-oracle-asmdevices.rules

-- 添加如下内容:

KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="14f504e46494c45003030365263642d32714a702d6866744c", NAME="asm-data", OWNER="grid", GROUP="asmadmin", MODE="0660"

KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="14f504e46494c450058444273784d2d64705a6a2d544c4756", NAME="asm-fra", OWNER="grid", GROUP="asmadmin", MODE="0660"

KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="14f504e46494c45006c58576a76452d716d50492d71436c76", NAME="asm-crs", OWNER="grid", GROUP="asmadmin", MODE="0660"

--重启UDEV:

[root@xsyy1 ~]# /sbin/start_udev

Starting udev: [  OK  ]

--检查共享设备的所属关系和权限:

[root@xsyy1 ~]# ls -l /dev/asm*

brw-rw---- 1 grid asmadmin 8, 96 Jun 29 21:56 /dev/asm-crs

brw-rw---- 1 grid asmadmin 8, 64 Jun 29 21:56 /dev/asm-data

brw-rw---- 1 grid asmadmin 8, 80 Jun 29 21:56 /dev/asm-fra

9、禁用NTP服务

(以Node1为例,两个节点相同)

[root@xsyy1 ~]# chkconfig ntpd off

[root@xsyy1 ~]# mv /etc/ntp.conf /etc/ntp.conf.bak

[root@xsyy1 ~]# rm /var/run/ntpd.pid

10、解压介质

节点1:

[root@xsyy1 ~]# chown -R grid.oinstall /install/

[root@xsyy1 ~]# chown oracle.oinstall /install/linuxamd64_12c_database_*

[root@xsyy1 ~]# chmod 775 /install

[root@xsyy1 ~]# su - grid

[root@xsyy1 ~]# cd /install/

[grid@xsyy1 install]$ unzip linuxamd64_12c_grid_1of2.zip

[grid@xsyy1 install]$ unzip linuxamd64_12c_grid_2of2.zip

[root@xsyy1 ~]# su - oracle

[oracle@xsyy1 install]$ unzip linuxamd64_12c_database_1of2.zip

解压之后大小为:

[oracle@xsyy1 install]$ du -sh grid

2.1G    grid

[oracle@xsyy1 install]$ du -sh database/

2.6G    database/

安装cvu相关rpm包:

[root@xsyy1 ~]# cd /install/grid/rpm/

[root@xsyy1 rpm]# rpm -ivh cvuqdisk-1.0.9-1.rpm

Preparing...########################################### [100%]

Using default group oinstall to install package

1:cvuqdisk########################################### [100%]

拷贝到节点2并安装:

[root@xsyy1 rpm]# scp cvuqdisk-1.0.9-1.rpm xsyy2:/install

[root@xsyy2 install]# rpm -ivh cvuqdisk-1.0.9-1.rpm

Preparing...########################################### [100%]

Using default group oinstall to install package

1:cvuqdisk               ########################################### [100%]

SSH配置

/usr/bin/ssh-keygen -t rsa

/usr/bin/ssh-keygen -t dsa

把4个文件合成一个文件

注意权限要是是644

chmod 644 authorized_keys

在节点RAC2进行的操作

[oracle@RAC2 ~]$ ls -a

.   .bash_history  .bash_profile.emacs  .mozilla  .viminfo

..  .bash_logout   .bashrc        .kde.ssh      .Xauthority

[oracle@RAC2 ~]$ cd .ssh

[oracle@RAC2 .ssh]$ ls

id_dsa  id_dsa.pub  id_rsaid_rsa.pub

[oracle@RAC2 .ssh]$ scp id_dsa.pub rac1:/home/oracle/.ssh/id_dsa.pub2

oracle@rac1's password:

id_dsa.pub100%  601     0.6KB/s00:00

[oracle@RAC2 .ssh]$ ls

id_dsa  id_dsa.pub  id_rsaid_rsa.pub  known_hosts

[oracle@RAC2 .ssh]$ scp id_rsa.pub rac1:/home/oracle/.ssh/id_rsa.pub2

oracle@rac1's password:

id_rsa.pub

再到RAC1节点

cat id_dsa.pubid_dsa.pub2  id_rsa.pub  id_rsa.pub2 > authorized_keys

scp authorized_keysrac2:/home/oracle/.ssh

之后要测试一下(如果都不需要密码,得到时间后。就说明ssh 对等性配置成功。)

 

 

11、校验

这里只贴失败的项,其中第一个是物理内存不足问题,Oracle推荐每节点至少4GB内存空间,我这里只有2G;第二个问题是配置DNS,这个问题我们可以忽略。

节点1:

[grid@xsyy1 grid]$ ./runcluvfy.sh stage -pre crsinst -n xsyy1,xsyy2 -verbose

Check: Total memory

Node Name     Available                 Required                  Status

------------  ------------------------  ------------------------  ----------

xsyy2       1.9567GB (2051748.0KB)    4GB (4194304.0KB)         failed

xsyy1       1.9567GB (2051748.0KB)    4GB (4194304.0KB)         failed

Result: Total memory check failed

Result: Default user file creation mask check passed

Checking integrity of file "/etc/resolv.conf" across nodes

Checking the file "/etc/resolv.conf" to make sure only one of domain and search entries is defined

"domain" and "search" entries do not coexist in any  "/etc/resolv.conf" file

Checking if domain entry in file "/etc/resolv.conf" is consistent across the nodes...

"domain" entry does not exist in any "/etc/resolv.conf" file

Checking if search entry in file "/etc/resolv.conf" is consistent across the nodes...

Checking file "/etc/resolv.conf" to make sure that only one search entry is defined

More than one "search" entry does not exist in any "/etc/resolv.conf" file

All nodes have same "search" order defined in file "/etc/resolv.conf"

Checking DNS response time for an unreachable node

Node Name                             Status

------------------------------------------------------------

xsyy1                               failed

xsyy2                               failed

PRVF-5636 : The DNS response time for an unreachable node exceeded "15000" ms on following nodes: xsyy1,xsyy2

Check for integrity of file "/etc/resolv.conf" failed

12、安装

1)安装GI

[root@12cr ~]# su - grid

[grid@12cr ~]$ cd /install/grid/

我打开Xmanager - Passive,设置DISPLAY,调用runInstaller启动OUI

[grid@12cr grid]$ export DISPLAY=192.168.1.1:0.0

[grid@12cr grid]$ ./runInstaller

这里有几项校验不过去,这几个都忽略。

脚本输出内容:

[root@xsyy1 ~]# /u01/app/oraInventory/orainstRoot.sh

Changing permissions of /u01/app/oraInventory.

Adding read,write permissions for group.

Removing read,write,execute permissions for world.

Changing groupname of /u01/app/oraInventory to oinstall.

The execution of the script is complete.

节点1:

[root@xsyy1 ~]# /u01/app/12.1.0/grid/root.sh

Performing root user operation for Oracle 12c

The following environment variables are set as:

ORACLE_OWNER= grid

ORACLE_HOME=  /u01/app/12.1.0/grid

Enter the full pathname of the local bin directory: [/usr/local/bin]:

Copying dbhome to /usr/local/bin ...

Copying oraenv to /usr/local/bin ...

Copying coraenv to /usr/local/bin ...

Creating /etc/oratab file...

Entries will be added to the /etc/oratab file as needed by

Database Configuration Assistant when a database is created

Finished running generic part of root script.

Now product-specific root actions will be performed.

Using configuration parameter file: /u01/app/12.1.0/grid/crs/install/crsconfig_params

2013/07/01 00:30:25 CLSRSC-363: User ignored prerequisites during installation

OLR initialization - successful

root wallet

root wallet cert

root cert export

peer wallet

profile reader wallet

pa wallet

peer wallet keys

pa wallet keys

peer cert request

pa cert request

peer cert

pa cert

peer root cert TP

profile reader root cert TP

pa root cert TP

peer pa cert TP

pa peer cert TP

profile reader pa cert TP

profile reader peer cert TP

peer user cert

pa user cert

2013/07/01 00:31:22 CLSRSC-330: Adding Clusterware entries to file 'oracle-ohasd.conf'

CRS-4133: Oracle High Availability Services has been stopped.

CRS-4123: Oracle High Availability Services has been started.

CRS-4133: Oracle High Availability Services has been stopped.

CRS-4123: Oracle High Availability Services has been started.

CRS-2673: Attempting to stop 'ora.drivers.acfs' on 'xsyy1'

CRS-2677: Stop of 'ora.drivers.acfs' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.evmd' on 'xsyy1'

CRS-2672: Attempting to start 'ora.mdnsd' on 'xsyy1'

CRS-2676: Start of 'ora.mdnsd' on 'xsyy1' succeeded

CRS-2676: Start of 'ora.evmd' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.gpnpd' on 'xsyy1'

CRS-2676: Start of 'ora.gpnpd' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.cssdmonitor' on 'xsyy1'

CRS-2672: Attempting to start 'ora.gipcd' on 'xsyy1'

CRS-2676: Start of 'ora.cssdmonitor' on 'xsyy1' succeeded

CRS-2676: Start of 'ora.gipcd' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.cssd' on 'xsyy1'

CRS-2672: Attempting to start 'ora.diskmon' on 'xsyy1'

CRS-2676: Start of 'ora.diskmon' on 'xsyy1' succeeded

CRS-2676: Start of 'ora.cssd' on 'xsyy1' succeeded

ASM created and started successfully.

Disk Group RACCRS created successfully.

CRS-2672: Attempting to start 'ora.crf' on 'xsyy1'

CRS-2672: Attempting to start 'ora.storage' on 'xsyy1'

CRS-2676: Start of 'ora.storage' on 'xsyy1' succeeded

CRS-2676: Start of 'ora.crf' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.crsd' on 'xsyy1'

CRS-2676: Start of 'ora.crsd' on 'xsyy1' succeeded

CRS-4256: Updating the profile

Successful addition of voting disk d883c23a7bfc4fdcbf418c9f631bd0af.

Successfully replaced voting disk group with +RACCRS.

CRS-4256: Updating the profile

CRS-4266: Voting file(s) successfully replaced

##  STATE    File Universal Id                File Name Disk group

--  -----    -----------------                --------- ---------

1. ONLINE   d883c23a7bfc4fdcbf418c9f631bd0af (/dev/asm-crs) [RACCRS]

Located 1 voting disk(s).

CRS-2791: Starting shutdown of Oracle High Availability Services-managed resources on 'xsyy1'

CRS-2673: Attempting to stop 'ora.crsd' on 'xsyy1'

CRS-2677: Stop of 'ora.crsd' on 'xsyy1' succeeded

CRS-2673: Attempting to stop 'ora.storage' on 'xsyy1'

CRS-2673: Attempting to stop 'ora.mdnsd' on 'xsyy1'

CRS-2673: Attempting to stop 'ora.gpnpd' on 'xsyy1'

CRS-2673: Attempting to stop 'ora.drivers.acfs' on 'xsyy1'

CRS-2677: Stop of 'ora.storage' on 'xsyy1' succeeded

CRS-2673: Attempting to stop 'ora.ctssd' on 'xsyy1'

CRS-2673: Attempting to stop 'ora.evmd' on 'xsyy1'

CRS-2673: Attempting to stop 'ora.asm' on 'xsyy1'

CRS-2677: Stop of 'ora.drivers.acfs' on 'xsyy1' succeeded

CRS-2677: Stop of 'ora.mdnsd' on 'xsyy1' succeeded

CRS-2677: Stop of 'ora.gpnpd' on 'xsyy1' succeeded

CRS-2677: Stop of 'ora.evmd' on 'xsyy1' succeeded

CRS-2677: Stop of 'ora.ctssd' on 'xsyy1' succeeded

CRS-2677: Stop of 'ora.asm' on 'xsyy1' succeeded

CRS-2673: Attempting to stop 'ora.cluster_interconnect.haip' on 'xsyy1'

CRS-2677: Stop of 'ora.cluster_interconnect.haip' on 'xsyy1' succeeded

CRS-2673: Attempting to stop 'ora.cssd' on 'xsyy1'

CRS-2677: Stop of 'ora.cssd' on 'xsyy1' succeeded

CRS-2673: Attempting to stop 'ora.crf' on 'xsyy1'

CRS-2677: Stop of 'ora.crf' on 'xsyy1' succeeded

CRS-2673: Attempting to stop 'ora.gipcd' on 'xsyy1'

CRS-2677: Stop of 'ora.gipcd' on 'xsyy1' succeeded

CRS-2793: Shutdown of Oracle High Availability Services-managed resources on 'xsyy1' has completed

CRS-4133: Oracle High Availability Services has been stopped.

CRS-4123: Starting Oracle High Availability Services-managed resources

CRS-2672: Attempting to start 'ora.mdnsd' on 'xsyy1'

CRS-2672: Attempting to start 'ora.evmd' on 'xsyy1'

CRS-2676: Start of 'ora.mdnsd' on 'xsyy1' succeeded

CRS-2676: Start of 'ora.evmd' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.gpnpd' on 'xsyy1'

CRS-2676: Start of 'ora.gpnpd' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.gipcd' on 'xsyy1'

CRS-2676: Start of 'ora.gipcd' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.cssdmonitor' on 'xsyy1'

CRS-2676: Start of 'ora.cssdmonitor' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.cssd' on 'xsyy1'

CRS-2672: Attempting to start 'ora.diskmon' on 'xsyy1'

CRS-2676: Start of 'ora.diskmon' on 'xsyy1' succeeded

CRS-2789: Cannot stop resource 'ora.diskmon' as it is not running on server 'xsyy1'

CRS-2676: Start of 'ora.cssd' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.cluster_interconnect.haip' on 'xsyy1'

CRS-2672: Attempting to start 'ora.ctssd' on 'xsyy1'

CRS-2676: Start of 'ora.ctssd' on 'xsyy1' succeeded

CRS-2676: Start of 'ora.cluster_interconnect.haip' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.asm' on 'xsyy1'

CRS-2676: Start of 'ora.asm' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.storage' on 'xsyy1'

CRS-2676: Start of 'ora.storage' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.crf' on 'xsyy1'

CRS-2676: Start of 'ora.crf' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.crsd' on 'xsyy1'

CRS-2676: Start of 'ora.crsd' on 'xsyy1' succeeded

CRS-6023: Starting Oracle Cluster Ready Services-managed resources

CRS-6017: Processing resource auto-start for servers: xsyy1

CRS-6016: Resource auto-start has completed for server xsyy1

CRS-6024: Completed start of Oracle Cluster Ready Services-managed resources

CRS-4123: Oracle High Availability Services has been started.

2013/07/01 00:38:01 CLSRSC-343: Successfully started Oracle clusterware stack

CRS-2672: Attempting to start 'ora.asm' on 'xsyy1'

CRS-2676: Start of 'ora.asm' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.RACCRS.dg' on 'xsyy1'

CRS-2676: Start of 'ora.RACCRS.dg' on 'xsyy1' succeeded

2013/07/01 00:39:51 CLSRSC-325: Configure Oracle Grid Infrastructure for a Cluster ... succeeded

节点2:

[root@xsyy2 ~]# /u01/app/12.1.0/grid/root.sh

Performing root user operation for Oracle 12c

The following environment variables are set as:

ORACLE_OWNER= grid

ORACLE_HOME=  /u01/app/12.1.0/grid

Enter the full pathname of the local bin directory: [/usr/local/bin]:

Copying dbhome to /usr/local/bin ...

Copying oraenv to /usr/local/bin ...

Copying coraenv to /usr/local/bin ...

Creating /etc/oratab file...

Entries will be added to the /etc/oratab file as needed by

Database Configuration Assistant when a database is created

Finished running generic part of root script.

Now product-specific root actions will be performed.

Using configuration parameter file: /u01/app/12.1.0/grid/crs/install/crsconfig_params

2013/07/01 00:42:51 CLSRSC-363: User ignored prerequisites during installation

OLR initialization - successful

2013/07/01 00:43:18 CLSRSC-330: Adding Clusterware entries to file 'oracle-ohasd.conf'

CRS-4133: Oracle High Availability Services has been stopped.

CRS-4123: Oracle High Availability Services has been started.

CRS-4133: Oracle High Availability Services has been stopped.

CRS-4123: Oracle High Availability Services has been started.

CRS-2791: Starting shutdown of Oracle High Availability Services-managed resources on 'xsyy2'

CRS-2673: Attempting to stop 'ora.drivers.acfs' on 'xsyy2'

CRS-2677: Stop of 'ora.drivers.acfs' on 'xsyy2' succeeded

CRS-2793: Shutdown of Oracle High Availability Services-managed resources on 'xsyy2' has completed

CRS-4133: Oracle High Availability Services has been stopped.

CRS-4123: Starting Oracle High Availability Services-managed resources

CRS-2672: Attempting to start 'ora.evmd' on 'xsyy2'

CRS-2672: Attempting to start 'ora.mdnsd' on 'xsyy2'

CRS-2676: Start of 'ora.evmd' on 'xsyy2' succeeded

CRS-2676: Start of 'ora.mdnsd' on 'xsyy2' succeeded

CRS-2672: Attempting to start 'ora.gpnpd' on 'xsyy2'

CRS-2676: Start of 'ora.gpnpd' on 'xsyy2' succeeded

CRS-2672: Attempting to start 'ora.gipcd' on 'xsyy2'

CRS-2676: Start of 'ora.gipcd' on 'xsyy2' succeeded

CRS-2672: Attempting to start 'ora.cssdmonitor' on 'xsyy2'

CRS-2676: Start of 'ora.cssdmonitor' on 'xsyy2' succeeded

CRS-2672: Attempting to start 'ora.cssd' on 'xsyy2'

CRS-2672: Attempting to start 'ora.diskmon' on 'xsyy2'

CRS-2676: Start of 'ora.diskmon' on 'xsyy2' succeeded

CRS-2789: Cannot stop resource 'ora.diskmon' as it is not running on server 'xsyy2'

CRS-2676: Start of 'ora.cssd' on 'xsyy2' succeeded

CRS-2672: Attempting to start 'ora.cluster_interconnect.haip' on 'xsyy2'

CRS-2672: Attempting to start 'ora.ctssd' on 'xsyy2'

CRS-2676: Start of 'ora.ctssd' on 'xsyy2' succeeded

CRS-2676: Start of 'ora.cluster_interconnect.haip' on 'xsyy2' succeeded

CRS-2672: Attempting to start 'ora.asm' on 'xsyy2'

CRS-2676: Start of 'ora.asm' on 'xsyy2' succeeded

CRS-2672: Attempting to start 'ora.storage' on 'xsyy2'

CRS-2676: Start of 'ora.storage' on 'xsyy2' succeeded

CRS-2672: Attempting to start 'ora.crf' on 'xsyy2'

CRS-2676: Start of 'ora.crf' on 'xsyy2' succeeded

CRS-2672: Attempting to start 'ora.crsd' on 'xsyy2'

CRS-2676: Start of 'ora.crsd' on 'xsyy2' succeeded

CRS-6017: Processing resource auto-start for servers: xsyy2

CRS-2673: Attempting to stop 'ora.LISTENER_SCAN1.lsnr' on 'xsyy1'

CRS-2672: Attempting to start 'ora.ons' on 'xsyy2'

CRS-2677: Stop of 'ora.LISTENER_SCAN1.lsnr' on 'xsyy1' succeeded

CRS-2673: Attempting to stop 'ora.scan1.vip' on 'xsyy1'

CRS-2677: Stop of 'ora.scan1.vip' on 'xsyy1' succeeded

CRS-2672: Attempting to start 'ora.scan1.vip' on 'xsyy2'

CRS-2676: Start of 'ora.scan1.vip' on 'xsyy2' succeeded

CRS-2672: Attempting to start 'ora.LISTENER_SCAN1.lsnr' on 'xsyy2'

CRS-2676: Start of 'ora.ons' on 'xsyy2' succeeded

CRS-2676: Start of 'ora.LISTENER_SCAN1.lsnr' on 'xsyy2' succeeded

CRS-6016: Resource auto-start has completed for server xsyy2

CRS-6024: Completed start of Oracle Cluster Ready Services-managed resources

CRS-4123: Oracle High Availability Services has been started.

2013/07/01 00:48:43 CLSRSC-343: Successfully started Oracle clusterware stack

2013/07/01 00:49:05 CLSRSC-325: Configure Oracle Grid Infrastructure for a Cluster ... succeeded

Oracle Cluster Verification 将在Oracle Database 安装中完成。

忽略错误,至此GRID 安装完成

查看状态:

检查scan配置

[grid@xsyy1 ~]$ cluvfy comp scan

[grid@xsyy1 ~]$ crsctl check cluster

[grid@xsyy1 ~]$ crsctl stat res -t

--------------------------------------------------------------------------------

Name           Target  StateServer                   State details

--------------------------------------------------------------------------------

Local Resources

--------------------------------------------------------------------------------

ora.LISTENER.lsnr

ONLINE  ONLINExsyy1                  STABLE

ONLINE  ONLINExsyy2                  STABLE

ora.RACCRS.dg

ONLINEONLINE       xsyy1                  STABLE

ONLINE  ONLINExsyy2                  STABLE

ora.asm

ONLINE  ONLINExsyy1Started,STABLE

ONLINE  ONLINExsyy2                  Started,STABLE

ora.net1.network

ONLINE  ONLINExsyy1                  STABLE

ONLINE  ONLINExsyy2                  STABLE

ora.ons

ONLINE  ONLINExsyy1                  STABLE

ONLINEONLINE       xsyy2                  STABLE

--------------------------------------------------------------------------------

Cluster Resources

--------------------------------------------------------------------------------

ora.xsyy1.vip

1        ONLINE  ONLINExsyy1                  STABLE

ora.xsyy2.vip

1        ONLINE  ONLINExsyy2                  STABLE

ora.LISTENER_SCAN1.lsnr

1        ONLINE  ONLINExsyy2                  STABLE

ora.LISTENER_SCAN2.lsnr

1        ONLINE  ONLINExsyy1                  STABLE

ora.LISTENER_SCAN3.lsnr

1        ONLINE  ONLINExsyy1                  STABLE

ora.MGMTLSNR

1        ONLINE  ONLINExsyy1                  169.254.88.173 192.1

68.80.150,STABLE

ora.cvu

1        ONLINE  ONLINExsyy1                  STABLE

ora.mgmtdb

1        ONLINE  ONLINExsyy1                  Open,STABLE

ora.oc4j

1        ONLINEONLINE       xsyy1                  STABLE

ora.scan1.vip

1        ONLINE  ONLINExsyy2                  STABLE

ora.scan2.vip

1        ONLINE  ONLINExsyy1                  STABLE

ora.scan3.vip

1        ONLINE  ONLINExsyy1                  STABLE

--------------------------------------------------------------------------------

如果此前配置了ssh,需要在此配置load ssh key

[oracle@xsyy1 ~]$ exec /usr/bin/ssh-agent $SHELL

[oracle@xsyy1 ~]$ /usr/bin/ssh-add

2)创建ASM磁盘组

节点1:

[grid@xsyy1 ~]$ export DISPLAY=192.168.1.1:0.0

[grid@xsyy1 ~]$ asmca

建立data和fra 磁盘组

3)安装RDBMS软件

节点1:

[root@xsyy1 ~]# su - oracle

[oracle@xsyy1 ~]$ cd /install/database/

[oracle@xsyy1 database]$ export DISPLAY=192.168.1.1:0.0

[oracle@xsyy1 database]$ ./runInstaller

执行脚本:

节点1:

[root@xsyy1 ~]# /u01/app/oracle/product/12.1.0/dbhome_1/root.sh

Performing root user operation for Oracle 12c

The following environment variables are set as:

ORACLE_OWNER= oracle

ORACLE_HOME=  /u01/app/oracle/product/12.1.0/dbhome_1

Enter the full pathname of the local bin directory: [/usr/local/bin]:

The contents of "dbhome" have not changed. No need to overwrite.

The contents of "oraenv" have not changed. No need to overwrite.

The contents of "coraenv" have not changed. No need to overwrite.

Entries will be added to the /etc/oratab file as needed by

Database Configuration Assistant when a database is created

Finished running generic part of root script.

Now product-specific root actions will be performed.

4)创建数据库

[oracle@xsyy1 ~]$ dbca

选择server pool为两个节点

勾上此选项,系统自动完善此前grid安装中的未执行的步骤.

点浏览选择ASM磁盘组

字符集根据需要进行选择

13、最终结果

资源状态:

[grid@xsyy1 ~]$ crsctl stat res -t

--------------------------------------------------------------------------------

Name           Target  StateServer                   State details

--------------------------------------------------------------------------------

Local Resources

--------------------------------------------------------------------------------

ora.LISTENER.lsnr

ONLINE  ONLINExsyy1                  STABLE

ONLINE  ONLINExsyy2                  STABLE

ora.RACCRS.dg

ONLINE  ONLINExsyy1                  STABLE

ONLINE  ONLINExsyy2                  STABLE

ora.RACDATA.dg

ONLINE  ONLINExsyy1                  STABLE

ONLINE  ONLINExsyy2                  STABLE

ora.RACFRA.dg

ONLINE  ONLINExsyy1                  STABLE

ONLINE  ONLINExsyy2                  STABLE

ora.asm

ONLINE  ONLINExsyy1Started,STABLE

ONLINE  ONLINExsyy2Started,STABLE

ora.net1.network

ONLINE  ONLINExsyy1                  STABLE

ONLINE  ONLINExsyy2                  STABLE

ora.ons

ONLINE  ONLINExsyy1                  STABLE

ONLINE  ONLINExsyy2                  STABLE

--------------------------------------------------------------------------------

Cluster Resources

--------------------------------------------------------------------------------

ora.xsyy1.vip

1        ONLINE  ONLINExsyy1                  STABLE

ora.xsyy2.vip

1        ONLINE  ONLINExsyy2                  STABLE

ora.LISTENER_SCAN1.lsnr

1        ONLINE  ONLINExsyy2                  STABLE

ora.LISTENER_SCAN2.lsnr

1        ONLINE  ONLINExsyy1                  STABLE

ora.LISTENER_SCAN3.lsnr

1        ONLINE  ONLINExsyy1                  STABLE

ora.MGMTLSNR

1        ONLINE  ONLINExsyy1                  169.254.88.173 192.1

68.80.150,STABLE

ora.cvu

1        ONLINE  ONLINExsyy1                  STABLE

ora.gdmsacs12c.db

1        ONLINE  ONLINExsyy1                  Open,STABLE

2        ONLINE  ONLINExsyy2                  Open,STABLE

ora.mgmtdb

1        ONLINE  ONLINExsyy1                  Open,STABLE

ora.oc4j

1        ONLINE  ONLINExsyy1                  STABLE

ora.scan1.vip

1        ONLINE  ONLINExsyy2                  STABLE

ora.scan2.vip

1        ONLINE  ONLINExsyy1                  STABLE

ora.scan3.vip

1        ONLINE  ONLINExsyy1                  STABLE

--------------------------------------------------------------------------------

RAC数据库配置信息

[grid@xsyy1 ~]$ srvctl config database -d gdmsacs12c

Database unique name: gdmsacs12c

Database name: gdmsacs12c

Oracle home: /u01/app/oracle/product/12.1.0/dbhome_1

Oracle user: oracle

Spfile: +RACDATA/gdmsacs12c/spfilegdmsacs12c.ora

Password file: +RACDATA/gdmsacs12c/orapwgdmsacs12c

Domain:

Start options: open

Stop options: immediate

Database role: PRIMARY

Management policy: AUTOMATIC

Server pools: gdmsacs12c

Database instances: gdmsacs12c1,gdmsacs12c2

Disk Groups: RACFRA,RACDATA

Mount point paths:

Services:

Type: RAC

Start concurrency:

Stop concurrency:

Database is administrator managed

[grid@xsyy1 ~]$ srvctl status database -d gdmsacs12c

Instance gdmsacs12c1 is running on node xsyy1

Instance gdmsacs12c2 is running on node xsyy2

[grid@xsyy1 ~]$ srvctl status listener

Listener LISTENER is enabled

Listener LISTENER is running on node(s): xsyy1,xsyy2

实例状态:

sys@GDMSACS12C> select instance_name, status from gv$instance;

INSTANCE_NAME    STATUS

---------------- ------------

gdmsacs12c1        OPEN

gdmsacs12c2        OPEN

sys@GDMSACS12C> col open_time for a25

sys@GDMSACS12C> col name for a10

sys@GDMSACS12C> select CON_ID, NAME, OPEN_MODE, OPEN_TIME, CREATE_SCN, TOTAL_SIZE from v$pdbs;

CON_ID NAME       OPEN_MODE  OPEN_TIME                 CREATE_SCN TOTAL_SIZE

---------- ---------- ---------- ------------------------- ---------- ----------

2 PDB$SEED   READ ONLY  01-JUL-13 04.33.07.302 PM    1720772  283115520

3 GDMSACS      READ WRITE 01-JUL-13 04.38.41.339 PM    1934986  288358400

查看归档启用与否:

sys@GDMSACS12C> archive log list

Database log mode              No Archive Mode

Automatic archivalDisabled

Archive destinationUSE_DB_RECOVERY_FILE_DEST

Oldest online log sequence     15

Current log sequence           16

现在手动开启:

[oracle@xsyy1 ~]$ srvctl stop database -d gdmsacs12c

[oracle@xsyy1 ~]$ srvctl start database -d gdmsacs12c -o mount

节点1:

[oracle@xsyy1 ~]$ sqlplus / as sysdba

SQL*Plus: Release 12.1.0.1.0 Production on Mon Jul 1 17:13:58 2014

Copyright (c) 1982, 2014, Oracle.  All rights reserved.

Connected to:

Oracle Database 12c Enterprise Edition Release 12.1.0.1.0 - 64bit Production

With the Partitioning, Real Application Clusters, Automatic Storage Management, OLAP,

Advanced Analytics and Real Application Testing options

idle> alter database archivelog;

Database altered.

idle> alter database open;

Database altered.

节点2:

[oracle@xsyy2 ~]$ sqlplus / as sysdba

SQL*Plus: Release 12.1.0.1.0 Production on Mon Jul 1 17:17:07 2014

Copyright (c) 1982, 2014, Oracle.  All rights reserved.

Connected to:

Oracle Database 12c Enterprise Edition Release 12.1.0.1.0 - 64bit Production

With the Partitioning, Real Application Clusters, Automatic Storage Management, OLAP,

Advanced Analytics and Real Application Testing options

SQL> alter database open;

Database altered.

这样RAC已经运行于归档模式了

sys@GDMSACS12C> archive log list

Database log mode              Archive Mode

Automatic archivalEnabled

Archive destinationUSE_DB_RECOVERY_FILE_DEST

Oldest online log sequence     15

Next log sequence to archive   16

Current log sequence           16

附录A

 

【节点1】【节点2】相同操作。

#yum install smartmontools

#yum install oracleasm-support

先对硬盘分区 fdisk /dev/sdb …… (根据需要  sde  sdf sdg 均同样操作)

#oracle asm configure –i

建立 ASM 盘

在每个节点: #oracleasm scandisks

习水医院12C RAC 数据库安装文档的更多相关文章

  1. Oracle 12c RAC 静默安装文档

    参考文档: https://docs.oracle.com/en/database/oracle/oracle-database/12.2/cwlin/index.html https://docs. ...

  2. ORACLE LINUX 6.3 + ORACLE 11.2.0.3 RAC + VBOX安装文档

    ORACLE LINUX 6.3 + ORACLE 11.2.0.3 RAC + VBOX安装文档 2015-10-21 12:51 525人阅读 评论(0) 收藏 举报  分类: Oracle RA ...

  3. oracle database 12c R1 安装文档

    INSTALLORACLE DATABASE 12C 完整的安装文档下载地址: http://download.csdn.net/detail/royjj/5665869 OS:ORALCE LINU ...

  4. Oracle 12c RAC 安装文档

    参考文档: https://docs.oracle.com/en/database/oracle/oracle-database/12.2/cwlin/index.html https://docs. ...

  5. 转载:【Oracle 集群】RAC知识图文详细教程(八)--Oracle 11G RAC数据库安装

    文章导航 集群概念介绍(一) ORACLE集群概念和原理(二) RAC 工作原理和相关组件(三) 缓存融合技术(四) RAC 特殊问题和实战经验(五) ORACLE 11 G版本2 RAC在LINUX ...

  6. Oracle 11g 单实例安装文档

    这里介绍在Red Hat Enterprise Linux Server release 5.7 (Tikanga)下安装ORACLE 11.2.0.1.0的过程,本文仅仅是为了写这样安装指导文档而整 ...

  7. Net 通用权限管理系统源码 带数据库设计文档,部署说明文档

    Net 通用权限管理系统源码 带数据库设计文档,部署说明文档 包括数据库设计文档部署安装文档源码数据库文件 下载地址:http://www.mallhd.com/archives/1389

  8. _00024 尼娜抹微笑伊拉克_云计算ClouderaManager以及CHD5.1.0群集部署安装文档V1.0

    笔者博文:妳那伊抹微笑 itdog8 地址链接 : http://www.itdog8.com(个人链接) 博客地址:http://blog.csdn.net/u012185296 博文标题:_000 ...

  9. mysql数据库迁移文档

    数据库迁移文档 一.需求 确保数据库稳定的运行,为开发人员提供方便的测试数据库和生产数据库的环境. 二.数据库整体架构(master/slave) 1.slave数据库安装 rpm -Uvh http ...

随机推荐

  1. java 使用xom对象数据序列化为xml、反序列化、Preferences相关操作小案例

    package org.rui.io.xml; import java.io.BufferedOutputStream; import java.io.FileOutputStream; import ...

  2. Linux设备驱动--块设备(四)之“自造请求”

    前面, 我们已经讨论了内核所作的在队列中优化请求顺序的工作; 这个工作包括排列请求和, 或许, 甚至延迟队列来允许一个预期的请求到达. 这些技术在处理一个真正的旋转的磁盘驱动器时有助于系统的性能. 但 ...

  3. linux下的git安装及配置

    一.yum安装方式 1.安装 $ yum install curl-devel expat-devel gettext-devel openssl-devel zlib-devel $ yum -y ...

  4. 给Xcode增加复制行、删除行快捷键的方法

    运行: sudo chmod 666 /Applications/Xcode.app/Contents/Frameworks/IDEKit.framework/Resources/IDETextKey ...

  5. URAL2104. Game with a Strip(博弈)

    There is a strip 1 × n with two sides. Each square of the strip (their total amount is 2n, n squares ...

  6. h5 页面设计尺寸

    请注意:(以下所有讨论内容和规范均将viewport设定为content=”width=device-width”的情况下) 也就是我们的H5页面前端代码里面必须包含 <meta content ...

  7. bzoj4390

    树上差分 感觉挺巧妙的... 每次更新就是在u,v上+1,x是lca(u,v),在x和fa[x]上-1,那么每个点的权值就是子树和,正确性yy一下就行了 不过树状数组的常数真是小,改成前缀和才快了20 ...

  8. 洛谷 P3952 时间复杂度【模拟】

    把No写成NO,WA了一发-- 现在看这题也不难-- 用一个栈,记一下前面F的字母,是否合法,合法的有多长,每次入栈弹栈即可 #include<iostream> #include< ...

  9. CCF2016.4 - A题

    思路:枚举每个点,看看它是否同时小于/大于前一个点和后一个点 import java.util.Scanner; public class Main { public static void main ...

  10. 【插件开发】—— 8 IPreferenceStore,插件的键/值存储!

    前文回顾: 1 插件学习篇 2 简单的建立插件工程以及模型文件分析 3 利用扩展点,开发透视图 4 SWT编程须知 5 SWT简单控件的使用与布局搭配 6 SWT复杂空间与布局搭配 7 SWT布局详解 ...