Complete below list of tasks:
DBA may need to add a new node or existing node (which got deleted from cluster) into Clusterware.
List of tasks for New Node:
- Install same version of OS
- Setup environment variable
- Install all required package including Oracleasm
- Configure Networking (IP and others)
- Check Time Synchronization
- Change Recommended Kernel Parameters for Oracle
- Create Group and User (make sure these are same with existing Nodes)
- Setup SSH configuration for grid user or (user which was used for GRID installation) between new node and other nodes.
If it is existing node, make sure cleanup properly. You may go through: how to delete a node from cluster.
Compare setting between new nodes with existing any node by running below command and fix these if any discrimination is there.
[grid@ocmnode2 bin]$ pwd /u01/app/12.1.0/grid/bin [grid@ocmnode2 bin]$ ./cluvfy comp peer -n ocmnode1 -refnode ocmnode2 -r 12.1
Click on below line to view details:
./cluvfy comp peer -n ocmnode1 -refnode ocmnode2 -r 12.1
[grid@ocmnode2 bin]$ ./cluvfy comp peer -n ocmnode1 -refnode ocmnode2 -r 12.1
Verifying peer compatibility
Checking peer compatibility...
Compatibility check: Physical memory [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 3.7411GB (3922860.0KB) 3.8613GB (4048820.0KB) mismatched
Physical memory <null>
Compatibility check: Available memory [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 3.4843GB (3653536.0KB) 2.316GB (2428544.0KB) matched
Available memory <null>
Compatibility check: Swap space [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 3.9375GB (4128760.0KB) 3.9375GB (4128764.0KB) mismatched
Swap space <null>
Compatibility check: Free disk space for "/usr" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 23.6152GB (2.4762368E7KB) 19.5928GB (2.0544512E7KB) matched
Free disk space <null>
Compatibility check: Free disk space for "/var" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 23.6152GB (2.4762368E7KB) 19.5928GB (2.0544512E7KB) matched
Free disk space <null>
Compatibility check: Free disk space for "/etc" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 23.6152GB (2.4762368E7KB) 19.5928GB (2.0544512E7KB) matched
Free disk space <null>
Compatibility check: Free disk space for "/u01/app/12.1.0/grid" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 23.6152GB (2.4762368E7KB) 19.5928GB (2.0544512E7KB) matched
Free disk space <null>
Compatibility check: Free disk space for "/sbin" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 23.6152GB (2.4762368E7KB) 19.5928GB (2.0544512E7KB) matched
Free disk space <null>
Compatibility check: Free disk space for "/tmp" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 23.6152GB (2.4762368E7KB) 19.5928GB (2.0544512E7KB) matched
Free disk space <null>
Compatibility check: User existence for "grid" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 grid(54321) grid(54321) matched
User existence for "grid" check passed
Compatibility check: Group existence for "oinstall" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 oinstall(54321) oinstall(54321) matched
Group existence for "oinstall" check passed
Compatibility check: Group existence for "dba" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 dba(54322) dba(54322) matched
Group existence for "dba" check passed
Compatibility check: Group membership for "grid" in "oinstall (Primary)" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 yes yes matched
Group membership for "grid" in "oinstall (Primary)" check passed
Compatibility check: Group membership for "grid" in "dba" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 no no matched
Group membership for "grid" in "dba" check passed
Compatibility check: Run level [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 5 5 matched
Run level check passed
Compatibility check: System architecture [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 x86_64 x86_64 matched
System architecture check passed
Compatibility check: Kernel version [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 2.6.32-431.el6.x86_64 3.8.13-16.2.1.el6uek.x86_64 mismatched
Kernel version check failed
Compatibility check: Kernel param "semmsl" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 250 250 matched
Kernel param "semmsl" check passed
Compatibility check: Kernel param "semmns" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 32000 32000 matched
Kernel param "semmns" check passed
Compatibility check: Kernel param "semopm" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 100 100 matched
Kernel param "semopm" check passed
Compatibility check: Kernel param "semmni" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 128 128 matched
Kernel param "semmni" check passed
Compatibility check: Kernel param "shmmax" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 68719476736 68719476736 matched
Kernel param "shmmax" check passed
Compatibility check: Kernel param "shmmni" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 4096 4096 matched
Kernel param "shmmni" check passed
Compatibility check: Kernel param "shmall" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 0 0 matched
Kernel param "shmall" check passed
Compatibility check: Kernel param "file-max" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 6815744 6815744 matched
Kernel param "file-max" check passed
Compatibility check: Kernel param "ip_local_port_range" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 9000 65500 9000 65500 matched
Kernel param "ip_local_port_range" check passed
Compatibility check: Kernel param "rmem_default" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 262144 262144 matched
Kernel param "rmem_default" check passed
Compatibility check: Kernel param "rmem_max" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 4194304 4194304 matched
Kernel param "rmem_max" check passed
Compatibility check: Kernel param "wmem_default" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 262144 262144 matched
Kernel param "wmem_default" check passed
Compatibility check: Kernel param "wmem_max" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 1048576 1048576 matched
Kernel param "wmem_max" check passed
Compatibility check: Kernel param "aio-max-nr" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 1048576 1048576 matched
Kernel param "aio-max-nr" check passed
Compatibility check: Kernel param "panic_on_oops" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 1 1 matched
Kernel param "panic_on_oops" check passed
Compatibility check: Package existence for "binutils" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 binutils-2.20.51.0.2-5.36.el6 binutils-2.20.51.0.2-5.36.el6 matched
Package existence for "binutils" check passed
Compatibility check: Package existence for "compat-libcap1" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 compat-libcap1-1.10-1 compat-libcap1-1.10-1 matched
Package existence for "compat-libcap1" check passed
Compatibility check: Package existence for "compat-libstdc++-33 (x86_64)" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 compat-libstdc++-33-3.2.3-69.el6 (x86_64) compat-libstdc++-33-3.2.3-69.el6 (x86_64) matched
Package existence for "compat-libstdc++-33 (x86_64)" check passed
Compatibility check: Package existence for "libgcc (x86_64)" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 libgcc-4.4.7-23.0.1.el6 (x86_64) libgcc-4.4.7-23.0.1.el6 (x86_64) matched
Package existence for "libgcc (x86_64)" check passed
Compatibility check: Package existence for "libstdc++ (x86_64)" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 libstdc++-4.4.7-23.0.1.el6 (x86_64) libstdc++-4.4.7-23.0.1.el6 (x86_64) matched
Package existence for "libstdc++ (x86_64)" check passed
Compatibility check: Package existence for "libstdc++-devel (x86_64)" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 libstdc++-devel-4.4.7-23.0.1.el6 (x86_64) libstdc++-devel-4.4.7-23.0.1.el6 (x86_64) matched
Package existence for "libstdc++-devel (x86_64)" check passed
Compatibility check: Package existence for "sysstat" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 sysstat-9.0.4-22.el6 sysstat-9.0.4-22.el6 matched
Package existence for "sysstat" check passed
Compatibility check: Package existence for "gcc" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 gcc-4.4.7-23.0.1.el6 gcc-4.4.7-23.0.1.el6 matched
Package existence for "gcc" check passed
Compatibility check: Package existence for "gcc-c++" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 gcc-c++-4.4.7-23.0.1.el6 gcc-c++-4.4.7-23.0.1.el6 matched
Package existence for "gcc-c++" check passed
Compatibility check: Package existence for "ksh" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 missing missing matched
Package existence for "ksh" check passed
Compatibility check: Package existence for "make" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 make-3.81-20.el6 make-3.81-20.el6 matched
Package existence for "make" check passed
Compatibility check: Package existence for "glibc (x86_64)" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 glibc-2.12-1.132.el6 (x86_64) glibc-2.12-1.132.el6 (x86_64) matched
Package existence for "glibc (x86_64)" check passed
Compatibility check: Package existence for "glibc-devel (x86_64)" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 glibc-devel-2.12-1.132.el6 (x86_64) glibc-devel-2.12-1.132.el6 (x86_64) matched
Package existence for "glibc-devel (x86_64)" check passed
Compatibility check: Package existence for "libaio (x86_64)" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 libaio-0.3.107-10.el6 (x86_64) libaio-0.3.107-10.el6 (x86_64) matched
Package existence for "libaio (x86_64)" check passed
Compatibility check: Package existence for "libaio-devel (x86_64)" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 libaio-devel-0.3.107-10.el6 (x86_64) libaio-devel-0.3.107-10.el6 (x86_64) matched
Package existence for "libaio-devel (x86_64)" check passed
Compatibility check: Package existence for "nfs-utils" [reference node: ocmnode2]
Node Name Status Ref. node status Comment
------------ ------------------------ ------------------------ ----------
ocmnode1 nfs-utils-1.2.3-39.el6 nfs-utils-1.2.3-39.el6 matched
Package existence for "nfs-utils" check passed
Verification of peer compatibility was successful.
Checks passed for the following node(s):
ocmnode1
Run below command to find the prerequisites status on New node or existing problematic node.
[grid@ocmnode2 bin]$ pwd /u01/app/12.1.0/grid/bin [grid@ocmnode2 bin]$ ./cluvfy stage -pre crsinst -n ocmnode1
Click on below line to view details:
[grid@ocmnode2 bin]$ ./cluvfy stage -pre crsinst -n ocmnode1
[grid@ocmnode2 bin]$ ./cluvfy stage -pre crsinst -n ocmnode1
Performing pre-checks for cluster services setup
Checking node reachability...
Node reachability check passed from node "ocmnode2"
Checking user equivalence...
User equivalence check passed for user "grid"
Checking node connectivity...
Checking hosts config file...
Verification of the hosts config file successful
Check: Node connectivity using interfaces on subnet "192.168.56.0"
Node connectivity passed for subnet "192.168.56.0" with node(s) ocmnode1
TCP connectivity check passed for subnet "192.168.56.0"
Check: Node connectivity using interfaces on subnet "192.168.10.0"
Node connectivity passed for subnet "192.168.10.0" with node(s) ocmnode1
TCP connectivity check passed for subnet "192.168.10.0"
Node connectivity check passed
Checking multicast communication...
Checking subnet "192.168.10.0" for multicast communication with multicast group "224.0.0.251"...
Check of subnet "192.168.10.0" for multicast communication with multicast group "224.0.0.251" passed.
Check of multicast communication passed.
Checking ASMLib configuration.
Check for ASMLib configuration passed.
Total memory check failed
Check failed on nodes:
ocmnode1
Available memory check passed
Swap space check passed
Free disk space check passed for "ocmnode1:/usr,ocmnode1:/var,ocmnode1:/etc,ocmnode1:/u01/app/12.1.0/grid,ocmnode1:/sbin,ocmnode1:/tmp"
Check for multiple users with UID value 54321 passed
User existence check passed for "grid"
Group existence check passed for "oinstall"
Group existence check passed for "dba"
Membership check for user "grid" in group "oinstall" [as Primary] passed
Membership check for user "grid" in group "dba" failed
Check failed on nodes:
ocmnode1
Run level check passed
Hard limits check passed for "maximum open file descriptors"
Soft limits check passed for "maximum open file descriptors"
Hard limits check passed for "maximum user processes"
Soft limits check passed for "maximum user processes"
System architecture check passed
Kernel version check failed
Check failed on nodes:
ocmnode1
Kernel parameter check passed for "semmsl"
Kernel parameter check passed for "semmns"
Kernel parameter check passed for "semopm"
Kernel parameter check passed for "semmni"
Kernel parameter check passed for "shmmax"
Kernel parameter check passed for "shmmni"
Kernel parameter check passed for "shmall"
Kernel parameter check passed for "file-max"
Kernel parameter check passed for "ip_local_port_range"
Kernel parameter check passed for "rmem_default"
Kernel parameter check passed for "rmem_max"
Kernel parameter check passed for "wmem_default"
Kernel parameter check passed for "wmem_max"
Kernel parameter check passed for "aio-max-nr"
Kernel parameter check passed for "panic_on_oops"
Package existence check passed for "binutils"
Package existence check passed for "compat-libcap1"
Package existence check passed for "compat-libstdc++-33(x86_64)"
Package existence check passed for "libgcc(x86_64)"
Package existence check passed for "libstdc++(x86_64)"
Package existence check passed for "libstdc++-devel(x86_64)"
Package existence check passed for "sysstat"
Package existence check passed for "gcc"
Package existence check passed for "gcc-c++"
Package existence check failed for "ksh"
Check failed on nodes:
ocmnode1
Package existence check passed for "make"
Package existence check passed for "glibc(x86_64)"
Package existence check passed for "glibc-devel(x86_64)"
Package existence check passed for "libaio(x86_64)"
Package existence check passed for "libaio-devel(x86_64)"
Package existence check passed for "nfs-utils"
Check for multiple users with UID value 0 passed
Current group ID check passed
Starting check for consistency of primary group of root user
Check for consistency of root user's primary group passed
Starting Clock synchronization checks using Network Time Protocol(NTP)...
NTP configuration file "/etc/ntp.conf" existence check passed
No NTP Daemons or Services were found to be running
PRVF-5507 : NTP daemon or service is not running on any node but NTP configuration file exists on the following node(s):
ocmnode1
Clock synchronization check using Network Time Protocol(NTP) failed
Core file name pattern consistency check passed.
User "grid" is not part of "root" group. Check passed
Default user file creation mask check passed
Checking integrity of file "/etc/resolv.conf" across nodes
WARNING:
PRVF-5640 : Both 'search' and 'domain' entries are present in file "/etc/resolv.conf" on the following nodes: ocmnode1
Check for integrity of file "/etc/resolv.conf" passed
Time zone consistency check passed
Checking integrity of name service switch configuration file "/etc/nsswitch.conf" ...
Check for integrity of name service switch configuration file "/etc/nsswitch.conf" passed
Checking daemon "avahi-daemon" is not configured and running
Daemon not configured check passed for process "avahi-daemon"
Daemon not running check passed for process "avahi-daemon"
Starting check for Reverse path filter setting ...
Check for Reverse path filter setting passed
Starting check for Network interface bonding status of private interconnect network interfaces ...
Check for Network interface bonding status of private interconnect network interfaces passed
Starting check for /dev/shm mounted as temporary file system ...
Check for /dev/shm mounted as temporary file system passed
Starting check for /boot mount ...
Check for /boot mount passed
Starting check for zeroconf check ...
Check for zeroconf check passed
Pre-check for cluster services setup was unsuccessful on all the nodes.
Below command will generate fix-up script for prerequisites.
[grid@ocmnode2 bin]$ pwd /u01/app/12.1.0/grid/bin [grid@ocmnode2 bin]$ ./cluvfy stage -pre crsinst -n ocmnode1 -fixup
[grid@ocmnode2 bin]$ ./cluvfy stage -pre crsinst -n ocmnode1 -fixup
[grid@ocmnode2 .ssh]$ cd /u01/app/12.1.0/grid/bin
[grid@ocmnode2 bin]$ ./cluvfy stage -pre crsinst -n ocmnode1 -fixup
Performing pre-checks for cluster services setup
Checking node reachability...
Node reachability check passed from node "ocmnode2"
Checking user equivalence...
User equivalence check passed for user "grid"
Checking node connectivity...
Checking hosts config file...
Verification of the hosts config file successful
Check: Node connectivity using interfaces on subnet "192.168.10.0"
Node connectivity passed for subnet "192.168.10.0" with node(s) ocmnode1
TCP connectivity check passed for subnet "192.168.10.0"
Check: Node connectivity using interfaces on subnet "192.168.56.0"
Node connectivity passed for subnet "192.168.56.0" with node(s) ocmnode1
TCP connectivity check passed for subnet "192.168.56.0"
Node connectivity check passed
Checking multicast communication...
Checking subnet "192.168.10.0" for multicast communication with multicast group "224.0. 0.251"...
Check of subnet "192.168.10.0" for multicast communication with multicast group "224.0. 0.251" passed.
Check of multicast communication passed.
Checking ASMLib configuration.
Check for ASMLib configuration passed.
Total memory check failed
Check failed on nodes:
ocmnode1
Available memory check passed
Swap space check passed
Free disk space check passed for "ocmnode1:/usr,ocmnode1:/var,ocmnode1:/etc,ocmnode1:/u01/app/12.1.0/grid,ocmnode1:/sbin,ocmnode1:/tmp"
Check for multiple users with UID value 54321 passed
User existence check passed for "grid"
Group existence check passed for "oinstall"
Group existence check passed for "dba"
Membership check for user "grid" in group "oinstall" [as Primary] passed
Membership check for user "grid" in group "dba" failed
Check failed on nodes:
ocmnode1
Run level check passed
Hard limits check passed for "maximum open file descriptors"
Soft limits check passed for "maximum open file descriptors"
Hard limits check passed for "maximum user processes"
Soft limits check passed for "maximum user processes"
System architecture check passed
Kernel version check passed
Kernel parameter check passed for "semmsl"
Kernel parameter check passed for "semmns"
Kernel parameter check passed for "semopm"
Kernel parameter check passed for "semmni"
Kernel parameter check passed for "shmmax"
Kernel parameter check passed for "shmmni"
Kernel parameter check passed for "shmall"
Kernel parameter check passed for "file-max"
Kernel parameter check passed for "ip_local_port_range"
Kernel parameter check passed for "rmem_default"
Kernel parameter check passed for "rmem_max"
Kernel parameter check passed for "wmem_default"
Kernel parameter check passed for "wmem_max"
PRVG-1206 : Check cannot be performed for configured value of kernel parameter "aio-max-nr" on node "ocmnode1"
Kernel parameter check passed for "aio-max-nr"
PRVG-1206 : Check cannot be performed for configured value of kernel parameter "panic_on_oops" on node "ocmnode1"
Kernel parameter check passed for "panic_on_oops"
Package existence check passed for "binutils"
Package existence check failed for "compat-libcap1"
Check failed on nodes:
ocmnode1
Package existence check passed for "compat-libstdc++-33(x86_64)"
Package existence check passed for "libgcc(x86_64)"
Package existence check passed for "libstdc++(x86_64)"
Package existence check passed for "libstdc++-devel(x86_64)"
Package existence check passed for "sysstat"
Package existence check passed for "gcc"
Package existence check passed for "gcc-c++"
Package existence check failed for "ksh"
Check failed on nodes:
ocmnode1
Package existence check passed for "make"
Package existence check passed for "glibc(x86_64)"
Package existence check passed for "glibc-devel(x86_64)"
Package existence check passed for "libaio(x86_64)"
Package existence check failed for "libaio-devel(x86_64)"
Check failed on nodes:
ocmnode1
Package existence check passed for "nfs-utils"
Check for multiple users with UID value 0 passed
Current group ID check passed
Starting check for consistency of primary group of root user
Check for consistency of root user's primary group passed
Starting Clock synchronization checks using Network Time Protocol(NTP)...
NTP configuration file "/etc/ntp.conf" existence check passed
No NTP Daemons or Services were found to be running
PRVF-5507 : NTP daemon or service is not running on any node but NTP configuration file exists on the following node(s):
ocmnode1
Clock synchronization check using Network Time Protocol(NTP) failed
Core file name pattern consistency check passed.
User "grid" is not part of "root" group. Check passed
Default user file creation mask check passed
Checking integrity of file "/etc/resolv.conf" across nodes
WARNING:
PRVF-5640 : Both 'search' and 'domain' entries are present in file "/etc/resolv.conf" on the following nodes: ocmnode1
Check for integrity of file "/etc/resolv.conf" passed
Time zone consistency check passed
Checking integrity of name service switch configuration file "/etc/nsswitch.conf" ...
Check for integrity of name service switch configuration file "/etc/nsswitch.conf" passed
Checking daemon "avahi-daemon" is not configured and running
Daemon not configured check passed for process "avahi-daemon"
Daemon not running check passed for process "avahi-daemon"
Starting check for Reverse path filter setting ...
Check for Reverse path filter setting passed
Starting check for Network interface bonding status of private interconnect network interfaces ...
Check for Network interface bonding status of private interconnect network interfaces passed
Starting check for /dev/shm mounted as temporary file system ...
ERROR:
PRVE-0426 : The size of in-memory file system mounted as /dev/shm is "1002" megabytes which is less than the required size of "2048" megabytes on node ""
Check for /dev/shm mounted as temporary file system failed
Starting check for /boot mount ...
Check for /boot mount passed
Starting check for zeroconf check ...
ERROR:
PRVE-10077 : NOZEROCONF parameter was not specified or was not set to 'yes' in file "/etc/sysconfig/network" on node "ocmnode1.localdomain"
Check for zeroconf check failed
Pre-check for cluster services setup was unsuccessful on all the nodes.
******************************************************************************************
Following is the list of fixable prerequisites selected to fix in this session
******************************************************************************************
-------------- --------------- ----------------
Check failed. Failed on nodes Reboot required?
-------------- --------------- ----------------
zeroconf check ocmnode1 no
Group Membership: dba ocmnode1 no
OS Kernel Parameter: ocmnode1 no
aio-max-nr
OS Kernel Parameter: ocmnode1 no
panic_on_oops
Execute "/tmp/CVU_12.1.0.2.0_grid/runfixup.sh" as root user on nodes "ocmnode1" to perform the fix up operations manually
Press ENTER key to continue after execution of "/tmp/CVU_12.1.0.2.0_grid/runfixup.sh" has completed on nodes "ocmnode1"
Fix: zeroconf check
"zeroconf check" was successfully fixed on all the applicable nodes
Fix: Group Membership: dba
"Group Membership: dba" was successfully fixed on all the applicable nodes
Fix: OS Kernel Parameter: aio-max-nr
"OS Kernel Parameter: aio-max-nr" was successfully fixed on all the applicable nodes
Fix: OS Kernel Parameter: panic_on_oops
"OS Kernel Parameter: panic_on_oops" was successfully fixed on all the applicable nodes
Pre-check for node addition:
[grid@ocmnode2 bin]$ pwd /u01/app/12.1.0/grid/bin [grid@ocmnode2 bin]$ ./cluvfy stage -pre nodeadd -n ocmnode1
[grid@ocmnode2 bin]$ ./cluvfy stage -pre nodeadd -n ocmnode1
[grid@ocmnode2 bin]$ ./cluvfy stage -pre nodeadd -n ocmnode1
Performing pre-checks for node addition
Checking node reachability...
Node reachability check passed from node "ocmnode2"
Checking user equivalence...
User equivalence check passed for user "grid"
Checking CRS integrity...
CRS integrity check passed
Clusterware version consistency passed.
Checking shared resources...
Checking CRS home location...
Path "/u01/app/12.1.0/grid" either already exists or can be successfully created on nodes: "ocmnode1"
Shared resources check for node addition passed
Checking node connectivity...
Checking hosts config file...
Verification of the hosts config file successful
Check: Node connectivity using interfaces on subnet "192.168.10.0"
Node connectivity passed for subnet "192.168.10.0" with node(s) ocmnode2,ocmnode1
TCP connectivity check passed for subnet "192.168.10.0"
Check: Node connectivity using interfaces on subnet "192.168.56.0"
Node connectivity passed for subnet "192.168.56.0" with node(s) ocmnode1,ocmnode2
TCP connectivity check passed for subnet "192.168.56.0"
Checking subnet mask consistency...
Subnet mask consistency check passed for subnet "192.168.56.0".
Subnet mask consistency check passed for subnet "192.168.10.0".
Subnet mask consistency check passed.
Node connectivity check passed
Checking multicast communication...
Checking subnet "192.168.10.0" for multicast communication with multicast group "224.0.0.251"...
Check of subnet "192.168.10.0" for multicast communication with multicast group "224.0.0.251" passed.
Check of multicast communication passed.
Total memory check failed
Check failed on nodes:
ocmnode2,ocmnode1
Available memory check passed
Swap space check passed
Free disk space check passed for "ocmnode2:/usr,ocmnode2:/var,ocmnode2:/etc,ocmnode2:/u01/app/12.1.0/grid,ocmnode2:/sbin,ocmnode2:/tmp"
Free disk space check passed for "ocmnode1:/usr,ocmnode1:/var,ocmnode1:/etc,ocmnode1:/u01/app/12.1.0/grid,ocmnode1:/sbin,ocmnode1:/tmp"
Check for multiple users with UID value 54321 passed
User existence check passed for "grid"
Run level check passed
Hard limits check passed for "maximum open file descriptors"
Soft limits check passed for "maximum open file descriptors"
Hard limits check passed for "maximum user processes"
Soft limits check passed for "maximum user processes"
System architecture check passed
Kernel version check passed
Kernel parameter check passed for "semmsl"
Kernel parameter check passed for "semmns"
Kernel parameter check passed for "semopm"
Kernel parameter check passed for "semmni"
Kernel parameter check passed for "shmmax"
Kernel parameter check passed for "shmmni"
Kernel parameter check passed for "shmall"
Kernel parameter check passed for "file-max"
Kernel parameter check passed for "ip_local_port_range"
Kernel parameter check passed for "rmem_default"
Kernel parameter check passed for "rmem_max"
Kernel parameter check passed for "wmem_default"
Kernel parameter check passed for "wmem_max"
Kernel parameter check passed for "aio-max-nr"
Kernel parameter check passed for "panic_on_oops"
Package existence check passed for "binutils"
Package existence check failed for "compat-libcap1"
Check failed on nodes:
ocmnode1
Package existence check passed for "compat-libstdc++-33(x86_64)"
Package existence check passed for "libgcc(x86_64)"
Package existence check passed for "libstdc++(x86_64)"
Package existence check passed for "libstdc++-devel(x86_64)"
Package existence check passed for "sysstat"
Package existence check passed for "gcc"
Package existence check passed for "gcc-c++"
Package existence check failed for "ksh"
Check failed on nodes:
ocmnode2,ocmnode1
Package existence check passed for "make"
Package existence check passed for "glibc(x86_64)"
Package existence check passed for "glibc-devel(x86_64)"
Package existence check passed for "libaio(x86_64)"
Package existence check failed for "libaio-devel(x86_64)"
Check failed on nodes:
ocmnode1
Package existence check passed for "nfs-utils"
Check for multiple users with UID value 0 passed
Current group ID check passed
Starting check for consistency of primary group of root user
Check for consistency of root user's primary group passed
Group existence check passed for "asmadmin"
Membership check for user "grid" in group "asmadmin" passed
Group existence check passed for "asmoper"
Membership check for user "grid" in group "asmoper" passed
Group existence check passed for "asmdba"
Membership check for user "grid" in group "asmdba" passed
Group existence check passed for "oinstall"
Membership check for user "grid" in group "oinstall" passed
Check for multiple users with UID value 0 passed
User existence check passed for "root"
Check for multiple users with UID value 54321 passed
User existence check passed for "grid"
Checking ASMLib configuration.
Check for ASMLib configuration passed.
Checking OCR integrity...
Disks "+OCR/ocmnode-cluster/OCRFILE/registry.255.1047853917" are managed by ASM.
OCR integrity check passed
Checking Oracle Cluster Voting Disk configuration...
Disks "ORCL:OCRDISK1,ORCL:OCRDISK2,ORCL:OCRDISK3" are managed by ASM.
Oracle Cluster Voting Disk configuration check passed
Time zone consistency check passed
Starting Clock synchronization checks using Network Time Protocol(NTP)...
NTP configuration file "/etc/ntp.conf" existence check passed
No NTP Daemons or Services were found to be running
PRVF-5507 : NTP daemon or service is not running on any node but NTP configuration file exists on the following node(s):
ocmnode2,ocmnode1
Clock synchronization check using Network Time Protocol(NTP) failed
User "grid" is not part of "root" group. Check passed
Checking integrity of file "/etc/resolv.conf" across nodes
WARNING:
PRVF-5640 : Both 'search' and 'domain' entries are present in file "/etc/resolv.conf" on the following nodes: ocmnode2,ocmnode1
Check for integrity of file "/etc/resolv.conf" passed
Checking integrity of name service switch configuration file "/etc/nsswitch.conf" ...
Check for integrity of name service switch configuration file "/etc/nsswitch.conf" passed
Pre-check for node addition was unsuccessful on all the nodes.
Run below command from workable node.
[grid@ocmnode2 addnode]$ ./addnode.sh -silent CLUSTER_NEW_NODES={ocmnode1} CLUSTER_NEW_VIRTUAL_HOSTNAMES={ocmnode1-vip} -ignoreSysPrereqs -ignorePrereq
Starting Oracle Universal Installer...
Checking Temp space: must be greater than 120 MB. Actual 15883 MB Passed
Checking swap space: must be greater than 150 MB. Actual 4029 MB Passed
Prepare Configuration in progress.
Prepare Configuration successful.
.................................................. 8% Done.
You can find the log of this install session at:
/u01/app/oraInventory/logs/addNodeActions2020-08-08_10-48-12PM.log
Instantiate files in progress.
Instantiate files successful.
.................................................. 14% Done.
Copying files to node in progress.
Copying files to node successful.
.................................................. 73% Done.
Saving cluster inventory in progress.
.................................................. 80% Done.
Saving cluster inventory successful.
The Cluster Node Addition of /u01/app/12.1.0/grid was successful.
Please check '/tmp/silentInstall.log' for more details.
Setup Oracle Base in progress.
Setup Oracle Base successful.
.................................................. 88% Done.
As a root user, execute the following script(s):
1. /u01/app/oraInventory/orainstRoot.sh
2. /u01/app/12.1.0/grid/root.sh
Execute /u01/app/oraInventory/orainstRoot.sh on the following nodes:
[ocmnode1]
Execute /u01/app/12.1.0/grid/root.sh on the following nodes:
[ocmnode1]
The scripts can be executed in parallel on all the nodes.
..........
Update Inventory in progress.
.................................................. 100% Done.
Update Inventory successful.
Successfully Setup Software.
[root@ocmnode1 ~]# /u01/app/oraInventory/orainstRoot.sh Changing permissions of /u01/app/oraInventory. Adding read,write permissions for group. Removing read,write,execute permissions for world. Changing groupname of /u01/app/oraInventory to oinstall. The execution of the script is complete.
[root@ocmnode1 ~]# /u01/app/12.1.0/grid/root.sh Check /u01/app/12.1.0/grid/install/root_ocmnode1.localdomain_2020-08-10_01-35-07.log for t he output of root script
[grid@ocmnode2 ~]$ cat root_ocmnode1.localdomain_2020-08-10_01-35-07.log
Performing root user operation.
The following environment variables are set as:
ORACLE_OWNER= grid
ORACLE_HOME= /u01/app/12.1.0/grid
Copying dbhome to /usr/local/bin ...
Copying oraenv to /usr/local/bin ...
Copying coraenv to /usr/local/bin ...
Entries will be added to the /etc/oratab file as needed by
Database Configuration Assistant when a database is created
Finished running generic part of root script.
Now product-specific root actions will be performed.
Relinking oracle with rac_on option
Using configuration parameter file: /u01/app/12.1.0/grid/crs/install/crsconfig_params
2020/08/10 01:35:08 CLSRSC-4001: Installing Oracle Trace File Analyzer (TFA) Collector.
2020/08/10 01:35:08 CLSRSC-4002: Successfully installed Oracle Trace File Analyzer (TFA) Collector.
2020/08/10 01:35:09 CLSRSC-363: User ignored prerequisites during installation
CRS-2791: Starting shutdown of Oracle High Availability Services-managed resources on 'ocmnode1'
CRS-2673: Attempting to stop 'ora.crsd' on 'ocmnode1'
CRS-2790: Starting shutdown of Cluster Ready Services-managed resources on 'ocmnode1'
CRS-2673: Attempting to stop 'ora.OCR.dg' on 'ocmnode1'
CRS-2673: Attempting to stop 'ora.LISTENER_SCAN1.lsnr' on 'ocmnode1'
CRS-2673: Attempting to stop 'ora.FRA.dg' on 'ocmnode1'
CRS-2673: Attempting to stop 'ora.OCR_VOTE.dg' on 'ocmnode1'
CRS-2677: Stop of 'ora.OCR_VOTE.dg' on 'ocmnode1' succeeded
CRS-2677: Stop of 'ora.LISTENER_SCAN1.lsnr' on 'ocmnode1' succeeded
CRS-2673: Attempting to stop 'ora.scan1.vip' on 'ocmnode1'
CRS-2677: Stop of 'ora.OCR.dg' on 'ocmnode1' succeeded
CRS-2677: Stop of 'ora.FRA.dg' on 'ocmnode1' succeeded
CRS-2673: Attempting to stop 'ora.DATA.dg' on 'ocmnode1'
CRS-2677: Stop of 'ora.DATA.dg' on 'ocmnode1' succeeded
CRS-2673: Attempting to stop 'ora.asm' on 'ocmnode1'
CRS-2677: Stop of 'ora.asm' on 'ocmnode1' succeeded
CRS-2677: Stop of 'ora.scan1.vip' on 'ocmnode1' succeeded
CRS-2672: Attempting to start 'ora.scan1.vip' on 'ocmnode2'
CRS-2676: Start of 'ora.scan1.vip' on 'ocmnode2' succeeded
CRS-2672: Attempting to start 'ora.LISTENER_SCAN1.lsnr' on 'ocmnode2'
CRS-2676: Start of 'ora.LISTENER_SCAN1.lsnr' on 'ocmnode2' succeeded
CRS-2673: Attempting to stop 'ora.ons' on 'ocmnode1'
CRS-2677: Stop of 'ora.ons' on 'ocmnode1' succeeded
CRS-2673: Attempting to stop 'ora.net1.network' on 'ocmnode1'
CRS-2677: Stop of 'ora.net1.network' on 'ocmnode1' succeeded
CRS-2792: Shutdown of Cluster Ready Services-managed resources on 'ocmnode1' has completed
CRS-2677: Stop of 'ora.crsd' on 'ocmnode1' succeeded
CRS-2673: Attempting to stop 'ora.evmd' on 'ocmnode1'
CRS-2673: Attempting to stop 'ora.mdnsd' on 'ocmnode1'
CRS-2673: Attempting to stop 'ora.gpnpd' on 'ocmnode1'
CRS-2673: Attempting to stop 'ora.drivers.acfs' on 'ocmnode1'
CRS-2677: Stop of 'ora.drivers.acfs' on 'ocmnode1' succeeded
CRS-2677: Stop of 'ora.mdnsd' on 'ocmnode1' succeeded
CRS-2677: Stop of 'ora.gpnpd' on 'ocmnode1' succeeded
CRS-2677: Stop of 'ora.evmd' on 'ocmnode1' succeeded
CRS-2673: Attempting to stop 'ora.ctssd' on 'ocmnode1'
CRS-2673: Attempting to stop 'ora.storage' on 'ocmnode1'
CRS-2677: Stop of 'ora.storage' on 'ocmnode1' succeeded
CRS-2673: Attempting to stop 'ora.asm' on 'ocmnode1'
CRS-2677: Stop of 'ora.ctssd' on 'ocmnode1' succeeded
CRS-2677: Stop of 'ora.asm' on 'ocmnode1' succeeded
CRS-2673: Attempting to stop 'ora.cluster_interconnect.haip' on 'ocmnode1'
CRS-2677: Stop of 'ora.cluster_interconnect.haip' on 'ocmnode1' succeeded
CRS-2673: Attempting to stop 'ora.cssd' on 'ocmnode1'
CRS-2677: Stop of 'ora.cssd' on 'ocmnode1' succeeded
CRS-2673: Attempting to stop 'ora.crf' on 'ocmnode1'
CRS-2677: Stop of 'ora.crf' on 'ocmnode1' succeeded
CRS-2673: Attempting to stop 'ora.gipcd' on 'ocmnode1'
CRS-2677: Stop of 'ora.gipcd' on 'ocmnode1' succeeded
CRS-2793: Shutdown of Oracle High Availability Services-managed resources on 'ocmnode1' has completed
CRS-4133: Oracle High Availability Services has been stopped.
CRS-4123: Starting Oracle High Availability Services-managed resources
CRS-2672: Attempting to start 'ora.mdnsd' on 'ocmnode1'
CRS-2672: Attempting to start 'ora.evmd' on 'ocmnode1'
CRS-2676: Start of 'ora.evmd' on 'ocmnode1' succeeded
CRS-2676: Start of 'ora.mdnsd' on 'ocmnode1' succeeded
CRS-2672: Attempting to start 'ora.gpnpd' on 'ocmnode1'
CRS-2676: Start of 'ora.gpnpd' on 'ocmnode1' succeeded
CRS-2672: Attempting to start 'ora.gipcd' on 'ocmnode1'
CRS-2676: Start of 'ora.gipcd' on 'ocmnode1' succeeded
CRS-2672: Attempting to start 'ora.cssdmonitor' on 'ocmnode1'
CRS-2676: Start of 'ora.cssdmonitor' on 'ocmnode1' succeeded
CRS-2672: Attempting to start 'ora.cssd' on 'ocmnode1'
CRS-2672: Attempting to start 'ora.diskmon' on 'ocmnode1'
CRS-2676: Start of 'ora.diskmon' on 'ocmnode1' succeeded
CRS-2676: Start of 'ora.cssd' on 'ocmnode1' succeeded
CRS-2672: Attempting to start 'ora.cluster_interconnect.haip' on 'ocmnode1'
CRS-2672: Attempting to start 'ora.ctssd' on 'ocmnode1'
CRS-2676: Start of 'ora.ctssd' on 'ocmnode1' succeeded
CRS-2676: Start of 'ora.cluster_interconnect.haip' on 'ocmnode1' succeeded
CRS-2672: Attempting to start 'ora.asm' on 'ocmnode1'
CRS-2676: Start of 'ora.asm' on 'ocmnode1' succeeded
CRS-2672: Attempting to start 'ora.storage' on 'ocmnode1'
CRS-2676: Start of 'ora.storage' on 'ocmnode1' succeeded
CRS-2672: Attempting to start 'ora.crf' on 'ocmnode1'
CRS-2676: Start of 'ora.crf' on 'ocmnode1' succeeded
CRS-2672: Attempting to start 'ora.crsd' on 'ocmnode1'
CRS-2676: Start of 'ora.crsd' on 'ocmnode1' succeeded
CRS-6017: Processing resource auto-start for servers: ocmnode1
CRS-2672: Attempting to start 'ora.net1.network' on 'ocmnode1'
CRS-2676: Start of 'ora.net1.network' on 'ocmnode1' succeeded
CRS-2672: Attempting to start 'ora.ons' on 'ocmnode1'
CRS-2676: Start of 'ora.ons' on 'ocmnode1' succeeded
CRS-2673: Attempting to stop 'ora.LISTENER_SCAN1.lsnr' on 'ocmnode2'
CRS-2677: Stop of 'ora.LISTENER_SCAN1.lsnr' on 'ocmnode2' succeeded
CRS-2673: Attempting to stop 'ora.scan1.vip' on 'ocmnode2'
CRS-2677: Stop of 'ora.scan1.vip' on 'ocmnode2' succeeded
CRS-2672: Attempting to start 'ora.scan1.vip' on 'ocmnode1'
CRS-2676: Start of 'ora.scan1.vip' on 'ocmnode1' succeeded
CRS-2672: Attempting to start 'ora.LISTENER_SCAN1.lsnr' on 'ocmnode1'
CRS-2676: Start of 'ora.LISTENER_SCAN1.lsnr' on 'ocmnode1' succeeded
CRS-6016: Resource auto-start has completed for server ocmnode1
CRS-6024: Completed start of Oracle Cluster Ready Services-managed resources
CRS-4123: Oracle High Availability Services has been started.
2020/08/10 01:37:21 CLSRSC-343: Successfully started Oracle Clusterware stack
clscfg: EXISTING configuration version 5 detected.
clscfg: version 5 is 12c Release 1.
Successfully accumulated necessary OCR keys.
Creating OCR keys for user 'root', privgrp 'root'..
Operation successful.
Preparing packages for installation...
cvuqdisk-1.0.9-1
2020/08/10 01:37:38 CLSRSC-325: Configure Oracle Grid Infrastructure for a Cluster ... succeeded
Node added successfully with Clusterware.
Validation:
[grid@ocmnode2 ~]$ crsctl stat res -t
--------------------------------------------------------------------------------
Name Target State Server State details
--------------------------------------------------------------------------------
Local Resources
--------------------------------------------------------------------------------
ora.DATA.dg
ONLINE ONLINE ocmnode1 STABLE
ONLINE ONLINE ocmnode2 STABLE
ora.FRA.dg
ONLINE ONLINE ocmnode1 STABLE
ONLINE ONLINE ocmnode2 STABLE
ora.LISTENER.lsnr
ONLINE ONLINE ocmnode1 STABLE
ONLINE ONLINE ocmnode2 STABLE
ora.OCR.dg
ONLINE ONLINE ocmnode1 STABLE
ONLINE ONLINE ocmnode2 STABLE
ora.OCR_VOTE.dg
ONLINE ONLINE ocmnode1 STABLE
ONLINE ONLINE ocmnode2 STABLE
ora.asm
ONLINE ONLINE ocmnode1 Started,STABLE
ONLINE ONLINE ocmnode2 Started,STABLE
ora.net1.network
ONLINE ONLINE ocmnode1 STABLE
ONLINE ONLINE ocmnode2 STABLE
ora.ons
ONLINE ONLINE ocmnode1 STABLE
ONLINE ONLINE ocmnode2 STABLE
--------------------------------------------------------------------------------
Cluster Resources
--------------------------------------------------------------------------------
ora.LISTENER_SCAN1.lsnr
1 ONLINE ONLINE ocmnode1 STABLE
ora.LISTENER_SCAN2.lsnr
1 ONLINE ONLINE ocmnode2 STABLE
ora.LISTENER_SCAN3.lsnr
1 ONLINE ONLINE ocmnode2 STABLE
ora.MGMTLSNR
1 ONLINE ONLINE ocmnode2 169.254.43.148 192.1
68.10.11,STABLE
ora.cvu
1 ONLINE ONLINE ocmnode2 STABLE
ora.mgmtdb
1 ONLINE ONLINE ocmnode2 Open,STABLE
ora.oc4j
1 ONLINE ONLINE ocmnode2 STABLE
ora.ocmnode1.vip
1 ONLINE ONLINE ocmnode1 STABLE
ora.ocmnode2.vip
1 ONLINE ONLINE ocmnode2 STABLE
ora.scan1.vip
1 ONLINE ONLINE ocmnode1 STABLE
ora.scan2.vip
1 ONLINE ONLINE ocmnode2 STABLE
ora.scan3.vip
1 ONLINE ONLINE ocmnode2 STABLE
--------------------------------------------------------------------------------
Adding Node to Oracle Home:
[oracle@ocmnode2 addnode]$ pwd
/u01/app/oracle/product/12.1.0/dbhome_1/addnode
[oracle@ocmnode2 addnode]$ ./addnode.sh -silent "CLUSTER_NEW_NODES={ocmnode1}" "CLUSTER_NEW_VIRTUAL_HOSTNAMES={ocmnode1-vip}"
[oracle@ocmnode2 addnode]$ ./addnode.sh -silent "CLUSTER_NEW_NODES={ocmnode1}" "CLUSTER_NEW_VIRTUAL_HOSTNAMES={ocmnode1-vip}"
Starting Oracle Universal Installer...
Checking Temp space: must be greater than 120 MB. Actual 12822 MB Passed
Checking swap space: must be greater than 150 MB. Actual 4031 MB Passed
[FATAL] [INS-30160] Installer has detected that the nodes [ocmnode1] specified for addnode operation have uncleaned inventory.
ACTION: Ensure that the inventory location /u01/app/oraInventory is cleaned before performing addnode procedure.
[oracle@ocmnode2 addnode]$ ./addnode.sh -silent "CLUSTER_NEW_NODES={ocmnode1}" "CLUSTER_NEW_VIRTUAL_HOSTNAMES={ocmnode1-vip}"
Starting Oracle Universal Installer...
Checking Temp space: must be greater than 120 MB. Actual 12822 MB Passed
Checking swap space: must be greater than 150 MB. Actual 4031 MB Passed
[WARNING] [INS-13014] Target environment does not meet some optional requirements.
CAUSE: Some of the optional prerequisites are not met. See logs for details. /u01/app/oraInventory/logs/addNodeActions2020-08-10_03-17-13PM.log
ACTION: Identify the list of failed prerequisite checks from the log: /u01/app/oraInventory/logs/addNodeActions2020-08-10_03-17-13PM.log. Then either from the log file or from installation manual find the appropriate configuration to meet the prerequisites and fix it manually.
Prepare Configuration in progress.
Prepare Configuration successful.
.................................................. 8% Done.
You can find the log of this install session at:
/u01/app/oraInventory/logs/addNodeActions2020-08-10_03-17-13PM.log
Instantiate files in progress.
Instantiate files successful.
.................................................. 14% Done.
Copying files to node in progress.
Copying files to node successful.
.................................................. 73% Done.
Saving cluster inventory in progress.
SEVERE:Remote 'AttachHome' failed on nodes: 'ocmnode1'. Refer to '/u01/app/oraInventory/logs/addNodeActions2020-08-10_03-17-13PM.log' for details.
It is recommended that the following command needs to be manually run on the failed nodes:
/u01/app/oracle/product/12.1.0/dbhome_1/oui/bin/runInstaller -attachHome -noClusterEnabled ORACLE_HOME=/u01/app/oracle/product/12.1.0/dbhome_1 ORACLE_HOME_NAME=OraDB12Home1 CLUSTER_NODES=ocmnode2,ocmnode1 "INVENTORY_LOCATION=/u01/app/oraInventory" -invPtrLoc "/etc/oraInst.loc" LOCAL_NODE=<node on which command is to be run>.
Please refer 'AttachHome' logs under central inventory of remote nodes where failure occurred for more details.
SEVERE:Remote 'UpdateNodeList' failed on nodes: 'ocmnode1'. Refer to '/u01/app/oraInventory/logs/addNodeActions2020-08-10_03-17-13PM.log' for details.
It is recommended that the following command needs to be manually run on the failed nodes:
/u01/app/oracle/product/12.1.0/dbhome_1/oui/bin/runInstaller -updateNodeList -noClusterEnabled ORACLE_HOME=/u01/app/oracle/product/12.1.0/dbhome_1 CLUSTER_NODES=ocmnode2,ocmnode1 CRS=false "INVENTORY_LOCATION=/u01/app/oraInventory" -invPtrLoc "/etc/oraInst.loc" LOCAL_NODE=<node on which command is to be run>.
Please refer 'UpdateNodeList' logs under central inventory of remote nodes where failure occurred for more details.
.................................................. 80% Done.
Saving cluster inventory successful.
WARNING:OUI-10234:Failed to copy the root script, /u01/app/oraInventory/orainstRoot.sh to the cluster nodes ocmnode1.[Error in copying file '/u01/app/oraInventory/orainstRoot.sh' present inside directory '/' on nodes 'ocmnode1'. [PRKC-1080 : Failed to transfer file "/u01/app/oraInventory/orainstRoot.sh" to any of the given nodes "ocmnode1 ".
Error on node ocmnode1:/bin/tar: ./u01/app/oraInventory/orainstRoot.sh: Cannot open: No such file or directory
/bin/tar: Exiting with failure status due to previous errors]]
Please copy them manually to these nodes and execute the script.
The Cluster Node Addition of /u01/app/oracle/product/12.1.0/dbhome_1 was unsuccessful.
Please check '/tmp/silentInstall.log' for more details.
Setup Oracle Base in progress.
Setup Oracle Base successful.
.................................................. 88% Done.
As a root user, execute the following script(s):
1. /u01/app/oraInventory/orainstRoot.sh
2. /u01/app/oracle/product/12.1.0/dbhome_1/root.sh
Execute /u01/app/oraInventory/orainstRoot.sh on the following nodes:
[ocmnode1]
Execute /u01/app/oracle/product/12.1.0/dbhome_1/root.sh on the following nodes:
[ocmnode1]
..........
Update Inventory in progress.
SEVERE:Remote 'UpdateNodeList' failed on nodes: 'ocmnode1'. Refer to '/u01/app/oraInventory/logs/UpdateNodeList2020-08-10_03-17-13PM.log' for details.
It is recommended that the following command needs to be manually run on the failed nodes:
/u01/app/oracle/product/12.1.0/dbhome_1/oui/bin/runInstaller -updateNodeList -setCustomNodelist -noClusterEnabled ORACLE_HOME=/u01/app/oracle/product/12.1.0/dbhome_1 CLUSTER_NODES=ocmnode2,ocmnode1 "NODES_TO_SET={ocmnode2,ocmnode1}" CRS=false "INVENTORY_LOCATION=/u01/app/oraInventory" -invPtrLoc "/etc/oraInst.loc" LOCAL_NODE=<node on which command is to be run>.
Please refer 'UpdateNodeList' logs under central inventory of remote nodes where failure occurred for more details.
[WARNING] [INS-10016] Installer failed to update the cluster related details, for this Oracle home, in the inventory on all/some of the nodes
ACTION: You may chose to retry the operation, without continuing further. Alternatively you can refer to information given below and manually execute the mentioned commands on the failed nodes now or later to update the inventory.
*MORE DETAILS*
Execute the following command on node(s) [ocmnode1]:
/u01/app/oracle/product/12.1.0/dbhome_1/oui/bin/runInstaller -jreLoc /u01/app/oracle/product/12.1.0/dbhome_1/jdk/jre -paramFile /u01/app/oracle/product/12.1.0/dbhome_1/oui/clusterparam.ini -silent -ignoreSysPrereqs -updateNodeList -bigCluster ORACLE_HOME=/u01/app/oracle/product/12.1.0/dbhome_1 CLUSTER_NODES=<Local Node> "NODES_TO_SET={ocmnode2,ocmnode1}" -invPtrLoc "/u01/app/oracle/product/12.1.0/dbhome_1/oraInst.loc" -local
.................................................. 100% Done.
Update Inventory successful.
Successfully Setup Software.
Execute root.sh script on new node.
[root@ocmnode1 app]# /u01/app/oracle/product/12.1.0/dbhome_1/root.sh
Check /u01/app/oracle/product/12.1.0/dbhome_1/install/root_ocmnode1.localdomain_2020-08-10_15-27-42.log for the output of root script
[root@ocmnode1 app]# cat /u01/app/oracle/product/12.1.0/dbhome_1/install/root_ocmnode1.localdomain_2020-08-10_15-27-42.log
Performing root user operation.
The following environment variables are set as:
ORACLE_OWNER= oracle
ORACLE_HOME= /u01/app/oracle/product/12.1.0/dbhome_1
Copying dbhome to /usr/local/bin ...
Copying oraenv to /usr/local/bin ...
Copying coraenv to /usr/local/bin ...
Entries will be added to the /etc/oratab file as needed by
Database Configuration Assistant when a database is created
Finished running generic part of root script.
Now product-specific root actions will be performed.
During adding a Node, I faced number of issues. I mentioned in below and it may help you.
Issue # 1: ORA-00845: MEMORY_TARGET not supported on this system
[grid@ocmnode2 addnode]$ ./addnode.sh -silent CLUSTER_NEW_NODES={ocmnode1} CLUSTER_NEW_VIRTUAL_HOSTNAMES={ocmnode1-vip} -ignoreSysPrereqs -ignorePrereq
Got below error which running root.sh script.
ORA-00845: MEMORY_TARGET not supported on this system . For details refer to "(:CLSN00107:)" in "/u01/app/grid/diag/crs/ocmnode1/crs/trace/ohasd_oraagent_grid.trc". CRS-2883: Resource 'ora.asm' failed during Clusterware stack start. CRS-4406: Oracle High Availability Services synchronous start failed. CRS-4000: Command Start failed, or completed with errors. 2020/08/08 22:55:25 CLSRSC-117: Failed to start Oracle Clusterware stack Died at /u01/app/12.1.0/grid/crs/install/crsinstall.pm line 914. The command '/u01/app/12.1.0/grid/perl/bin/perl -I/u01/app/12.1.0/grid/perl/lib -I/u01/app/12.1.0/grid/crs/install /u01/app/12.1.0/grid/crs/install/rootcrs.pl ' execution failed
Solution: Need to extend /tmp filesystem size.
[root@ocmnode1 ~]# df -h Filesystem Size Used Avail Use% Mounted on /dev/mapper/vg_racprodn1-lv_root 45G 20G 23G 47% / tmpfs 1001M 702M 30% /dev/shm /dev/sda1 477M 55M 397M 13% /boot /dev/sr0 56M 56M 0 100% /media/VBox_GAs_5.2.10
[root@ocmnode1 ~]# cat /etc/fstab # /dev/mapper/vg_racprodn1-lv_root / ext4 defaults 1 1 UUID=31e76520-7f50-4592-9ced-d4f03d6d95cc /boot ext4 defaults 1 2 /dev/mapper/vg_racprodn1-lv_swap swap swap defaults 0 0 tmpfs /dev/shm tmpfs defaults 0 0 devpts /dev/pts devpts gid=5,mode=620 0 0 sysfs /sys sysfs defaults 0 0 proc /proc proc defaults 0 0
Modify Below line on fstab:
tmpfs /dev/shm tmpfs size=2g 0 0
[root@ocmnode1 ~]# cat /etc/fstab # /dev/mapper/vg_racprodn1-lv_root / ext4 defaults 1 1 UUID=31e76520-7f50-4592-9ced-d4f03d6d95cc /boot ext4 defaults 1 2 /dev/mapper/vg_racprodn1-lv_swap swap swap defaults 0 0 tmpfs /dev/shm tmpfs size=2g 0 0 devpts /dev/pts devpts gid=5,mode=620 0 0 sysfs /sys sysfs defaults 0 0 proc /proc proc defaults 0 0
Reboot the system.
Run the root.sh script.
Issue # 2: One or more node names “ocmnode1.localdomain” contain one or more of the following invalid characters “.”
[grid@ocmnode2 addnode]$ . oraenv
ORACLE_SID = [grid] ? +ASM2
The Oracle base has been set to /u01/app/grid
[grid@ocmnode2 addnode]$ pwd
/u01/app/12.1.0/grid/addnode
[grid@ocmnode2 addnode]$ ./addnode.sh -silent CLUSTER_NEW_NODES={ocmnode1.localdomain} CLUSTER_NEW_VIRTUAL_HOSTNAMES={ocmnode1-vip.localdomain} -ignoreSysPrereqs -ignorePrereq
Starting Oracle Universal Installer...
Checking Temp space: must be greater than 120 MB. Actual 16183 MB Passed
Checking swap space: must be greater than 150 MB. Actual 4031 MB Passed
[FATAL] [INS-30131] Initial setup required for the execution of installer validations failed.
CAUSE: Failed to access the temporary location.
ACTION: Ensure that the current user has required permissions to access the temporary location.
*ADDITIONAL INFORMATION:*
Exception details
 - PRVG-11322 : One or more node names "ocmnode1.localdomain" contain one or more of the following invalid characters "."
Solution:
Don’t use the domain name for silent mode. Correct syntax will be:
[grid@ocmnode2 addnode]$ ./addnode.sh -silent CLUSTER_NEW_NODES={ocmnode1} CLUSTER_NEW_VIRTUAL_HOSTNAMES={ocmnode1-vip} -ignoreSysPrereqs -ignorePrereq
Issue # 3: [FATAL] [INS-40906] Duplicate host name ocmnode1 found in the node information table for Oracle Clusterware install
[grid@ocmnode2 addnode]$ pwd
/u01/app/12.1.0/grid/addnode
[grid@ocmnode2 addnode]$ ls -l addnode.sh
-rwxr-xr-x 1 grid oinstall 3575 Aug 7 22:18 addnode.sh
[grid@ocmnode2 addnode]$ ./addnode.sh -silent "CLUSTER_NEW_NODES={ocmnode1}" "CLUSTER_NEW_VIRTUAL_HOSTNAMES={ocmnode1-vip}"
Starting Oracle Universal Installer...
Checking Temp space: must be greater than 120 MB. Actual 15562 MB Passed
Checking swap space: must be greater than 150 MB. Actual 4019 MB Passed
[FATAL] [INS-40906] Duplicate host name ocmnode1 found in the node information table for Oracle Clusterware install.
CAUSE: Duplicate entries have been made in the node information table for Oracle Clusterware installation.
ACTION: Remove duplicate entries in the Oracle Clusterware node information table.
Cause: The problematic node information still exist on other node’s inventory file.
[grid@ocmnode2 addnode]$ cat /u01/app/oraInventory/ContentsXML/inventory.xml
...
<HOME NAME="OraGI12Home1" LOC="/u01/app/12.1.0/grid" TYPE="O" IDX="1" CRS="true">
<NODE_LIST>
<NODE NAME="ocmnode1"/>
<NODE NAME="ocmnode2"/>
</NODE_LIST>
</HOME>
<HOME NAME="OraDB12Home1" LOC="/u01/app/oracle/product/12.1.0/dbhome_1" TYPE="O" IDX="2">
<NODE_LIST>
<NODE NAME="ocmnode1"/>
<NODE NAME="ocmnode2"/>
</NODE_LIST>
</HOME>
...Need to update inventory with Good Nodes (currently working) list.
[grid@ocmnode2 bin]$ pwd
/u01/app/12.1.0/grid/oui/bin
[grid@ocmnode2 bin]$ export ORACLE_HOME=/u01/app/12.1.0/grid/
[grid@ocmnode2 bin]$ ./runInstaller -updateNodeList ORACLE_HOME=$ORACLE_HOME "CLUSTER_NODES={ocmnode2}" CRS=TRUE
Starting Oracle Universal Installer...
Checking swap space: must be greater than 500 MB. Actual 4019 MB Passed
The inventory pointer is located at /etc/oraInst.loc
'UpdateNodeList' was successful.
[grid@ocmnode2 addnode]$ cat /u01/app/oraInventory/ContentsXML/inventory.xml | grep NODE
<NODE_LIST>
<NODE NAME="ocmnode2"/>
</NODE_LIST>
<NODE_LIST>
<NODE NAME="ocmnode2"/>
</NODE_LIST>Issue # 4:
[grid@ocmnode2 addnode]$ ./addnode.sh -silent CLUSTER_NEW_NODES={ocmnode1} CLUSTER_NEW_VIRTUAL_HOSTNAMES={ocmnode1-vip} -ignoreSysPrereqs -ignorePrereq
Starting Oracle Universal Installer...
Checking Temp space: must be greater than 120 MB. Actual 16182 MB Passed
Checking swap space: must be greater than 150 MB. Actual 4029 MB Passed
Prepare Configuration in progress.
Prepare Configuration successful.
.................................................. 8% Done.
You can find the log of this install session at:
/u01/app/oraInventory/logs/addNodeActions2020-08-08_10-28-51PM.log
SEVERE:Error ocurred while retrieving node numbers of the existing nodes. Please check if clusterware home is properly configured.
[FATAL] [INS-10008] Session initialization failed
CAUSE: An unexpected error occured while initializing the session.
ACTION: Contact Oracle Support Services or refer logs
SUMMARY:
[grid@ocmnode2 addnode]$ more /u01/app/oraInventory/logs/addNodeActions2020-08-08_10-28-51PM.log
Status of node 'ocmnode1':
Node is okay
--------------------------------------------------------------------------
INFO: Setting variable 'REMOTE_CLEAN_MACHINES' to 'ocmnode1'. Received the value from a code block.
INFO: /u01/app/12.1.0/grid/oui/bin/../bin/lsnodes: struct size 0
INFO: Vendor clusterware is not detected.
INFO: Error ocurred while retrieving node numbers of the existing nodes. Please check if clusterware home is properly configured.
SEVERE: Error ocurred while retrieving node numbers of the existing nodes. Please check if clusterware home is properly configured.
INFO: Alert Handler not registered, using Super class functionality
INFO: Alert Handler not registered, using Super class functionality
INFO: User Selected: Yes/OK
INFO: Shutting down OUISetupDriver.JobExecutorThread
SEVERE: [FATAL] [INS-10008] Session initialization failed
CAUSE: An unexpected error occured while initializing the session.
ACTION: Contact Oracle Support Services or refer logs
SUMMARY:
- .
Refer associated stacktrace #oracle.install.commons.util.exception.DefaultErrorAdvisor:299
INFO: Advice is ABORT
SEVERE: Unconditional Exit
INFO: Adding ExitStatus FAILURE to the exit status set
INFO: Finding the most appropriate exit status for the current application
INFO: Exit Status is -1
INFO: Shutdown Oracle Grid Infrastructure 12c Release 1 Installer
INFO: Unloading Setup Driver
Cause: Check on problematic Node if oracleasm is working properly. Run below command to verify.
$ oracleasm init # If get any error, you may try with re-installation oracle asm packages. $ oracleasm scandisks # If not return any records by below query, you have to fix the oracleasm. $ oracleasm listdisks
Solution: Fix the oracleasm issue and make sure all disks are accessible from new node.
Issue # 5: On Oracle Home.
[FATAL] [INS-30160] Installer has detected that the nodes [ocmnode1] specified for addnode operation have uncleaned inventory
[oracle@ocmnode2 addnode]$ pwd
/u01/app/oracle/product/12.1.0/dbhome_1/addnode
[oracle@ocmnode2 addnode]$ ./addnode.sh -silent "CLUSTER_NEW_NODES={ocmnode1}" "CLUSTER_NEW_VIRTUAL_HOSTNAMES={ocmnode1-vip}"
Starting Oracle Universal Installer...
Checking Temp space: must be greater than 120 MB. Actual 12822 MB Passed
Checking swap space: must be greater than 150 MB. Actual 4031 MB Passed
[FATAL] [INS-30160] Installer has detected that the nodes [ocmnode1] specified for addnode operation have uncleaned inventory.
ACTION: Ensure that the inventory location /u01/app/oraInventory is cleaned before performing addnode procedure.
Solution: Remove or Rename oraInventory directory from New Node.
[root@ocmnode1 app]# pwd /u01/app [root@ocmnode1 app]# mv oraInventory oraInventory_bck [root@ocmnode1 app]# ls -lrt total 20 drwxr-xr-x. 3 root oinstall 4096 Aug 7 00:27 12.1.0 drwxrwxr-x 2 grid oinstall 4096 Aug 8 17:40 grid:oinstall drwxrwx--- 5 grid oinstall 4096 Aug 9 20:53 oraInventory_bck drwxrwxr-x. 8 grid oinstall 4096 Aug 10 01:34 grid drwxrwxr-x. 5 oracle oinstall 4096 Aug 10 15:19 oracle

