A few times ago, Oracle Clusterware 10.2.0.1 was installed on RHEL5 or oel5. the main reason was that RHEL5 was not released when Oracle RAC 10.2.0.1 was released, and RedHat was RHEL4, the same problem occurs in SUSE Linux SLES10.
The problem occurs when the installation is started and when the last node runs root. sh. There are three main problems:
Issue #1: To install 10gR2, you must first install the base release, which is 10.2.0.1. As these version of OS are newer, you should use the following command to invoke the installer:
$ RunInstaller-ignoreSysPrereqs // This will bypass the OS check //
Issue #2: At end of root. sh on the last node vipca will fail to run with the following error:
Oracle CRS stack installed and running under init (1 M)
Running vipca (silent) for processing nodeapps
/Home/oracle/crs/oracle/product/10/crs/jdk/jre // bin/java: error while loading
Shared libraries: libpthread. so.0: cannot open shared object file:
No such file or directory
Issue #3: After working around Issue #2 abve, vipca will fail to run with the following error if the vip ip's are in a non-routable range [10. x. x. x, 172. (16-31 ). x. x or 192.168.x.x]:
# Vipca
Error 0 (Native: listNetInterfaces: [3])
[Error 0 (Native: listNetInterfaces: [3])]
The reason is as follows:
These releases of the Linux kernel fix an old bug in the Linux threading that Oracle worked around using LD_ASSUME_KERNEL settings in both vipca and srvctl, this workaround is no longer valid on OEL5 or RHEL5 or SLES10 hence the failures.
Problem 1 is easy to solve. You only need to ignore the check during runInstaller.
The solution to problem 2 is:
To workaround Issue #2 abve, edit vipca (in the CRS bin directory on all nodes) to undo the setting of LD_ASSUME_KERNEL. after the IF statement around line 120 add an unset command to ensure LD_ASSUME_KERNEL is not set as follows:
If ["$ arch" = "i686"-o "$ arch" = "ia64"-o "$ arch" = "x86_64"]
Then
LD_ASSUME_KERNEL = 2.4.19
Export LD_ASSUME_KERNEL
Fi
Unset LD_ASSUME_KERNEL <= Line to be added
Solution to problem 3:
To workaround issue #3 (vipca failing on non-routable vip ip ranges, manually or during root. sh), if you still have the OUI window open, click OK and it will create the "oifcfg" information, then cluvfy will fail due to vipca not completed successfully, skip below in this note and run vipca manually then return to the installer and cluvfy will succeed. otherwise you may configure the interfaces for RAC manually using the oifcfg command as root, like in the following example (from any node ):
/Bin #./oifcfg setif-global eth0/192.168.1.0: public
/Bin #./oifcfg setif-global eth1/10.10.0: cluster_interconnect
/Bin #./oifcfg getif
Eth0 192.168.1.0 global public
Eth1 10.10.10.0 global cluster_interconnect
Then, manually run vipca to add nodeapps resource.
Detailed information is recorded in Oracle notes: 414163.1.