Oracle® Grid Infrastructure Installation Guide 11g Release 2 (11.2) for IBM AIX on POWER Systems (64-Bit) Part Number E24614-03 |
|
|
PDF · Mobi · ePub |
This chapter describes the procedures for installing Oracle Grid Infrastructure for a cluster. Oracle Grid Infrastructure consists of Oracle Clusterware and Oracle Automatic Storage Management. If you plan afterward to install Oracle Database with Oracle Real Application Clusters (Oracle RAC), then this is phase one of a two-phase installation.
This chapter contains the following topics:
Before you install Oracle Grid Infrastructure with the installer, use the following checklist to ensure that you have all the information you will need during installation, and to ensure that you have completed all tasks that must be done before starting your installation. Check off each task in the following list as you complete it, and write down the information needed, so that you can provide it during installation.
Shut Down Running Oracle Processes
You may need to shut down running Oracle processes:
Installing on a node with a standalone database not using Oracle ASM: You do not need to shut down the database while you install Oracle Grid Infrastructure software.
Installing on a node that already has a standalone Oracle Database 11g release 2 (11.2) installation running on Oracle ASM: Stop the existing Oracle ASM instances. The Oracle ASM instances are restarted during the installation.
Installing on an Oracle RAC Database node: This installation requires an upgrade of Oracle Clusterware, as Oracle Clusterware is required to run Oracle RAC. As part of the upgrade, you must shut down the database one node at a time as the rolling upgrade proceeds from node to node.
Note:
If you are upgrading an Oracle RAC 9i release 2 (9.2) node, and the TNSLSNR is listening to the same port on which the SCAN listens (default 1521), then the TNSLSNR should be shut down.If a Global Services Daemon (GSD) from Oracle9i Release 9.2 or earlier is running, then stop it before installing Oracle Grid Infrastructure by running the following command:
$ Oracle_home/bin/gsdctl stop
where Oracle_home
is the Oracle Database home that is running the GSD.
Caution:
If you have an existing Oracle9i release 2 (9.2) Oracle Cluster Manager (Oracle CM) installation, then do not shut down the Oracle CM service. Shutting down the Oracle CM service prevents the Oracle Grid Infrastructure 11g release 2 (11.2) software from detecting the Oracle9i release 2 node list, and causes failure of the Oracle Grid Infrastructure installation.Note:
If you receive a warning to stop all Oracle services after starting OUI, then run the commandOracle_home/bin/localconfig delete
where Oracle_home
is the existing Oracle Clusterware home.
Prepare for Oracle Automatic Storage Management and Oracle Clusterware Upgrade If You Have Existing Installations
During the Oracle Grid Infrastructure installation, existing Oracle Clusterware and clustered Oracle ASM installations are both upgraded.
When all member nodes of the cluster are running Oracle Grid Infrastructure 11g release 2 (11.2), then the new clusterware becomes the active version.
If you intend to install Oracle RAC, then you must first complete the upgrade to Oracle Grid Infrastructure 11g release 2 (11.2) on all cluster nodes before you install the Oracle Database 11g release 2 (11.2) version of Oracle RAC.
Note:
All Oracle Grid Infrastructure upgrades (upgrades of existing Oracle Clusterware and Oracle ASM installations) are out-of-place upgrades.Determine the Oracle Inventory (oraInventory) location
If you have already installed Oracle software on your system, then OUI detects the existing Oracle Inventory (oraInventory) directory from the /var/opt/oracle/oraInst.loc
file, and uses this location. This directory is the central inventory of Oracle software installed on your system. Users who have the Oracle Inventory group as their primary group are granted the OINSTALL privilege to write to the central inventory.
If you are installing Oracle software for the first time on your system, and your system does not have an oraInventory directory, then the installer designates the installation owner's primary group as the Oracle Inventory group. Ensure that this group is available as a primary group for all planned Oracle software installation owners.
Note:
The oraInventory directory cannot be placed on a shared file system.See Also:
The preinstallation chapters in Chapter 2 for information about creating the Oracle Inventory, and completing required system configurationObtain root account access
During installation, you are asked to run configuration scripts as the root
user. You must run these scripts as root
, or be prepared to have your system administrator run them for you. You must run the root.sh
script on the first node and wait for it to finish. If your cluster has four or more nodes, then root.sh
can be run concurrently on all nodes but the first and last.
Decide if you want to install other languages
During installation, you are asked if you want translation of user interface text into languages other than the default, which is English.
Note:
If the language set for the operating system is not supported by the installer, then by default the installer runs in the English language.See Also:
Oracle Database Globalization Support Guide for detailed information on character sets and language configurationDetermine your cluster name, public node names, the SCAN, virtual node names, GNS VIP and planned interface use for each node in the cluster
During installation, you are prompted to provide the public and virtual host name, unless you use a third party cluster software. In that case, the public host name information will be filled in. You are also prompted to identify which interfaces are public, private, or interfaces in use for another purpose, such as a network file system.
If you use Grid Naming Service (GNS), then OUI displays the public and virtual host name addresses labeled as "AUTO" because they are configured automatically.
Note:
If you configure IP addresses manually, then avoid changing host names after you complete the Oracle Grid Infrastructure installation, including adding or deleting domain qualifications. A node with a new host name is considered a new host, and must be added to the cluster. A node under the old name will appear to be down until it is removed from the cluster.If you use third-party clusterware, then use your vendor documentation to complete setup of your public and private domain addresses.
When you enter the public node name, use the primary host name of each node. In other words, use the name displayed by the hostname
command.
In addition:
Provide a cluster name with the following characteristics:
It must be globally unique throughout your host domain.
It must be at least one character long and less than or equal to 15 characters long.
It must consist of the same character set used for host names, in accordance with RFC 1123: Hyphens (-), and single-byte alphanumeric characters (a to z, A to Z, and 0 to 9). If you use third-party vendor clusterware, then Oracle recommends that you use the vendor cluster name.
If you are not using Grid Naming Service (GNS), then determine a virtual host name for each node. A virtual host name is a public node name that is used to reroute client requests sent to the node if the node is down. Oracle Database uses VIPs for client-to-database connections, so the VIP address must be publicly accessible. Oracle recommends that you provide a name in the format hostname-vip. For example: myclstr2-vip
.
Provide SCAN addresses for client access to the cluster. These addresses should be configured as round robin addresses on the domain name service (DNS). Oracle recommends that you supply three SCAN addresses.
Note:
The following is a list of additional information about node IP addresses:For the local node only, OUI automatically fills in public and VIP fields. If your system uses vendor clusterware, then OUI may fill additional fields.
Host names and virtual host names are not domain-qualified. If you provide a domain in the address field during installation, then OUI removes the domain from the address.
Interfaces identified as private for private IP addresses should not be accessible as public interfaces. Using public interfaces for Cache Fusion can cause performance problems.
Identify public and private interfaces. OUI configures public interfaces for use by public and virtual IP addresses, and configures private IP addresses on private interfaces.
The private subnet that the private interfaces use must connect all the nodes you intend to have as cluster members.
Obtain proxy realm authentication information if you have a proxy realm on your network
During installation, OUI attempts to download updates. You are prompted to provide a proxy realm, and user authentication information to access the Internet through the proxy service. If you have a proxy realm configured, then be prepared to provide this information. If you do not have a proxy realm, then you can leave the proxy authentication fields blank.
Identify shared storage for Oracle Clusterware files and prepare storage if necessary
During installation, you are asked to provide paths for the following Oracle Clusterware files. These files must be shared across all nodes of the cluster, either on Oracle ASM, or on a supported file system:
Voting disks are files that Oracle Clusterware uses to verify cluster node membership and status.
Voting disk files must be owned by the user performing the installation (oracle
or grid
), and must have permissions set to 640.
Oracle Cluster Registry files (OCR) contain cluster and database configuration information for Oracle Clusterware.
Before installation, OCR files must be owned by the user performing the installation (grid
or oracle
). That installation user must have oinstall
as its primary group. During installation, OUI changes ownership of the OCR files to root
.
If your file system does not have external storage redundancy, then Oracle recommends that you provide two additional locations for the OCR disk, and two additional locations for the voting disks, for a total of six partitions (three for OCR, and three for voting disks). Creating redundant storage locations protects the OCR and voting disk in the event of a failure. To completely protect your cluster, the storage locations given for the copies of the OCR and voting disks should have completely separate paths, controllers, and disks, so that no single point of failure is shared by storage locations.
When you select to store the OCR on Oracle ASM, the default configuration is to create the OCR on one ASM disk group. If you create the disk group with normal or high redundancy, then the OCR is protected from physical disk failure.
To protect the OCR from logical disk failure, create another ASM disk group after installation and add the OCR to the second diskgroup using the ocrconfig
command.
See Also:
Chapter 2, "Advanced Installation Oracle Grid Infrastructure for a Cluster Preinstallation Tasks" and Oracle Database Storage Administrator's Guide for information about adding disks to diskgroupsEnsure cron jobs do not run during installation
If the installer is running when daily cron jobs start, then you may encounter unexplained installation problems if your cron job is performing cleanup, and temporary files are deleted before the installation is finished. Oracle recommends that you complete installation before daily cron jobs are run, or disable daily cron jobs that perform cleanup until after the installation is completed.
Ensure that the Oracle home path you select for the Oracle Grid Infrastructure home uses only ASCII characters
This restriction includes installation owner user names, which are used as a default for some home paths, as well as other directory names you may select for paths.
Unset Oracle environment variables. If you have set ORA_CRS_HOME as an environment variable, then unset it before starting an installation or upgrade. You should never use ORA_CRS_HOME as an environment variable.
If you have had an existing installation on your system, and you are using the same user account to install this installation, then unset the following environment variables: ORA_CRS_HOME; ORACLE_HOME; ORA_NLS10; TNS_ADMIN
Decide if you want to use the Software Updates option. OUI can install critical patch updates, system requirements updates (hardware, operating system parameters, and kernel packages) for supported operating systems, and other significant updates that can help to ensure your installation proceeds smoothly. Oracle recommends that you enable software updates during installation.
If you choose to enable software updates, then during installation you must provide a valid My Oracle Support user name and password during installation, so that OUI can download the latest updates, or you must provide a path to the location of an software updates packages that you have downloaded previously.
If you plan to run the installation in a secured data center, then you can download updates before starting the installation by starting OUI on a system that has Internet access in update download mode. To start OUI to download updates, enter the following command:
$ ./runInstaller -downloadUpdates
Provide the My Oracle Support user name and password, and provide proxy settings if needed. After you download updates, transfer the update file to a directory on the server where you plan to run the installation.
This section provides you with information about how to use the installer to install Oracle Grid Infrastructure. It contains the following sections:
Complete the following steps to install grid infrastructure (Oracle Clusterware and Oracle Automatic Storage Management) on your cluster. At any time during installation, if you have a question about what you are being asked to do, click the Help button on the OUI page.
Change to the /Disk1 directory on the installation media, or where you have downloaded the installation binaries, and run the runInstaller
command. For example:
$ cd /home/grid/oracle_sw/Disk1 $ ./runInstaller
Select Typical or Advanced installation.
Provide information or run scripts as root
when prompted by OUI. If you need assistance during installation, click Help. Click Details to see the log file. If root.sh
fails on any of the nodes, then you can fix the problem and follow the steps in Section 6.5, "Deconfiguring Oracle Clusterware Without Removing Binaries," rerun root.sh
on that node, and continue.
Note:
You must run theroot.sh
script on the first node and wait for it to finish. If your cluster has four or more nodes, then root.sh
can be run concurrently on all nodes but the first and last. As with the first node, the root.sh
script on the last node must be run separately.After you run root.sh
on all the nodes, OUI runs Net Configuration Assistant (netca
) and Cluster Verification Utility. These programs run without user intervention.
Oracle Automatic Storage Management Configuration Assistant (asmca
) configures Oracle ASM during the installation.
When you have verified that your Oracle Grid Infrastructure installation is completed successfully, you can either use it to maintain high availability for other applications, or you can install an Oracle database.
If you intend to install Oracle Database 11g release 2 (11.2) with Oracle RAC, then refer to Oracle Real Application Clusters Installation Guide for IBM AIX on POWER Systems (64-Bit).
See Also:
Oracle Real Application Clusters Administration and Deployment Guide for information about using cloning and node addition procedures, and Oracle Clusterware Administration and Deployment Guide for cloning Oracle Grid InfrastructureDuring installation of Oracle Grid Infrastructure, you are given the option either of providing cluster configuration information manually, or of using a cluster configuration file. A cluster configuration file is a text file that you can create before starting OUI, which provides OUI with cluster node addresses that it requires to configure the cluster.
Oracle suggests that you consider using a cluster configuration file if you intend to perform repeated installations on a test cluster, or if you intend to perform an installation on many nodes.
To create a cluster configuration file manually, start a text editor, and create a file that provides the name of the public and virtual IP addresses for each cluster member node, in the following format:
node1 node1-vip node2 node2-vip . . .
For example:
mynode1 mynode1-vip mynode2 mynode2-vip
This section contains the following tasks:
Note:
Oracle recommends that only advanced users should perform the software-only installation, as this installation option requires manual postinstallation steps to enable the Oracle Grid Infrastructure software.A software-only installation consists of installing Oracle Grid Infrastructure for a cluster on one node.
If you use the Install Grid Infrastructure Software Only option during installation, then this installs the software binaries on the local node. To complete the installation for your cluster, you must perform the additional steps of configuring Oracle Clusterware and Oracle ASM, creating a clone of the local installation, deploying this clone on other nodes, and then adding the other nodes to the cluster.
To perform a software-only installation:
On the local node, verify that the cluster node meets installation requirements using the command runcluvfy.sh stage -pre crsinst
. Ensure that you have completed all storage and server preinstallation requirements.
Run the runInstaller
command from the relevant directory on the Oracle Database 11g release 2 (11.2) installation media or download directory. For example:
$ cd /home/grid/oracle_sw/Disk1 $ ./runInstaller
Complete a software-only installation of Oracle Grid Infrastructure on the first node.
When the software has been installed, run the orainstRoot.sh
script when prompted.
For installations with Oracle RAC release 11.2.0.2 and later, proceed to step 6. For installations with Oracle RAC release 11.2.0.1, to relink Oracle Clusterware with the Oracle RAC option enabled, run commands similar to the following (in this example, the Grid home is /u01/app/11.2.0/grid
):
$ cd /u01/app/11.2.0/grid/ $ set env ORACLE_HOME pwd $ cd rdbms/lib $ make -f ins_rdbms.mk rac_on ioracle
The root.sh
script output provides information about how to proceed, depending on the configuration you plan to complete in this installation. Make note of this information.
However, ignore the instruction to run the roothas.pl
script, unless you intend to install Oracle Grid Infrastructure on a standalone server (Oracle Restart).
On each remaining node, verify that the cluster node meets installation requirements using the command runcluvfy.sh stage -pre crsinst
. Ensure that you have completed all storage and server preinstallation requirements.
Use Oracle Universal Installer as described in steps 1 through 4 to install the Oracle Grid Infrastructure software on every remaining node that you want to include in the cluster, and complete a software-only installation of Oracle Grid Infrastructure on every node.
Configure the cluster using the full OUI configuration wizard GUI as described in Section 4.3.2, "Configuring the Software Binaries," or configure the cluster using a response file as described in section Section 4.3.3, "Configuring the Software Binaries Using a Response File."
Configure the software binaries by starting Oracle Grid Infrastructure configuration wizard in GUI mode, available in release 11.2.0.2 and later:
Log in to a terminal as the Grid infrastructure installation owner, and change directory to grid_home
/crs/config
.
Enter the following command:
$ ./config.sh
The configuration script starts OUI in Configuration Wizard mode. Provide information as needed for configuration. Each page shows the same user interface and performs the same validation checks that OUI normally does. However, instead of running an installation, The configuration wizard mode validates inputs and configures the installation on all cluster nodes.
When you complete inputs, OUI shows you the Summary page, listing all inputs you have provided for the cluster. Verify that the summary has the correct information for your cluster, and click Install to start configuration of the local node.
When configuration of the local node is complete, OUI copies the Oracle Grid Infrastructure configuration file to other cluster member nodes.
When prompted, run root scripts.
When you confirm that all root scripts are run, OUI checks the cluster configuration status, and starts other configuration tools as needed.
When you install or copy Oracle Grid Infrastructure software on any node, you can defer configuration for a later time. This section provides the procedure for completing configuration after the software is installed or copied on nodes, using the configuration wizard utility (config.sh
), available in release 11.2.0.2 and later.
See Also:
Oracle Clusterware Administration and Deployment Guide for more information about the configuration wizard.To configure the Oracle Grid Infrastructure software binaries using a response file:
As the Oracle Grid Infrastructure installation owner (grid
) start OUI in Oracle Grid Infrastructure configuration wizard mode from the Oracle Grid Infrastructure software-only home using the following syntax, where Grid_home is the Oracle Grid Infrastructure home, and filename is the response file name:
Grid_home/crs/config/config.sh [-debug] [-silent -responseFile filename]
For example:
$ cd /u01/app/grid/crs/config/ $ ./config.sh -responseFile /u01/app/grid/response/response_file.rsp
The configuration script starts OUI in Configuration Wizard mode. Each page shows the same user interface and performs the same validation checks that OUI normally does. However, instead of running an installation, The configuration wizard mode validates inputs and configures the installation on all cluster nodes.
When you complete inputs, OUI shows you the Summary page, listing all inputs you have provided for the cluster. Verify that the summary has the correct information for your cluster, and click Install to start configuration of the local node.
When configuration of the local node is complete, OUI copies the Oracle Grid Infrastructure configuration file to other cluster member nodes.
When prompted, run root scripts.
When you confirm that all root scripts are run, OUI checks the cluster configuration status, and starts other configuration tools as needed.
The Oracle Berkeley DB embedded database installation that is included with Oracle Grid Infrastructure for a Cluster 11g release 2 (11.2.0.2) is only for use with the Oracle Grid Infrastructure installation products. Refer to terms of the Berkeley DB license at the following URL for details:
http://www.oracle.com/technetwork/database/berkeleydb/downloads/index.html
After installation, log in as root
, and use the following command syntax on each node to confirm that your Oracle Clusterware installation is installed and running correctly:
crsctl check crs
For example:
$ crsctl check crs CRS-4638: Oracle High Availability Services is online CRS-4537: Cluster Ready Services is online CRS-4529: Cluster Synchronization Services is online CRS-4533: Event Manager is online
Caution:
After installation is complete, do not remove manually or run cron jobs that remove/tmp/.oracle
or /var/tmp/.oracle
or its files while Oracle Clusterware is up. If you remove these files, then Oracle Clusterware could encounter intermittent hangs, and you will encounter error CRS-0184: Cannot communicate with the CRS daemon.If you installed the OCR and voting disk files on Oracle ASM, then use the following command syntax as the Grid Infrastructure installation owner to confirm that your Oracle ASM installation is running:
srvctl status asm
For example:
$ srvctl status asm ASM is running on node1,node2
Oracle ASM is running only if it is needed for Oracle Clusterware files. If you have not installed OCR and voting disks files on Oracle ASM, then the Oracle ASM instance should be down.
Note:
To manage Oracle ASM or Oracle Net 11g release 2 (11.2) or later installations, use thesrvctl
binary in the Oracle Grid Infrastructure home for a cluster (Grid home). If you have Oracle Real Application Clusters or Oracle Database installed, then you cannot use the srvctl
binary in the database home to manage Oracle ASM or Oracle Net.After installation, log in as root
, and use the following command syntax on each node to confirm that your Oracle Clusterware installation is installed and running correctly:
crsctl check crs
For example:
$ crsctl check crs CRS-4638: Oracle High Availability Services is online CRS-4537: Cluster Ready Services is online CRS-4529: Cluster Synchronization Services is online CRS-4533: Event Manager is online
Caution:
After installation is complete, do not remove manually or run cron jobs that remove/tmp/.oracle
or /var/tmp/.oracle
or its files while Oracle Clusterware is up. If you remove these files, then Oracle Clusterware could encounter intermittent hangs, and you will encounter error CRS-0184: Cannot communicate with the CRS daemon.