PK
x=Aoa, mimetypeapplication/epub+zipPK x=A iTunesMetadata.plistp
This appendix describes the steps for noninteractive installations during which you can create Oracle Real Application Clusters (Oracle RAC) databases using scripts.
Note: The scripts generated by Database Configuration Assistant (DBCA) are for reference only. Oracle strongly recommends that you use DBCA to create a database. |
See Also: Oracle Universal Installer and OPatch User's Guide for Windows and UNIX for detailed information about generating scripts |
This appendix contains the following topics:
About DBCA Noninteractive (Silent) Configuration for Oracle RAC
Using DBCA Noninteractive (Silent) Configuration for Oracle RAC
To generate scripts to create an Oracle RAC database, create a database using the scripts, and prepare the database for use, complete the following steps:
Start DBCA and select your preferred options to build the Oracle RAC database.
On the Creation Options page of your DBCA session, deselect Create Database and select Generate Database Creation Scripts before you click Finish. You can accept the default destination directory for the scripts, or browse for a different location. In either case, you should note the path name for use in the next step.
See Also: Chapter 3, "Creating Oracle Real Application Clusters Databases with Database Configuration Assistant" for additional information about running a DBCA session |
Navigate to the directory where DBCA created the scripts, and review the SQL scripts to ensure that they contain the statements to build a database with the characteristics you require. If they do not, then Oracle recommends that you rerun DBCA to create scripts with the desired configuration rather than editing the scripts yourself.
On each cluster node you identified during your DBCA session, run the script sid
.sh
, where sid
is the SID prefix that you entered on the DBCA Database Name page.
Set the initialization parameter, cluster_database
, to the value TRUE
in your SPFILE
by running an ALTER SYSTEM
command, or by uncommenting it in your PFILE
for each instance.
Configure Oracle Net Services to support your new database and instances as described in Chapter 7, "Understanding the Oracle Real Application Clusters Installed Configuration".
Set the local_listener
and remote_listener
parameters in your SPFILE
by issuing an ALTER SYSTEM
command, or by uncommenting it in your PFILE
for each instance.
Run SRVCTL
to configure and start database and instance applications as described in Oracle Real Application Clusters Administration and Deployment Guide.
See Also: The Oracle Universal Installer and OPatch User's Guide for Windows and UNIX for additional information about creating and using scripts to install Oracle software with Oracle Universal Installer (OUI) |
You can perform a noninteractive, or silent configuration of Oracle RAC using DBCA. To perform a silent configuration, you must have completed an Oracle Grid Infrastructure (Oracle Clusterware and Oracle ASM) installation, run the root.sh
script from the Oracle Database home, and defined the Oracle home directory environment variable.
You can use DBCA to create a database from templates supplied by Oracle, or from templates that you create. The templates contain settings optimized for a particular type of workload.
Oracle provides templates for the following two workload types:
General purpose or transaction processing
Data warehouse
For more complex environments, you can select the Custom Database option. This option does not use templates and results in a more extensive installation interview, which means that it takes longer to create your database.
The following command syntax creates an Oracle RAC database in an existing Oracle ASM disk group, where the cluster nodes are node1 and node2, and where the disk group name is ASMgrp1, the passwords are my_password, and the ASM SYS password is My_ASM_Password:
# su oracle -c "$ORACLE_HOME/bin/dbca -silent -createDatabase -templateName General_Purpose.dbc -gdbName $DBNAME -sid $SID -sysPassword my_password -systemPassword my_password -sysmanPassword my_password -dbsnmpPassword my_password -emConfiguration LOCAL -storageType ASM -diskGroupName ASMgrp1 -datafileJarLocation $ORACLE_HOME/assistants/dbca/templates -nodeinfo node1,node2 -characterset WE8ISO8859P1 -obfuscatedPasswords false -sampleSchema false -asmSysPassword My_ASM_password"
When you start OUI, you can use a response file to automate the installation and configuration of Oracle software, either fully or partially. OUI uses the values contained in the response file to provide answers to some or all installation prompts.
Typically, the installer runs in interactive mode, which means that it prompts you to provide information in graphical user interface (GUI) screens. When you use response files to provide this information, you run the installer from a command prompt using either of the following modes:
If you include responses for all of the prompts in the response file and specify the -silent
option when starting the installer, then it runs in silent mode. During a silent mode installation, the installer does not display any screens. Instead, it displays progress information in the terminal that you used to start it.
If you include responses for some or all of the prompts in the response file and omit the -silent
option, then the installer runs in response file mode. During a response file mode installation, the installer displays all the screens, screens for which you specify information in the response file, and also screens for which you did not specify the required information in the response file.
You define the settings for a silent or response file installation by entering values for the variables listed in the response file. For example, to specify the Oracle home name, supply the appropriate value for the ORACLE_HOME
variable:
ORACLE_HOME="OraDBHome1"
Another way of specifying the response file variable settings is to pass them as command line arguments when you run the installer. For example:
-silent "ORACLE_HOME=OraDBHome1" ...
In this command, directory_path
is the path of the database
directory on the DVD or the path of the Disk1
directory on the hard drive.
This method is particularly useful if you do not want to embed sensitive information, such as passwords, in the response file. For example:
-silent "s_dlgRBOPassword=binks342" ...
Enclose the variable and its setting in double quotes.
See Also: The Oracle Universal Installer and OPatch User's Guide for Windows and UNIX for more information about response files |
The following table provides use cases for running the installer in silent mode or response file mode.
Before you create a database that uses Oracle Automatic Storage Management (Oracle ASM), you must run the root.sh
script. For this reason, you cannot create a database using Oracle ASM as the storage option for database files during a silent-mode installation. Instead, you can complete a software-only installation using silent-mode, and then run the Oracle Net Configuration Assistant and DBCA in silent mode after you have completed the software-only installation and you have run the root.sh
script.
Note: This limitation applies only to databases that use Oracle Automatic Storage Management as the storage option for database files. You can create a database that uses the file system option during a silent-mode installation. |
The following are the general steps to install and configure Oracle products using the installer in silent or response file mode:
Note: You must complete all required preinstallation tasks on a system before running the installer in silent or response file mode. |
Create the oraInst.loc
file if it is not present on the server.
Prepare a response file.
Run the installer in silent or response file mode.
If you completed a software-only installation, then run Net Configuration Assistant and DBCA in silent or response file mode.
These steps are described in the following sections.
This section describes the following methods to prepare response files for use during silent mode or response file mode installations:
Oracle provides response file templates for each product and installation type and for each configuration tool. These files are located at database/response
directory on the installation media.
Note: If you copied the software to a hard disk, then the response files are located in the directory/Disk1/response . |
Table B-1 lists the response files provided with this software:
Table B-1 Response Files for Oracle Database
Response File | Description |
---|---|
Silent installation of Oracle Database 11g | |
Silent installation of DBCA | |
Silent installation of Oracle Net Configuration Assistant |
Caution: When you modify a response file template and save a file for use, the response file may contain plain text passwords. Ownership of the response file should be given to the Oracle software installation owner only, and permissions on the response file should be changed to 600. Oracle strongly recommends that database administrators or other administrators delete or secure response files when they are not in use. |
To copy and modify a response file:
Copy the response file from the response file directory to a directory on your system:
$ cp /directory_path/response/response_file.rsp local_directory
In this example, directory_path
is the path to the database
directory on the installation media. If you have copied the software to a hard drive, then you can edit the file in the response
directory if you prefer.
Open the response file in a text editor:
$ vi /local_dir/response_file.rsp
Remember that you can specify sensitive information, such as passwords, at the command line rather than within the response file. "How Response Files Work" explains this method.
See Also: The Oracle Universal Installer and OPatch User's Guide for Windows and UNIXfor detailed information on creating response files |
Follow the instructions in the file to edit it.
Note: The installer or configuration assistant fails if you do not correctly configure the response file. |
Change the permissions on the file to 600:
$ chmod 600 /local_dir/response_file.rsp
Note: A fully specified response file for an Oracle Database installation contains the passwords for database administrative accounts and for a user who is a member of the OSDBA group (required for automated backups). Ensure that only the Oracle software owner user can view or modify response files or consider deleting them after the installation succeeds. |
You can use OUI in interactive mode to record response files, which you can edit and then use to complete silent mode or response file mode installations. This method is useful for Advanced or software-only installations.
Starting with Oracle Database 11g Release 2 (11.2), you can save all of the installation steps into a response file during installation by clicking Save Response File on the Summary page. You can use the generated response file for a silent installation later.
When you record the response file, you can either complete the installation, or you can exit from the installer on the Summary page, before it starts to copy the software to the system.
If you use record mode during a response file mode installation, then the installer records the variable values that were specified in the original source response file into the new response file.
Note: You cannot save passwords while recording the response file. |
To record a response file:
Complete preinstallation tasks as for a normal installation.
When you run the installer to record a response file, it checks the system to verify that it meets the requirements to install the software. For this reason, Oracle recommends that you complete all of the required preinstallation tasks and record the response file while completing an installation.
Ensure that the Oracle software owner user (typically oracle
) has permissions to create or write to the Oracle home path that you will specify when you run the installer.
On each installation screen, specify the required information.
When the installer displays the Summary screen, perform the following:
Click Save Response File and specify a file name and location to save the values for the response file.
Click Finish to create the response file and continue with the installation.
Click Cancel if you do not want to continue with the installation. The installation stops, but the recorded response file is retained.
If you do not complete the installation, then delete the Oracle home directory that the installer created using the path you specified in the Specify File Locations screen.
Before you use the saved response file on another system, edit the file and make any required changes.
Use the instructions in the file as a guide when editing it.
Now, you are ready to run OUI at the command line, specifying the response file you created, to perform the installation. The OUI executable, runInstaller
, provides several options. For help information on the full set of these options, run the runInstaller
command with the -help
option, for example:
$ directory_path/runInstaller -help
The help appears in a window after some time.
To run the installer using a response file:
Complete the preinstallation tasks as for a normal installation
Log in as the software installation owner user.
If you are completing a response file mode installation, set the DISPLAY
environment variable.
Note: You do not have to set theDISPLAY environment variable if you are completing a silent mode installation. |
To start the installer in silent or response file mode, enter a command similar to the following:
$ /directory_path/runInstaller [-silent] [-noconfig] \ -responseFile responsefilename
Note: Do not specify a relative path for the response file. If you specify a relative path, then the installer fails. |
In this example:
directory_path
is the path of the database
directory on the DVD or the path of the Disk1
directory on the hard drive.
-silent
indicates that you want to run the installer in silent mode.
-noconfig
suppresses running the configuration assistants during installation, and a software-only installation is performed instead.
responsefilename
is the full path and file name of the installation response file that you configured.
Note: For more information about other options for therunInstaller command, enter the following command:
$ /directory_path/runInstaller -help
|
When the installation completes, log in as the root
user and run the root.sh
script:
$ sudo sh
password:
# /oracle_home_path/root.sh
You can run Net Configuration Assistant in silent mode to configure and start an Oracle Net listener on the system, configure naming methods, and configure Oracle Net service names. To run Net Configuration Assistant in silent mode, you must copy and edit a response file template. Oracle provides a response file template named netca.resp
in the response
dir3ectory in the database
/response
directory on the DVD.
Note: If you copied the software to a hard disk, then the response file template is located in thedatabase/response directory. |
To run Net Configuration Assistant using a response file:
Copy the netca.rsp
response file template from the response file directory to a directory on your system:
$ cp /directory_path/response/netca.rsp local_directory
In this example, directory_path
is the path of the database
directory on the DVD. If you have copied the software to a hard drive, you can edit the file in the response
directory if you prefer.
Open the response file in a text editor:
$ vi /local_dir/netca.rsp
Follow the instructions in the file to edit it.
Note: Net Configuration Assistant fails if you do not correctly configure the response file. |
Log in as the Oracle software owner user, and set the ORACLE_HOME
environment variable to specify the correct Oracle home directory.
Enter a command similar to the following to run Net Configuration Assistant in silent mode:
$ $ORACLE_HOME/bin/netca /silent /responsefile /local_dir/netca.rsp
In this command:
The /silent
option indicates to run Net Configuration Assistant in silent mode.
local_dir
is the full path of the directory where you copied the netca.rsp
response file template.
You can run configuration assistants in response file or silent mode to configure and start Oracle software after it is installed on the system. To run configuration assistants in response file or silent mode, you must copy and edit a response file template.
Note: If you copied the software to a hard disk, then the response file template is located in theDisk1/response directory. |
This section contains the following topics:
About Database Configuration Assistant in Response File Mode
Running Database Configuration Assistant in Response File or Silent Mode
Running Postinstallation Configuration Using a Response File
In the response file mode, DBCA uses values that you specify, in the response file or as command line options, to create a database. As it configures and starts the database, it displays a window that contains status messages and a progress bar. The window that it displays is the same window that is displayed when you create a preconfigured database during an Enterprise Edition or Standard Edition installation.
To run DBCA in response file mode, use a graphical display and set the DISPLAY
environment variable. Use -progressOnly
flag to set the run mode to response file.
Oracle provides a response file template named dbca.rsp
in the /Disk1
/response
directory on the DVD.
To run DBCA in response file or silent mode:
Copy the dbca.rsp
response file template from the response file directory to a directory on your system:
$ cp /directory_path/response/dbca.rsp local_directory
In this example, directory_path
is the path of the database
directory on the DVD. If you have copied the software to a hard drive, you can edit the file in the response
directory if you prefer.
Note: As an alternative to editing the response file template, you can also create a database by specifying all required information as command line options when you run DBCA. For information about the list of options supported, enter the following command:$ $ORACLE_HOME/bin/dbca -help |
Open the response file in a text editor:
$ vi /local_dir/dbca.rsp
Edit the file, following the instructions in the file.
Note: DBCA fails if you do not correctly configure the response file. |
Log in as the Oracle software owner user, and set the ORACLE_HOME
environment variable to specify the correct Oracle home directory.
To run DBCA in response file mode, set the DISPLAY
environment variable.
Use the following command syntax to run DBCA in silent or response file mode using a response file:
$ORACLE_HOME/bin/dbca {-progressOnly | -silent} -responseFile \
/local_dir/dbca.rsp
In this example:
The -silent
option indicates that you want to run DBCA in silent mode.
The -progressOnly
option indicates that you want to run DBCA in response file mode.
local_dir
is the full path of the directory where you copied the dbca.rsp
response file template.
The following sections explain how to create and run a response file configuration after installing Oracle software. The topics in this section are:
When you run a silent or response file installation, you provide information about your servers in a response file that you otherwise provide manually during a graphical user interface installation. However, the response file does not contain passwords for user accounts that configuration assistants require after software installation is complete. The configuration assistants are started with a script called configToolAllCommands
. You can run this script in response file mode by using a password response file. The script uses the passwords to run the configuration tools in succession to complete configuration.
If you keep the password file to use for clone installations, then Oracle strongly recommends that you store the password file in a secure location. In addition, if you have to stop an installation to fix an error, then you can run the configuration assistants using configToolAllCommands
and a password response file.
The configToolAllCommands
password response file has the following syntax options:
internal_component_name is the name of the component that the configuration assistant configures
variable_name is the name of the configuration file variable
value is the desired value to use for configuration.
The command syntax is as follows:
internal_component_name|variable_name=value
For example:
oracle.assistants.asm|S_ASMPASSWORD=myPassWord
Oracle strongly recommends that you maintain security with a password response file:
Permissions on the response file should be set to 600.
The owner of the response file should be the installation owner user, with the group set to the central inventory (oraInventory) group.
To run configuration assistants with the configToolAllCommands
script:
Create a response file using the syntax filename.properties. For example:
$ touch cfgrsp.properties
Open the file with a text editor, and cut and paste the password template, modifying as needed.
Example B-1 Password response file for Oracle Real Application Clusters
Oracle Database configuration requires the SYS, SYSTEM, SYSMAN, and DBSNMP passwords for use with DBCA. The S_ASMSNMPPASSWORD response is necessary only if the database is using Oracle ASM for storage. Also, if you selected to configure Oracle Enterprise Manager, then you must provide the password for the Oracle software installation owner for the S_HOSTUSERPASSWORD response.
oracle.assistants.server|S_SYSPASSWORD=password oracle.assistants.server|S_SYSTEMPASSWORD=password oracle.assistants.server|S_SYSMANPASSWORD=password oracle.assistants.server|S_DBSNMPPASSWORD=password oracle.assistants.server|S_HOSTUSERPASSWORD=password oracle.assistants.server|S_ASMSNMPPASSWORD=password
If you do not want to enable Oracle Enterprise Manager or Oracle ASM, then leave those password fields blank
Change permissions to secure the file. For example:
$ ls -al cfgrsp.properties -rw------- 1 oracle oinstall 0 Apr 30 17:30 cfgrsp.properties
Change directory to $ORACLE_BASE/cfgtoollogs
Run the configuration script using the following syntax:
configToolAllCommands RESPONSE_FILE=/path/name.properties
For example:
$ ./configToolAllCommands RESPONSE_FILE=/home/oracle/cfgrsp.properties
This chapter describes the Oracle Real Application Clusters (Oracle RAC) installed configuration. This chapter contains the following topics:
Database Components Created Using Database Configuration Assistant
Oracle Database Default Audit Policy and Initialization Parameters
Oracle Net Configuration Assistant (NETCA) and Database Configuration Assistant (DBCA) configure your environment to meet the requirements for database creation and Oracle Enterprise Manager discovery of Oracle RAC databases.
Note: Configuration files are created on each node in your cluster database. |
Avoid changing host names after you complete your Oracle RAC installation, including adding or deleting domain qualifications. Hostnames are created during an Oracle Clusterware installation and are used extensively with database processes. Nodes with changed host names must be deleted from the cluster and added back with the new name.
Oracle RAC requires that all cluster nodes have the same time zone setting. During an Oracle Clusterware installation, the installation process determines the time zone setting of the Grid installation owner on the node where Oracle Universal Installer (OUI) runs, and uses that time zone value on all of the nodes as the default time zone setting for all processes that Oracle Clusterware manages. This default setting is used for databases, Oracle ASM, and any other managed processes. However, if you start an instance with SQL*Plus, you must ensure that the time zone value that Oracle RAC uses is the same as the Oracle Clusterware time zone. You can change the time zone that Oracle Clusterware uses for a database by running the command srvctl setenv database -T 'TZ=
time zone
.
Oracle creates an entry for each Oracle RAC database in the oratab
configuration file. The oratab
file is created by the root.sh
script during installation, and it is updated by the Database Configuration Assistant when creating or deleting a database. The oratab
file entry is also created automatically by the Database Agent when a database is first started on a node where it has not run previously. Oracle ASM Agent creates Oracle ASM oratab
entries.
Oracle Enterprise Manager uses the oratab file during service discovery to determine the name of the Oracle RAC database, and to determine if the database should be started automatically when the system is restarted.
The database entry has the following syntax:
$DB_UNIQUE_NAME:$ORACLE_HOME:N
A colon (:) is used as the field terminator. A new line terminates the entry. Lines beginning with a pound sign (#) are comments. Because all the instances of an Oracle RAC database have the same DB_UNIQUE_NAME
, but each instance has its own ORACLE_SID
, use the $DB_UNIQUE_NAME
environment variable in the oratab
file as the database entry.
The $DB_UNIQUE_NAME
identifier for your Oracle RAC database must be unique across your enterprise. $ORACLE_HOME
is the directory path to the database, and N
indicates that the database should not be started at restart time. The following is an example entry for a database named sales
:
sales:/u01/app/oracle/sales:N
This section describes the database components that DBCA creates, which include:
For both single-instance and cluster database environments, an Oracle Database is divided into smaller logical areas of space known as tablespaces. Each tablespace corresponds to one or more data files on the shared storage. Table 7-1 shows the tablespace names used by an Oracle RAC database and the types of data they contain.
Table 7-1 Tablespace Names Used with Oracle Real Application Clusters Databases
Tablespace Name | Contents |
---|---|
Consists of the data dictionary, including definitions of tables, views, and stored procedures needed by the database. Oracle Database automatically maintains information in this tablespace. | |
|
An auxiliary system tablespace that contains the |
Consists of application data. As you create and enter data into tables, Oracle Database fills this space with your data. | |
Contains temporary tables and indexes created during SQL statement processing. You may need to expand this tablespace if you are running a SQL statement that involves significant sorting, such as | |
Contains undo tablespaces for each instance that DBCA creates for automatic undo management. | |
If you do not use automatic undo management, then Oracle Database uses the |
You cannot alter these tablespace names when using the preconfigured database configuration options from OUI. However, you can change the names of the tablespaces if you use the advanced database creation method.
As mentioned, each tablespace has one or more data files. The data file names created by the preconfigured database configuration options vary by storage type such as Oracle ASM, or a file system.
The database is configured with two control files that are stored on shared storage.
Each instance is configured with at least two redo log files that are stored in the shared storage. If you use a file system, then these files are shared file system files. If you use Oracle ASM, then these files are stored on the Oracle ASM disk group.
The file names of the redo log files that are created with the preconfigured database configuration options vary by storage type. Unless you are using a cluster file system, you must enter the block or raw device names.
Oracle Database stores rollback or undo information in undo tablespaces. To manage undo tablespaces, Oracle recommends that you use Automatic Undo Management. Automatic Undo Management is an automated undo tablespace management mode that is easier to administer than manual undo management.
When Oracle ASM and Oracle Managed Files are used along with Automatic Undo Management, an instance that is started for the first time, and thus does not have an undo tablespace, will have its undo tablespace created for it by another instance automatically. The same is also true for redo logs.
See Also: Oracle Real Application Clusters Administration and Deployment Guide for more information about managing undo tablespaces |
Oracle recommends using the server parameter file (SPFILE) for storing Oracle Database initialization parameters. Oracle recommends that you store all SPFILEs on Oracle ASM, including the Oracle ASM SPFILE. SPFILEs must be located on shared storage; all instances in a cluster database can access this parameter file.
See Also: Chapter 5, " Configuring the Server Parameter File in Oracle Real Application Clusters Environments" for more information about the creation and use of parameter files |
Oracle Database installs a baseline database auditing policy. This policy covers the access control configuration information stored in Oracle Database database tables, information stored in Oracle Catalog (rollback segments, tablespaces, and so on), the use of system privileges, and Oracle Label Security configuration.
See Also: Oracle Database Vault Administrator's Guide for more information about the database audit policy |
Oracle RAC databases provide the important benefits of connection load balancing and failover. Standalone Oracle Databases perform load balancing by distributing connections among local dispatchers.
Oracle RAC databases increase load balancing by using the Oracle Clusterware SCANs (Single Client Access Name) to balance connections among all instances in a cluster database.
Review the following sections to understand the Oracle RAC load-balancing features:
Understanding Single Client Access Name (SCAN) and Load-balancing
Understanding Load-Balancing, Failover, and High Availability
The Single Client Access Name (SCAN) is a fully qualified name (hostname+domain) that is configured to resolve to all the addresses allocated for the SCAN. The addresses resolve using Round Robin DNS either on the DNS server, or within the cluster in a GNS configuration. SCAN listeners can run on any node in the cluster.
Oracle Database 11g release 2 and later instances only register with SCAN listeners as remote listeners. Upgraded databases register with SCAN listeners as remote listeners, and also continue to register with all node listeners.
SCANs run on the cluster, not on a cluster node. SCANs provide location independence for the databases, so that client configuration does not have to depend on which nodes are running a particular database. For example, if you configure policy managed server pools in a cluster, then the SCAN enables connections to databases in these server pools regardless of which nodes are allocated to the server pool.
SCANs function like a node virtual IP (VIP) address. However, SCANs are resolved to SCAN VIPs that run on any node in the cluster, so unlike a virtual IP address for a node, clients connecting to the SCAN no longer require updated virtual IP addresses as nodes are added to or removed from the cluster. Because the SCAN addresses resolve to the cluster, rather than to a node address in the cluster, nodes can be added to or removed from the cluster without affecting the SCAN address configuration.
During Oracle Grid Infrastructure installation, SCAN listeners are created for as many IP addresses as there are addresses assigned to resolve to the SCAN. Oracle recommends that the SCAN resolves to three addresses, to provide high availability and scalability. If the SCAN resolves to three addresses, then there are three SCAN listeners created.
Oracle RAC provides failover with the node VIP addresses by configuring multiple listeners on multiple nodes to manage client connection requests for the same database service. If a node fails, then the service connecting to the VIP is relocated transparently to a surviving node, enabling fast notification of the failure to the clients connecting through the VIP. If the application and client are configured with transparent application failover options, then the client is reconnected to the surviving node. Oracle Clusterware routes client application requests from the SCAN to the VIP on the least loaded instance providing the service.
Clients configured to use IP addresses for Oracle Database releases prior to Oracle Database 11g release 2 can continue to use their existing connection addresses; using SCANs is not required. When an earlier version of Oracle Database is upgraded, it registers with the SCAN listeners, and clients can start using the SCAN to connect to that database. The database registers with the SCAN listener through the remote listener parameter in the init.ora file. The REMOTE_LISTENER parameter must be set to SCAN:PORT. Do not set it to a TNSNAMES alias with a single address with the SCAN as HOST=SCAN.
See Also: The Oracle Grid Infrastructure Installation Guide for detailed information about SCANs and their configuration |
Connection load balancing and failover increase availability by taking advantage of the redundant resources within a cluster database.
The Database Agent (previously known as racgimon
) configures the LOCAL_LISTENER
initialization parameters automatically. You do not need to alter this setting. Set the REMOTE_LISTENER
to the SCAN port and SCAN using the SQL command ALTER SYSTEM SET REMOTE_LISTENER
. If you change the SCAN port, then you need to change the REMOTE_LISTENER configuration.
The LOCAL_LISTENER
parameter identifies the local listener, and the REMOTE_LISTENER
parameter identifies the global list of listeners. The REMOTE_LISTENER
parameter is always set to the SCAN address. For example, if the name of the cluster is mycluster, and the domain is example.com, then the SCAN address will be similar to myscan.mycluster.example.com:1521
.
By default, DBCA configures your environment with dedicated servers. However, if you select the Shared server option when using DBCA, then Oracle configures the shared server. In this case, Oracle Database uses both dedicated and shared server processing. When shared servers are configured, the DISPATCHERS
parameter is specified as in the following example:
DISPATCHERS="(protocol=tcp)"
If the DISPATCHERS
initialization parameter does not specify the LISTENER
attribute as in the previous example, then the PMON process registers information for all dispatchers with the listeners specified by the LOCAL_LISTENER
and REMOTE_LISTENER
parameters.
However, when the LISTENER
attribute is specified, the PMON process registers dispatcher information with the listeners specified by the LISTENER
attribute. In this case, setting the LISTENER
attribute overrides REMOTE_LISTENER
settings for the specified dispatchers as in the following example:
DISPATCHERS="(protocol=tcp)(listener=listeners_db_name)"
Note: Where the notationdb_name appears in the previous example and throughout this chapter, it refers to the database name (or Oracle SID) that you entered when prompted by DBCA, or it refers to the entry you made for the DATABASE keyword of the CREATE DATABASE statement. |
See Also: The Oracle Database Net Services Administrator's Guide for further information about cross-instance registration, shared and dedicated server configurations, and connection load balancing |
You can configure two types of listeners in the listener.ora
file as described in the following sections:
If you enable Grid Naming Service (GNS), then you do not have to manually configure the listener. In GNS configurations, there is one GNS daemon for the cluster. Service requests to the cluster domain that GNS manages are routed to the GNS virtual IP address, which routes these requests to the GNS daemon.
The GNS daemon listens for registrations. When a SCAN virtual IP starts, it registers its addresses with GNS. When GNS receives a request from a DNS for the SCAN, it returns the registered addresses to the DNS.
During Oracle Database creation, the LOCAL_LISTENER
parameter is automatically configured to point to the local listener for the database. The Database Agent sets the LOCAL_LISTENER
parameter to a connect descriptor that does not require a TNS alias.
You can set a value manually for LOCAL_LISTENER. However, Oracle recommends that you leave the parameter unset so that the Database Agent can maintain it automatically.
For example, after you create the database, if you need to add a second listener, listening on port 2012, then use the following command to have the database register with both listeners on startup:
> alter system set local_listener='(DESCRIPTION= (ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=192.168.0.61)(PORT=1521)) (ADDRESS=(PROTOCOL=TCP)(HOST=192.168.0.61)(PORT=2011))))' scope=BOTH SID='OCRL1';
See Also: Oracle Database Net Services Administrator's Guide for information about understanding and configuring listeners |
The tnsnames.ora
file is located in the path $ORACLE_HOME/network/admin
. By default, the tnsnames.ora
file is read from the Oracle Grid Infrastructure home, in the path Grid_home/network/admin
.
The listener association no longer requires tnsnames.ora
file entries. The REMOTE_LISTENER is configured by DBCA to reference the SCAN name and port, without any need for a tnsnames.ora
entry.
The following is an example of an Oracle RAC 11g release 2 (11.2) tnsnames.ora
file:
# tnsnames.ora Network Configuration file: /u01/app/oracle/product/11.2.0/dbhome_1/network/admin/tnsnames.ora # Generated by Oracle Configuration tools RACDB = (DESCRIPTION = (ADDRESS = (PROTOCOL = TCP)(HOST = myscan.mycluster.example.com)(PORT = 1521) ) (CONNECT_DATA = (SERVER = DEDICATED) (SERVICE_NAME = RACDB) ) )
If DBCA detects more than one listener on the node, then it displays a list of the listeners. You can select one or all of these listeners with which to register your database.
Note: To administer 11g release 2 (11.2) local and scan listeners using thelsnrctl command, set your ORACLE_HOME environment variable to the path for the Oracle Grid Infrastructure home (Grid home). Do not attempt to use the lsnrctl commands from Oracle home locations for previous releases, because they cannot be used with the new release. |
Services coordinate their sessions by registering their load with the Local Listener and the SCAN Listeners. Clients are redirected by the SCAN Listener to a local listener on the least loaded node that is running the instance for a particular service. That Listener either redirects the client to a Dispatcher process (if the database was configured for Shared Server), or redirects the client to a dedicated server process. Local Listeners are configured to respond to database connection requests, and to non-database connection requests, such as external procedures or XDB requests.
The Listener agent dynamically updates endpoints with the listener. With the Oracle Database 11g release 2 installation, the listener.ora
file now only contains an IPC key and the following information:
(ADDRESS = (PROTOCOL=TCP)(HOST=)(PORT=1521))
This information refers implicitly to the local node's HOST endpoint. The listener.ora
file is the same on every node.
Before you install Oracle RAC, during the Oracle Grid Infrastructure installation, Oracle Net Configuration Assistant creates and starts a default listener in the Grid home called LISTENER
. The listener is configured with default protocol listening addresses. The listener is configured to respond to connection requests that are directed to one protocol address specified during installation.
During Oracle RAC installation, the Oracle RAC database uses the Oracle Clusterware listener, and configures service information about the Oracle RAC database. An Oracle Database 11g release 2 (11.2) database service automatically registers its information with the listener, such as its service name, instance names, and load information. This feature, called service registration, does not require configuration in the listener.ora
file. After listener creation, Oracle Net Configuration Assistant starts the listener.
After installation, you can modify listeners by using srvctl
commands, or by using NetCA. Manual editing of the listener.ora
file is not required.
The following is an example listener.ora
file with an entry for a node named node1
:
LISTENER_SCAN1=(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=IPC)(KEY=LISTENER_ SCAN1)))) # line added by Agent LISTENER_NODE1=(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=IPC) (KEY=LISTENER)))) # line added by Agent # listener.ora.mycluster Network Configuration File: /u01/app/oracle/11.2.0/dbhome_1/network/admin/listener.ora.mycluster # Generated by Oracle configuration tools. LISTENER_NODE1 = (DESCRIPTION_LIST = (DESCRIPTION = (ADDRESS = (PROTOCOL = IPC)(KEY = EXTPROC1521)) ) ) ENABLE_GLOBAL_DYNAMIC_ENDPOINT_LISTENER_NODE1=ON # line added by Agent ENABLE_GLOBAL_DYNAMIC_ENDPOINT_LISTENER_SCAN2=ON # line added by Agent ENABLE_GLOBAL_DYNAMIC_ENDPOINT_LISTENER_SCAN1=ON # line added by Agen
When a listener starts after the Oracle instance starts, and the listener is listed for service registration, registration does not occur until the next time the Oracle Database process monitor (PMON) discovery routine starts. By default, PMON discovery occurs every 60 seconds.
To override the 60-second delay, use the SQL ALTER SYSTEM REGISTER
statement. This statement forces the PMON process to register the service immediately.
Oracle recommends that you create a script to run this statement immediately after starting the listener. If you run this statement while the listener is up and the instance is already registered, or while the listener is down, then the statement has no effect.
See Also: The Oracle Database Net Services Administrator's Guide for further information about the listener and thelistener.ora file |
A tnsnames.ora
file is created on each node with net service names. A connect identifier is an identifier that maps to a connect descriptor. A connect descriptor contains the following information:
The network route to the service, including the location of the listener through a protocol address
The SERVICE_NAME
for an Oracle Database
Note: TheSERVICE_NAME parameter that you use in the tnsnames.ora file is singular, because you can specify only one service name. There is also the SERVICE_NAMES parameter in the database initialization file. The SERVICE_NAMES parameter defaults the global database name, a name comprising the DB_NAME and DB_DOMAIN parameters in the initialization parameter file. |
With Oracle Clusterware 11g release 2 and later, Database Configuration Agent (DBCA) no longer sets the LOCAL_LISTENER parameter. The Oracle Clusterware agent that starts the database sets the LOCAL_LISTENER parameter dynamically, unless you explicitly set the LOCAL_LISTENER parameter, using the command ALTER SYSTEM SET LOCAL_LISTENER to set the LOCAL_LISTENER parameter manually. Listener_alias entries are no longer needed in the tnsnames.ora
file.
For example, after you create the database, if you need to add a second listener, listening on port 2012, then use the following command to have the database register with both listeners on startup:
> alter system set local_listener='(DESCRIPTION= (ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=192.168.0.61)(PORT=1521)) (ADDRESS=(PROTOCOL=TCP)(HOST=192.168.0.61)(PORT=2011))))' scope=BOTH SID='OCRL1';
For the REMOTE_LISTENER parameter, Oracle Clusterware uses the EZ connect syntax scanname:scanport, so that no entries are needed for the REMOTE_LISTENER parameter in the tnsnames.ora
file.
DBCA creates net service names for connections as described in the following sections:
Review the following sections for information about net service names:
Clients that connect to any instance of the database use the SCAN name for the database. This entry also enables Oracle Enterprise Manager to discover an Oracle RAC database.
Three SCAN addresses are configured for the cluster, and allocated to servers. If the chosen address fails, then the connection request to the SCAN name fails over to the next address. Thus, if an instance fails, then clients can still connect using another instance.
In the following example, myscan.mycluster.example.com
is used by the client to connect to the target database, mycluster.example.com
.
mycluster.example.com = (DESCRIPTION = (ADDRESS = (PROTOCOL = TCP)(HOST = host=myscan.mycluster.example.com)(PORT = 1522)) (CONNECT_DATA = (SERVER = DEDICATED) (SERVICE_NAME = myApp) )
An Easy Connect address for this database should be similar to the following:
myscan.mycluster.example.com/mycluster.example.com
Clients that connect to a particular instance of the database use the SCAN name for the instance. This entry, for example, enables Oracle Enterprise Manager to discover the instances in the cluster. These entries are also used to start and stop instances.
In the following example, Oracle Enterprise Manager uses myscan.mycluster.example.com
to connect to an instance named mycluster1
on node1:
mycluster1.example.com= (description= (address=(protocol=tcp)(host=myscan.mycluster.example.com)(port=1521)) (connect_data= (service_name=mycluster.example.com) (instance_name=mycluster1)))
This example is a sample tnsnames.ora
file that is created during a preconfigured database configuration installation.
# tnsnames.ora Network Configuration File: /u01/app/oracle/11.2.0/dbhome_1/network/admin/tnsnames.ora # Generated by Oracle configuration tools. sales = (DESCRIPTION = (ADDRESS = (PROTOCOL = TCP)(HOST = myscan.mycluster.example.com)(PORT = 1522)) (CONNECT_DATA = (SERVER = DEDICATED) (SERVICE_NAME = myApp) )
See Also: The Oracle Database Net Services Administrator's Guide for further information about thetnsnames.ora file |
Oracle Universal Installer starts Oracle Net Configuration Assistant after the database is installed, and creates the Net Services profile, or sqlnet.ora
file.
By default, the sqlnet.ora
file is located in the following directory:
$ORACLE_HOME/network/admin
In this directory, in addition to the default sqlnet.ora
file generated during installation, you also can find a sample sqlnet.ora
file in the directory sample
.
During installation, Oracle Net Configuration Assistant creates the following entries in the sqlnet.ora
file, where Oracle_base is the path to the Oracle RAC installation owner's Oracle base directory:
NAMES.DIRECTORY_PATH=(TNSNAMES, EZCONNECT)
ADR_BASE =Oracle_base
The parameter NAMES.DIRECTORY_PATH
specifies the priority order of the naming methods to use to resolve connect identifiers to connect descriptors.
The Easy Connect naming method eliminates the need for service name lookup in the tnsnames.ora
files for TCP/IP environments. With Easy Connect, clients use a connect string for a simple TCP/IP address, consisting of a host name and optional port and service name. If you use this method, then no naming or directory system is required.
See Also: The Oracle Database Net Services Administrator's Guide for further information about thesqlnet.ora file |
This section describes new features as they pertain to the installation and configuration of Oracle Real Application Clusters (Oracle RAC). The topics in this section are:
Starting with Oracle Database 11g Release 2 (11.2.0.3) you can enter the Proxy Realm information when providing the details for downloading software updates. The proxy realm identifies the security database used for authentication. If you do not have a proxy realm, then you do not have to provide an entry for the Proxy Username, Proxy Password, and Proxy Realm fields. It is case-sensitive.
This proxy realm is for software updates download only.
The following is a list of new features for Release 2 (11.2.0.2):
Starting with the release of the 11.2.0.2 patch set for Oracle Database 11g Release 2, Oracle Database patch sets are full installations of the Oracle Database software. Note the following changes with the new patch set packaging:
Direct upgrades from previous releases (11.x, 10.x) to the most recent patch set are supported.
Out-of-place patch set upgrades, in which you install the patch set into a new, separate Oracle home, are the best practices recommendation. In-place upgrades are supported, but not recommended.
New installations consist of installing the most recent patch set, rather than installing a base release and then upgrading to a patch release.
See Also: My Oracle Support note 1189783.1, "Important Changes to Oracle Database Patch Sets Starting With 11.2.0.2", available from the following URL: |
Use the Software Updates feature to dynamically download and apply software updates as part of the Oracle Database installation. You can also download the updates separately using the downloadUpdates option and later apply them during the installation by providing the location where the updates are present.
Oracle RAC One Node is a single instance of Oracle RAC running on one node in a cluster. You can use Oracle RAC One Node to consolidate many databases onto a single cluster with minimal overhead, and still provide the high availability benefits of failover protection, online rolling patch application, as well as rolling upgrades for the operating system and for Oracle Clusterware. With Oracle RAC One Node you can standardize all Oracle Database deployments across your enterprise.
You can use Oracle Database and Oracle Grid Infrastructure configuration assistants, such as Oracle Database Configuration Assistant (DBCA) and RCONFIG, to configure Oracle RAC One Node databases.
Oracle RAC One Node is a single Oracle RAC database instance. You can use a planned online relocation to start a second Oracle RAC One Node instance temporarily on a new target node, so that you can migrate the current Oracle RAC One Node instance to this new target node. After the migration, the source node instance is shut down. Oracle RAC One Node databases can also fail over to another cluster node within its hosting server pool if their current node fails.
Oracle RAC One Node is not supported if you use a third-party clusterware software, such as Veritas, SFRAC, IBMPowerHA, or HP Serviceguard. Sun Solaris Cluster is not supported at this time.
With Oracle Database release 2 (11.2.0.2), Oracle RAC One Node is supported on all platforms where Oracle Real Application Clusters (Oracle RAC) is certified.
In previous releases, to make use of redundant networks for the interconnect, bonding, trunking, teaming, or similar technology was required. Oracle Grid Infrastructure and Oracle RAC can now make use of redundant network interconnects, without the use of other network technology, to enhance optimal communication in the cluster. This functionality is available starting with Oracle Database 11g Release 2 (11.2.0.2).
Redundant Interconnect Usage enables load-balancing and high availability across multiple (up to four) private networks (also known as interconnects).
The following is a list of new features for Release 2 (11.2)
With Oracle Clusterware 11g release 2 (11.2), Oracle Automatic Storage Management (Oracle ASM) is part of the Oracle Grid Infrastructure installation. In an Oracle Clusterware and Oracle RAC installation, Oracle ASM is installed in the Oracle Clusterware home. In addition, Oracle ASM can be configured to require separate administrative privileges, so that membership in OSDBA may no longer provide administrator access to both the database and the storage tiers.
Oracle Automatic Storage Management Cluster File System (Oracle ACFS) is a new multi-platform, scalable file system and storage management design that extends Oracle Automatic Storage Management (Oracle ASM) technology to support all application data. Oracle ACFS provides dynamic file system resizing, and improved performance using the distribution, balancing and striping technology across all available storage, and provides storage reliability through Oracle ASM's mirroring and parity protection.
Oracle ACFS is available for Linux. It is not available for UNIX platforms at the time of this release.
Cluster Verification Utility (CVU) has the following new features:
CVU can generate shell scripts (Fixup scripts) that perform the system configuration that is required for a successful installation, in addition to identifying system issues that can cause installation failures.
CVU provides additional checks to address install, configuration and operational issues.
CVU is automatically called by OUI to verify prerequisites, and will prompt you to create fixup scripts to correct many system configuration issues that prevent installation.
DBCA no longer sets the value for LOCAL_LISTENER. When Oracle Clusterware starts the database resource, it updates the instance parameters. The LOCAL_LISTENER is set to the virtual IP endpoint of the local node listener address. You should not modify the setting for LOCAL_LISTENER. New installation instances only register with Single Client Access Name (SCAN) listeners as remote listeners. SCANs are virtual IP addresses assigned to the cluster, rather than to individual nodes, so cluster members can be added or removed without requiring updates of clients served by the cluster. Upgraded databases will continue to register with all node listeners, and additionally with the SCAN listeners.
When time zone version files are updated due to Daylight Saving Time changes, TIMESTAMP WITH TIMEZONE (TSTZ) data could become stale. In previous releases, database administrators ran the SQL script utltzuv2.sql
to detect TSTZ data affected by the time zone version changes and then had to carry out extensive manual procedures to update the TSTZ data.
TSTZ data is updated transparently with very minimal manual procedures using newly provided DBMS_DST PL/SQL packages. In addition, there is no longer a need for clients to patch their time zone data files.
See Also: Oracle Database Upgrade Guide for information about preparing to upgrade Timestamp with Time Zone data, Oracle Database Globalization Support Guide for information about how to upgrade the Time Zone file and Timestamp with Time Zone data, and Oracle Call Interface Programmer's Guide for information about performance effects of clients and servers operating with different versions of Time Zone files |
Database Control 11g provides the capability to automatically provision Oracle Clusterware and Oracle RAC installations on new nodes, and then extend the existing Oracle Clusterware and Oracle RAC database to these provisioned nodes. This provisioning procedure requires a successful Oracle RAC installation before you can use this feature.
See Also: Oracle Real Application Clusters Administration and Deployment Guide for information about this feature |
You can use Enterprise Manager Cluster Home page to perform full administrative and monitoring support for High Availability Application and Oracle Clusterware resource management. Such administrative tasks include creating and modifying server pools.
In the past, adding or removing servers in a cluster required extensive manual preparation. Grid Plug and Play reduces the costs of installing, configuring, and managing server nodes by using a grid naming service within the cluster to enable each node to perform the following tasks dynamically:
Negotiating appropriate network identities for itself
Acquiring additional information it needs to operate from a configuration profile
Configuring or reconfiguring itself using profile data, making hostnames and addresses resolvable on the network
As servers perform these tasks dynamically, adding and removing nodes simply requires an administrator to connect the server to the cluster, and to enable the cluster to configure the node. Using Grid Plug and Play, and using best practices recommendations, adding a node to the database cluster is part of the normal server restart, and removing a node from the cluster occurs automatically when a server is turned off.
Oracle configuration assistants provide additional guidance to ensure recommended deployment, and to prevent configuration issues. In addition, configuration assistants validate configurations, and provide scripts to fix issues, which you can choose to use, or reject. If you accept the fix scripts, then configuration issues will be fixed automatically.
Oracle configuration assistants provide the capability of deconfiguring and deinstalling Oracle Real Application Clusters, without requiring additional manual steps.
The Single Client Access Name (SCAN) is the address to provide for all clients connecting to the cluster. The SCAN is a domain name registered to three IP addresses, either in the domain name service (DNS) or the Grid Naming Service (GNS). SCANs eliminate the need to change clients when nodes are added to or removed from the cluster. Clients using SCANs can also access the cluster using Easy Connect.
Opatch now can apply patches in multi-node, multi-patch fashion, and will not start up instances that have a non-rolling patch applied to it, if other instances of the database do not have that patch. Opatch also detects if the database schema is an earlier patch level than the new patch, and it runs SQL commands to bring the schema up to the new patch level.
Note the following changes with this release:
Installing files on raw devices is no longer an option available during installation. You must use a shared file system, or use Oracle Automatic Storage Management. If you are upgrading from a previous release and currently use raw devices, then your existing raw devices can continue to be used. After upgrade is complete, you can migrate to Oracle ASM or to a shared file system if you choose.
The SYSDBA privilege of acting as administrator on the Oracle ASM instance is removed with this release, unless the operating system OSDBA group for the database is the same group that is designated as the OSASM group for Oracle Automatic Storage Management.
If separate Oracle ASM access privileges are enabled, and database administrators are not members of the OSASM group, then database administrators must be a member of the OSDBA for Oracle ASM group to be able to access Oracle ASM files You can designate the OSDBA group for the Oracle RAC database as the OSDBA group for the Oracle ASM instance.
The following is a list of new features for Oracle RAC 11g release 1 (11.1):
Note: Some features in this list have been superseded by changes in the 11.2 release, particularly those listed for Oracle ASM. |
With Oracle Database 11g release 1, Oracle Clusterware can be installed or configured as an independent product, and additional documentation is provided on storage administration. For installation planning, note the following documentation:
This book provides an overview and examples of the procedures to install and configure a two-node Oracle Clusterware and Oracle RAC environment.
This platform-specific book provides procedures either to install Oracle Clusterware as a standalone product, or to install Oracle Clusterware with either Oracle Database, or Oracle RAC. It contains system configuration instructions that require system administrator privileges.
This book (the guide that you are reading) provides procedures to install Oracle RAC after you have completed successfully an Oracle Clusterware installation. It contains database configuration instructions for database administrators.
This book provides information for database and storage administrators who administer and manage storage, or who configure and administer Oracle Automatic Storage Management (Oracle ASM).
This is the administrator's reference for Oracle Clusterware. It contains information about administrative tasks, including those that involve changes to operating system configurations, and cloning Oracle Clusterware.
This is the administrator's reference for Oracle RAC. It contains information about administrative tasks. These tasks include database cloning, node addition and deletion, Oracle Cluster Registry (OCR) administration, use of SRVCTL and other database administration utilities, and tuning changes to operating system configurations.
The following are installation option changes for Oracle Database 11g:
Oracle Application Express: This feature is installed with Oracle Database 11g. It was previously named HTML DB, and was available as a separate Companion CD component.
Oracle Configuration Manager: Oracle Configuration Manager (OCM) is integrated with Oracle Universal Installer. However, it is an optional component with database and client installations. Oracle Configuration Manager, used in previous releases as Customer Configuration Repository (CCR), is a tool that gathers and stores details relating to the configuration of the software stored in the Oracle ASM and database Oracle home directories.
Oracle Data Mining: The Enterprise Edition installation type selects Oracle Database Mining option by default. In Oracle Database 11g, the Data Mining metadata is created with SYS
metadata when you select the Create Database option.
Oracle Database: This feature is integrated with Oracle Universal Installer. However, it is an optional component for database installation.
Oracle SQL Developer: This feature is installed by default with template-based database installations, such as General Purpose, Transaction Processing, and Data Warehousing. It is also installed with database client Administrator, Runtime, and Advanced installations.
Oracle Warehouse Builder: This information integration tool is now installed with both Standard and Enterprise Edition versions of the Oracle Database. With Enterprise Edition, you can purchase additional extension processes. Installing the Oracle Database also installs a pre-seeded repository, OWBSYS
, necessary for using Oracle Warehouse Builder.
Oracle Ultra Search: Starting with Oracle Database 11g, Oracle Ultra Search is integrated with Oracle Database. In the previous releases, this product was available as a Companion CD component.
Oracle XML DB: Starting with Oracle Database 11g, Oracle XML DB is no longer an optional feature. It is installed and configured using Database Configuration Assistant for all database installations.
The following are the new components available while installing Oracle Database 11g:
Oracle Application Express: Starting Oracle Database 11g, HTML DB is no longer available as a Companion CD component. Renamed as Oracle Application Express, this component is installed with Oracle Database 11g.
With Oracle Database 11g, Oracle Application Express replaces iSQL*Plus.
See Also: Oracle Application Express User's Guide for more information about Oracle Application Express |
Oracle Configuration Manager: This feature is offered during Advanced installation. It was previously named Customer Configuration repository (CCR). It is an optional component for database and client installations. Oracle Configuration Manager gathers and stores details relating to the configuration of the software stored in database Oracle home directories.
Oracle Database Vault: This feature is an optional component for database installation.
Oracle Real Application Testing: This feature is installed by default with the Enterprise Edition installation type of Oracle Database 11g.
See Also: Oracle Database Real Application Testing User's Guide for more information about Oracle Real Application Testing |
Oracle SQL Developer: This feature is installed by default with template-based database installations, such as General Purpose, Transaction Processing, and Data Warehousing. It is also installed with database client Administrator, Runtime, and Advanced installations.
Oracle Warehouse Builder: This feature is now included as an option in the Database installation.
Note: With Standard Edition and Enterprise Edition versions of Oracle Database 11g release 2 (11.2), Oracle Warehouse Builder with basic features is installed. However, with Enterprise Edition, you can purchase options that extend Oracle Warehouse Builder. |
See Also: Oracle Warehouse Builder Sources and Targets Guide for more information about Oracle Warehouse Builder |
The following is a list of enhancements and new features for Oracle Database 11g release 2 (11.2):
The Automatic Diagnostic Repository is a feature added to Oracle Database 11g. The main objective of this feature is to reduce the time required to resolve bugs. Automatic Diagnostic Repository is the layer of the Diagnostic Framework implemented in Oracle Database 11g that stores diagnostic data and also provides service APIs to access data. The default directory that stores the diagnostic data is $ORACLE_BASE/diag
.
The Automatic Diagnostic Repository implements the following:
Diagnostic data for all Oracle products is written into an on-disk repository.
Interfaces that provide easy navigation of the repository, and the capability to read and write data
For Oracle RAC installations, if you use a shared Oracle Database home, then the Automatic Data Repository must be located on a shared storage location that is available to all the nodes.
Oracle Clusterware continues to store diagnostic data in the directory Grid_home
/log
, where Grid_home
is the Oracle Clusterware home.
Oracle ASM fast mirror resync quickly resynchronizes Oracle ASM disks within a disk group after transient disk path failures as long as the disk drive media is not corrupted. Any failures that render a failure group temporarily unavailable are considered transient failures. Disk path malfunctions, such as cable disconnections, host bus adapter or controller failures, or disk power supply interruptions, can cause transient failures. The duration of a fast mirror resync depends on the duration of the outage. The duration of a resynchronization is typically much shorter than the amount of time required to completely rebuild an entire Oracle ASM disk group.
ASM Configuration Assistant (ASMCA) is a new configuration tool that can run from the Oracle Grid Infrastructure for a cluster home. ASMCA configures ASM instances, diskgroups, volumes, and file systems. ASMCA is run during installation, and can be used as an administration configuration tool, like DBCA.
Database Configuration Assistant (DBCA), Database Upgrade Assistant (DBUA), and Oracle Net Configuration Assistant (NETCA) have been improved. These improvements include the following:
7IDBCA is enhanced with the following feature:
Provides the option to switch from a database configured for Oracle Enterprise Manager Database Control to Oracle Enterprise Manager Grid Control.
DBUA is enhanced with the following features:
Includes an improved pre-upgrade script to provide space estimation, initialization parameters, statistics gathering, and new warnings. DBUA also provides upgrades from Oracle Database releases 9.0, 9.2, 10.1, and 10.2.
Update for Oracle Database 11g release 11.2: Also provides upgrades for 11.1.
Only out-of-place upgrades are supported.
Starts any services running prior to upgrades
Includes a deinstallation tool (deinstall
), which is available in the installation media before installation, and is available in Oracle home directories after installation. It is located in the path $ORACLE_HOME/deinstall. The script stops Oracle software, and removes Oracle software and configuration files on the operating system.
This feature introduces a new SYSASM
privilege that is specifically intended for performing Oracle ASM administration tasks. Using the SYSASM
privilege instead of the SYSDBA
privilege provides a clearer division of responsibility between Oracle ASM administration and database administration.
OSASM is a new operating system group that is used exclusively for Oracle ASM. Members of the OSASM group can connect as SYSASM using operating system authentication and have full access to Oracle ASM.
In previous releases, Oracle ASM used the disk with the primary copy of a mirrored extent as the preferred disk for data reads. With this release, using the new initialization file parameter asm_preferred_read_failure_groups
, you can specify disks located near a specific cluster node as the preferred disks from which that node obtains mirrored data. This option is presented in Database Configuration Assistant (DBCA), and can be configured after installation. This change facilitates faster processing of data with widely distributed shared storage systems or with extended clusters (clusters whose nodes are geographically dispersed), and improves disaster recovery preparedness.
Rolling migration for Oracle ASM enables you to upgrade or patch Oracle ASM instances on clustered Oracle ASM nodes without affecting database availability. Rolling migration provides greater availability and more graceful migration of Oracle ASM software from one release to the next. This feature applies to Oracle ASM configurations that run on Oracle Database 11g release 1 (11.1) and later. In other words, you must already have Oracle Database 11g release 1 (11.1) installed before you can perform rolling migrations.
Note: You cannot change the owner of the Oracle ASM or Oracle Database home during an upgrade. You must use the same Oracle software owner that owns the existing Oracle ASM or Oracle Database home. |
In Oracle Database 11g, the data mining schema is created when you run the SQL script catproc.sql
as the SYS user. Therefore, the data mining option is removed from the Database Features screen of Database Configuration Assistant.
Oracle Disk Manager (ODM) can manage network file systems (NFS) on its own, without using the operating system kernel NFS driver. This is referred to as Direct NFS. Direct NFS implements NFS version 3 protocol within the Oracle Database kernel. This change enables monitoring of NFS status using the ODM interface. The Oracle Database kernel driver tunes itself to obtain optimal use of available resources.
This feature provides the following:
Ease of tuning, and diagnosability, by providing the Oracle kernel control over the input-output paths to Network File Server, and avoiding the need to tune network performance at the operating system level.
A highly stable, highly optimized NFS client for database operations.
Use of the Oracle network file system layer for user tasks, reserving the operating system kernel network file system layer for network communication.
Use of the Oracle buffer cache, rather than the file system cache, for simpler tuning.
A common, consistent NFS interface, capable for use across Linux, UNIX and Windows platforms.
Oracle RAC aware NFS performance. With operating system NFS drivers, even though NFS is a shared file system, NFS drives have to be mounted with the option noac
(NO Attribute Caching) to prevent the operating system NFS driver from optimizing the filesystem cache by storing file attributes locally. ODM NFS automatically recognizes Oracle RAC instances, and performs appropriate operations for datafiles without requiring additional reconfiguration from users, system administrators, or DBAs. If you store the Oracle Clusterware voting disks or Oracle Cluster Registry (OCR) files on NFS, then you must continue to require mounting voting disks with the noac
option.
With the development of Stripe and Mirror Everything architecture (SAME), and improved storage and throughput capacity for storage devices, the original OFA mission to enhance performance has shifted to its role of providing well-organized Oracle installations with separated software, configuration files and data. This separation enhances security, and simplifies upgrade, cloning, and other administrative tasks.
Oracle Database 11g release 2 (11.2) incorporates several changes to OFA to address this changed purpose.
As part of this change:
During Oracle RAC installation, you are prompted to accept the default, or select a location for the Oracle base directory, instead of the Oracle home directory. This change facilitates installation of more than one Oracle home directory in a common location, and separates software units for simplified administration.
With this release, as part of the implementation of Automatic Diagnostic Repository (ADR), the following admin
directories are changed:
bdump
(location set by the background_dump_dest
initialization parameter; storage of Oracle background process trace files)
cdump
(location set by the core_dump_dest
initialization parameter; storage of Oracle core dump files)
udump
(location set by the user_dump_dest
initialization parameter; storage of Oracle user SQL trace files)
By default, the location of these trace and core files is in the /diag
directory, which is in the path $ORACLE_BASE/diag
.
The initialization parameters BACKGROUND_DUMP_DEST
and USER_DUMP_DEST
are deprecated. They continue to be set, but you should not set these parameters manually.
A new initialization parameter is introduced. DIAGNOSTIC_DEST
contains the location of "ADR base", which is the base directory under which one or more Automatic Database Repository homes are kept. Oracle documentation commonly refers to these homes as ADR homes. Each database instance has an ADR home, which is the root directory for a number of other directories that contain trace files, the alert log, health monitor reports, and dumps for critical errors. You can also view alert and trace files with the SQL statement select name, value from v$diag_info
.
The default Fast Recovery Area (formerly known as Flash Recovery Area) is moved from $ORACLE_HOME/../recovery_area
to $ORACLE_BASE/recovery_area
.
The default data file location is moved from $ORACLE_HOME/../oradata
to $ORACLE_BASE/oradata
.
A new utility, The ADR Command Interpreter (ADRCI), is introduced. The ADRCI facilitates reviewing alert log and trace files.
For Oracle RAC installations, Oracle requires that the Fast Recovery Area and the data file location are on a location shared among all the nodes. The Oracle Universal Installer confirms that this is the case during installation.
This change does not affect the location of trace files for Oracle Clusterware.
See Also: Oracle Database Administrator's Guide for detailed information about these changes, and Oracle Database Utilities for information about viewing alert log and list trace files with ADRCI |
During installation, you are asked if you want to install Oracle Configuration Manager (OCM). OCM is an optional tool that enables you to associate your configuration information with your My Oracle Support account (formerly OracleMetalink). This can facilitate handling of service requests by ensuring that server system information is readily available.
Configuring the OCM tool requires that you have the following information from your service agreement:
My Oracle Support e-mail address/username
Password
In addition, you are prompted for server proxy information, if the host system does not have a direct connection to the Internet.
Large data file support is an automated feature that enables Oracle to support larger files on Oracle ASM more efficiently, and to increase the maximum file size.
In previous releases, Database Configuration Assistant contains the functionality to configure databases while creating them either with Database Control or with Grid Control, or to reconfigure databases after creation. However, if you want to change the configuration from Database to Grid control, this requires significant work. With Oracle Database 11g, Database Configuration Assistant enables you to switch configuration of a database from Database Control to Grid Control by running the Oracle Enterprise Manager Configuration Plug-in.
The following components that were part of Oracle Database 10g release 2 (10.2) are not available for installation with Oracle Database 11g:
iSQL*Plus
Oracle Workflow
Data Mining Scoring Engine
Oracle Enterprise Manager Java console
This appendix describes how to stop all processes in an Oracle Real Application Clusters (Oracle RAC) database, in preparation for adding additional products to an existing database, or in preparation for patch updates. This appendix contains the following topics:
Before you make any changes to the Oracle software, Oracle recommends that you create a backup of the Oracle Database installation.
Shut down any existing Oracle Database instances on each node, with normal or immediate priority. You can use Oracle Enterprise Manager or SRVCTL to shut down the Oracle RAC databases and their instances.
Note: If you are using Oracle Clusterware or Cluster Ready Services, then you must shut down all Oracle Database instances on all of the cluster nodes before modifying the Oracle software. If you are performing a patch update, then review the instructions in the Patch Set Notes for detailed instructions. |
Stop all listener and other processes that are running in the Oracle home directories where you want to modify the database software.
Note: Before you can shut down Oracle ASM instances, Oracle Clusterware, or Oracle Restart, you must first shut down all database instances that use them. |
This section contains the following topics:
This section provides an overview of what needs to be done before adding additional products to Oracle Database 11g release 2 (11.2). If you are performing a patch upgrade, then refer to the Database Patch Set Notes for the patch for additional instructions.
Note: You must perform these steps in the order listed. |
Shut down any processes in the Oracle home on each node that can access a database.
Note: Before you shut down any processes that are monitored by Enterprise Manager Grid Control, set a blackout in Grid Control for the processes that you intend to shut down. This is necessary so that the availability records for these processes indicate that the shutdown was planned downtime, rather than an unplanned system outage. |
For example:
$ORACLE_HOME/bin/emctl stop dbconsole
Shut down all Oracle RAC instances on all nodes. To shut down all Oracle RAC instances for a database, enter the following command, where db_name
is the name of the database:
$ oracle_home/bin/srvctl stop database -d db_name
If you are modifying an Oracle Clusterware or Oracle Cluster Ready Services (CRS) installation, then shut down the following Oracle Database services.
Note: You must perform these steps in the order listed. |
Shut down any processes in the Oracle home on each node that might be accessing a database; for example, Oracle Enterprise Manager Database Control.
Note: Before you shut down any processes that are monitored by Oracle Enterprise Manager Grid Control, set a blackout in Grid Control for the processes that you intend to shut down. This is necessary so that the availability records for these processes indicate that the shutdown was planned downtime, rather than an unplanned system outage. |
Shut down all Oracle RAC instances on all nodes. To shut down all Oracle RAC instances for a database, enter the following command, where db_name
is the name of the database:
$ oracle_home/bin/srvctl stop database -d db_name
Stop all node applications on all nodes. To stop node applications running on a node, enter the following command, where u01/app/grid/11.2.0
is the Oracle Grid Infrastructure home, and node
is the name of the node where the applications are running
$ u01/app/grid/11.2.0/bin/srvctl stop nodeapps -n node
Shut down the Oracle Clusterware or CRS process by entering the following command on all nodes as the root
user
# crsctl stop cluster
To stop Oracle Clusterware and Oracle ASM instances on all nodes in the cluster, you can enter a single command: crsctl stop cluster -all
.
Review the contents in this section to validate that your Oracle RAC cluster is ready for upgrades.
Use the command cluvfy stage -predbinst -upgrade
to check the readiness of your Oracle RAC installation for upgrades. Running cluvfy
with the -predbinst -upgrade
flags performs system checks to confirm if the cluster is in a correct state for upgrading from an existing Oracle RAC installation.
The command uses the following syntax, where variable content is indicated by italics:
cluvfy stage -pre dbinst -upgrade -src_dbhome src_RAChome -dest_dbhome dest_RAChome -dest_version dest_version [-fixup[-fixupdirpath]] [-verbose]
The options are:
-src_dbhome
src_RAChome
Use this flag to indicate the location of the source Oracle RAC home that you are upgrading, where src_RAChome
is the path to the home that you want to upgrade.
-dest_dbhome
dest_RAChome
Use this flag to indicate the location of the upgrade Oracle RAC home, where dest_RAChome
is the path to the Oracle RAC home.
-dest_version
dest_version
Use the dest_version flag to indicate the release number of the upgrade, including any patchset. The release number must include the five digits designating the release to the level of the platform-specific patch. For example: 11.2.0.2.0.
-fixup
[-fixupdir
fixupdirpath
]]
Use the -fixup
flag to indicate that you want to generate instructions for any required steps you need to complete to ensure that your cluster is ready for an upgrade. The default location is the CVU work directory. If you want to place the fixup instructions in a different directory, then add the flag -fixupdir, and provide the path to the fixup directory
-verbose
Use the -verbose
flag to produce detailed output of individual checks
You can verify that the permissions required for installing Oracle RAC have been configured on the nodes node1
and node2
by running the following command:
$ ./cluvfy stage -pre dbinst -upgrade -src_dbhome /u01/app/oracle/product/11.1.0/dbhome_1 -dest_dbhome /u01/app/oracle/product/11.2.0/dbhome_1 -dest_version 11.2.0.3.0 -fixup -fixupdirpath /home/oracle/fixup -verbose
Use Cluster Verification Utility to assist you with system checks in preparation for starting a database upgrade. The installer runs the appropriate CVU checks automatically, and either prompts you to fix problems, or provides a fixup script to be run on all nodes in the cluster before proceeding with the upgrade.
This chapter describes how to use Database Configuration Assistant (DBCA) in standalone mode to create and delete Oracle Real Application Clusters (Oracle RAC) databases. The topics in this chapter include the following:
DBCA has the following primary database functions:
Create and delete databases
Add and delete database instances
Set up network configurations for databases and their instances
Register databases in Oracle Enterprise Manager Grid Control or configure Database Control
Configure database options, such as Oracle Database Vault
Start up databases and their instances
Note: Cluster Managed Services are no longer managed through DBCA. Instead, use the cluster managed services page in Oracle Enterprise Manager DB Control (accessible from the Cluster Database Availability Page). For more information, refer to the Oracle Real Application Clusters Administration and Deployment Guide. |
See Also:
|
Oracle recommends that you use DBCA to create your Oracle RAC database, because preconfigured databases optimize your environment for features such as the server parameter file (SPFILE), and automatic undo management. If you use Oracle ASM or cluster file system storage, then DBCA also configures automated backup, which uses the Fast Recovery Area.
DBCA enables you to create both policy-managed and administrator-managed databases. With DBCA, you can also create site-specific tablespaces as part of database creation. If you have data file requirements that differ from those offered by DBCA templates, then create your database with DBCA and modify the data files later. You can also run user-specified scripts as part of your database creation process.
DBCA also configures your Oracle RAC environment for various Oracle high availability features, such as cluster administration tools. DBCA also starts any database instances required to support your defined configuration.
If your system has an Oracle Database 10g or 11g installation, and you install Oracle Database 11g release 2 (11.2) either to coexist with or to upgrade the Oracle Database 10.1, 10.2, or 11.1 installation, then most installation types automatically migrate the existing Oracle Database listener to the 11g release 2 (11.2) Oracle home. During migration, they configure and start a default Oracle Net listener using the same TCP/IP port as the existing listener, with the IPC key value.
During the Oracle Clusterware upgrade, the default listener (LISTENER_NODENAME was migrated to the Oracle Grid Infrastructure home (Grid home). DBCA always uses the default listener.
The listener migration process stops the listener in the existing Oracle home, and restarts it in the new Oracle home. If the database was using the default listener (LISTENER_NODENAME), then it is migrated automatically to the Oracle Clusterware home by NETCA as part of the Oracle Clusterware upgrade. If the database was using a nondefault listener, then DBUA migrates the nondefault listener to the Oracle Database home.
During migration, client applications may not be able to connect to any databases that are registered to the listener that is being migrated.
In previous releases of Oracle Database, you were required to set environment variables for ORACLE_HOME and ORACLE_SID to start, stop, and check the status of Enterprise Manager. With Oracle Database 11g release 2 (11.2) and later, you need to set the environment variables ORACLE_HOME and ORACLE_UNQNAME to use Enterprise Manager. For example, on each node, enter commands similar to the following to set these values for the oracle
user with the Bourne shell as the default shell, where the Oracle home is /u01/app/oracle
, and where the database unique name is sales
:
$ export ORACLE_HOME=/u01/app/oracle/product/11.2/dbhome1 $ export ORACLE_UNQNAME=sales
Place these environment variables in the oracle
user profile file on each cluster member node to ensure that the environment variable settings are available after system restarts.
To use DBCA to change the database configuration, use Cluster Verification Utility (CVU) to verify that your system is prepared for configuration changes using the following command syntax:
/Grid_home/bin/cluvfy stage -pre dbcfg -fixup -n node_list -d Oracle_home [-verbose]
In the preceding syntax example, the variable Grid_home
is the Oracle Grid Infrastructure home, the variable node_list
is the list of nodes in your cluster, separated by commas, and the variable Oracle_home
is the path for the Oracle home directory where OUI creates or modifies the database. The -fixup
flag generates a fixup script that can be run as root to resolve many operating system configuration tasks if they were not completed before you run the check.
For example, to perform a check to determine if your system is prepared for an Oracle Database with Oracle RAC installation on a two-node cluster with nodes node1 and node2, with the Oracle Grid Infrastructure home path/u01/app/grid/11.2.0
, and with the Oracle home path /u01/app/oracle/product/11/db1
, enter the following command:
$ /u01/app/grid/11.2.0/bin/cluvfy stage -pre dbcfg -fixup -n node1,node2 -d\ /u01/app/oracle/product/11/db1
You can select the option -verbose
to receive progress updates as the CVU performs its system checks, and detailed reporting of the test results.
If the CVU summary indicates that the cluster verification check fails, and you cannot resolve these issues by running the fixup script, then review and correct the relevant system configuration steps, and run the test again.
The command cluvfy.sh stage -pre dbcfg
verifies the following:
Node Reachability: All the specified nodes are reachable from the local node.
User Equivalence: User equivalence exists on all the specified nodes.
Node Connectivity: Connectivity exists between all the specified nodes through the available public and private network interfaces.
Administrative Privileges: The oracle
user has proper administrative privileges on the specified nodes for creating an Oracle RAC database.
Oracle Clusterware Integrity: All the components of the Oracle Clusterware stack are fully operational.
To create a database with DBCA in standalone mode without Oracle ASM or a cluster file system, you must have configured shared storage devices. In addition, you must have run the Oracle Net Configuration Assistant (NETCA) to configure your Oracle Net listener.ora
file. To run DBCA, you no longer need to set operating system environment variables ORACLE_HOME to the Oracle RAC database home, or ORACLE_UNQNAME to the database unique name.
You can use DBCA to create a database from templates supplied by Oracle, or from templates that you create. The templates contain settings optimized for a particular type of workload.
Oracle ships templates for the following two workload types:
General purpose or transaction processing
Data warehouse
For more complex environments, you can select the Custom Database option. This option does not use templates and results in a more extensive interview, which means that it takes longer to create your database.
Click Show Details to see the configuration for each type of database. Select the template suited to the type of workload your database will support. If you are not sure which to choose, then select the default General Purpose or Transaction Processing template.
To start DBCA, connect as the installation owner account (for example, oracle
) to one of your nodes where Oracle RAC is installed, load SSH keys into memory, and enter the command dbca
command from the $ORACLE_HOME/bin
directory.
When you start DBCA, the first page it displays is the Welcome page for Oracle RAC, which includes the option to select an Oracle RAC database. DBCA displays this Oracle RAC Welcome page only if the Oracle home from which it is started was installed on a cluster.
If the Oracle RAC Welcome page opens, then provide information as prompted by DBCA. Click Help if you need assistance.
If DBCA does not display the Welcome page for Oracle RAC, then DBCA was unable to detect if the Oracle home is installed on a cluster. In this case, check that the OUI inventory is correctly located in the directory /etc/oraInst.loc
, and that the oraInventory
file is not corrupted. Also, perform clusterware diagnostics by using the following CVU command syntax:
/Grid_home/bin/cluvfy/cluvfy.sh stage -post crsinst -n nodelist
.
For example, with the mountpoint /u01/app/grid/11.2.0
, and nodes node1
and node2
, run the following command:
$ /u01/app/grid/11.2.0/bin/cluvfy stage -post crsinst -n node1,node2
Note the following important information when using DBCA:
If nodes that are part of your cluster installation do not appear on the Node Selection page, then run Opatch lsinventory to perform inventory diagnostics and CVU to perform clusterware diagnostics.
The global database name can be up to 30 characters in length, and must begin with an alphabetic character. The domain portion of the global database name can be no more than 128 characters and can contain only alphabetic and numeric characters, as well as the period (.) character.
The SID
prefix must begin with an alphabetic character.
The maximum number of characters you can use for the SID prefix is 8 characters. DBCA uses the SID prefix to generate a unique value for the variable ORACLE_SID
for each instance.
On the Management Options page, you are provided with selections for Oracle Enterprise Manager monitoring and management interfaces.
Database Control is always an option. If DBCA discovers Grid Control agents on the cluster, then it also provides you with the option Enterprise Manager with the Grid Control. Choose between Database Control and Grid Control interfaces.
You can enable daily backup operations. For daily backups, you enter the backup time and operating system credentials for the user that performs backup operations.
You can no longer set up e-mail notification either from DBCA or OUI with Oracle RAC 11g release 11.2.0.2 and later.
Note that with Oracle RAC 11g release 2 and later, the storage administration role, SYSASM, requires membership in the OSASM group. This group may not be the same as OSDBA, whose members are granted the SYSDBA privilege. Tasks that require the SYSASM privilege in this list are marked with SYSASM Privilege.
SYSASM Privilege: To use a Fast Recovery Area, Oracle recommends that you create at least two separate Oracle ASM disk groups: one for the database area and one for the recovery area. Oracle recommends that you place the database area and the recovery area in separate failure groups. If you have divided administrative privileges, then ask your storage administrator to set up the Fast Recovery Area.
A failure group is defined by shared hardware, such as a controller shared between two storage devices, or two disks that are on the same spindle. If two storage devices share hardware that could fail, making both storage devices unavailable, then these storage devices are said to be in the same failure group.
If you do not use Oracle ASM, then Oracle recommends that the data files and the Fast Recovery Area are located outside of the Oracle home, in separate locations, as with separate Oracle ASM failure groups, so that a hardware failure does not affect availability.
See Also: Oracle Database Concepts for more information about using a Fast Recovery Area, and Oracle Automatic Storage Management Administrator's Guide for information about failure groups and best practices for high availability and recovery |
SYSASM Privilege: On the Oracle ASM Disk Groups page, if you do not see the storage devices that you want to add, then click Change Disk Discovery Path to alter the search path used by DBCA to find available storage devices. You can select storage devices with a status of Candidate
or Former
(never used in an Oracle ASM disk group or no longer in a group) by selecting the box. If you want to add storage devices that still have Oracle ASM disk headers, but the disk group is no longer in use (a case that can occur if you are selecting storage devices after an aborted install, you attempted deinstallation without dropping the disk group, or other configuration problems), then use the Force
command.
On the Recovery Configuration page, if you are using Oracle ASM or cluster file system storage, then you can also select the Fast Recovery Area and size on the Recovery Configuration page. If you are using Oracle ASM, then the Fast Recovery Area defaults to the Oracle ASM Disk Group. If you are using OCFS2, then the Fast Recovery Area defaults to $ORACLE_BASE/flash_recovery_area
.
On the Initialization Parameters page, if you are installing on a Linux system, note that Memory Size (SGA and PGA), which sets the initialization parameter MEMORY_TARGET
or MEMORY_MAX_TARGET
, cannot be greater than the shared memory filesystem on your operating system.
For example, if the shared memory filesystem allocation on your system is 1 GB, but you set Memory Size (MEMORY_TARGET
) to 2 GB, then you receive the following error during database startup:
ORA-00845: MEMORY_TARGET not supported on this system ORA-01078: Failure in processing system parameters
This issue is not relevant for other platforms.
For all platforms, if you intend to add more nodes in your cluster than you have during the current DBCA session, then click All Initialization Parameters, and change the parameter CLUSTER_DATABASE_INSTANCES
to the total number of nodes that you plan to add to the cluster.
In addition, if you click All Initialization Parameters, note that if your global database name is longer than 8 characters, then the database name value (in the DB_NAME
parameter) is truncated to the first 8 characters, and the DB_UNIQUE_NAME
parameter value is set to the global name.
See Also: Oracle Database Administrator's Reference for Linux and UNIX-Based Operating Systems for information about initialization parameters |
After you respond to DBCA prompts, review the Summary dialog information and click OK, DBCA does the following:
Creates an operative Oracle RAC database and its instances
Creates the Oracle RAC data dictionary views
Configures the network for the cluster database
Migrates previous release Oracle Database listeners and related files to the Oracle Database 11g Oracle home
Starts the listeners and database instances, and then starts the high availability services
Configures Oracle Enterprise Manager Database Control or Grid Control
Caution: After you have created the database, if you decide that you want to install additional Oracle Database products in the database you have created, then you must stop all processes running in the Oracle home before you attempt to install the additional products, so that Oracle Universal Installer can relink certain executables and libraries. Refer to Appendix E, "How to Stop Processes in an Existing Oracle Real Application Clusters Database" for additional information. |
This section explains how to delete an Oracle RAC database with DBCA. This process deletes a database and removes a database's initialization parameter files, instances, OFA structure, and an Oracle network configuration. However, this process does not remove data files if you placed the files on raw devices or on raw partitions.
To delete a database with DBCA:
Start DBCA on one of the nodes:.
Run the dbca
command from the $ORACLE_HOME/bin
directory.
The DBCA Welcome page appears.
Select Oracle Real Application Clusters, and click Next.
Select Delete a database, and click Next. DBCA displays the List of Cluster Databases page.
If your user ID and password are not operating-system authenticated, then the List of Cluster Databases page displays the user name and password fields. If these fields appear, then enter a user ID and password for a user account that has SYSDBA
privileges.
Select the database to delete, and click Finish.
After you click Finish, DBCA displays a dialog box to confirm the database and instances that DBCA is going to delete.
Click OK to begin the deletion of the database and its associated files, services, and environment settings, or click Cancel to stop the operation.
When you click OK, DBCA continues the operation and deletes all the associated instances for this database. DBCA also removes the parameter files, password files, and oratab
entries.
At this point, you have accomplished the following:
Deleted the selected database from the cluster
Deleted high availability services that were assigned to the database
Deleted the Oracle Net configuration for the database
Deconfigured Database Control for that database
Deleted the OFA directory structure for that database from the cluster
Deleted the data files if the data files were not on raw devices
Installation Guide
11g Release 2 (11.2) for Linux and UNIX
E24660-03
May 2012
Oracle Real Application Clusters Installation Guide, 11g Release 2 (11.2) for Linux and UNIX
E24660-03
Copyright © 2007, 2012, Oracle and/or its affiliates. All rights reserved.
Primary Author: Douglas Williams
Contributing Authors: Mark Bauer, Jonathan Creighton, Barb Lundhild, Venkat Maddali, Markus Michalewicz, Dipak Saggi, Shachi Sanklecha, Janet Stern
Contributors: David Austin, Troy Anthony, Mark Bauer, Mark Fuller, Yingwei Hu, Prakash Jashnani, Caroline Johnston, Aneesh Khandelwal, Sana Karam, Ram Kumar, Bob Nelson, Satish Panchumarthy, Sunil Ravindrachar, Trivikrama Samudrala, Janelle Simmons, Vivian Schupmann, Dharma Sirnapalli, James Spiller, Malai Stalin, Kamal Tbeileh
This software and related documentation are provided under a license agreement containing restrictions on use and disclosure and are protected by intellectual property laws. Except as expressly permitted in your license agreement or allowed by law, you may not use, copy, reproduce, translate, broadcast, modify, license, transmit, distribute, exhibit, perform, publish, or display any part, in any form, or by any means. Reverse engineering, disassembly, or decompilation of this software, unless required by law for interoperability, is prohibited.
The information contained herein is subject to change without notice and is not warranted to be error-free. If you find any errors, please report them to us in writing.
If this is software or related documentation that is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, the following notice is applicable:
U.S. GOVERNMENT RIGHTS Programs, software, databases, and related documentation and technical data delivered to U.S. Government customers are "commercial computer software" or "commercial technical data" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, the use, duplication, disclosure, modification, and adaptation shall be subject to the restrictions and license terms set forth in the applicable Government contract, and, to the extent applicable by the terms of the Government contract, the additional rights set forth in FAR 52.227-19, Commercial Computer Software License (December 2007). Oracle America, Inc., 500 Oracle Parkway, Redwood City, CA 94065.
This software or hardware is developed for general use in a variety of information management applications. It is not developed or intended for use in any inherently dangerous applications, including applications that may create a risk of personal injury. If you use this software or hardware in dangerous applications, then you shall be responsible to take all appropriate fail-safe, backup, redundancy, and other measures to ensure its safe use. Oracle Corporation and its affiliates disclaim any liability for any damages caused by use of this software or hardware in dangerous applications.
Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners.
Intel and Intel Xeon are trademarks or registered trademarks of Intel Corporation. All SPARC trademarks are used under license and are trademarks or registered trademarks of SPARC International, Inc. AMD, Opteron, the AMD logo, and the AMD Opteron logo are trademarks or registered trademarks of Advanced Micro Devices. UNIX is a registered trademark of The Open Group.
This software or hardware and documentation may provide access to or information on content, products, and services from third parties. Oracle Corporation and its affiliates are not responsible for and expressly disclaim all warranties of any kind with respect to third-party content, products, and services. Oracle Corporation and its affiliates will not be responsible for any loss, costs, or damages incurred due to your access to or use of third-party content, products, or services.
This appendix lists the default port numbers and describes how to change the assigned port after installation. This appendix contains the following topics:
During installation, Oracle Universal Installer (OUI) assigns port numbers to components from a set of default port numbers. Many Oracle RAC components and services use ports. As an administrator, it is important to know the port numbers used by these services, and to make sure that the same port number is not used by two services on your server.
Most port numbers are assigned during installation. Every component and service has an allotted port range, which is the set of port numbers Oracle RAC attempts to use when assigning a port. Oracle RAC starts with the lowest number in the range and performs the following checks:
Is the port used by another Oracle Database installation on the host?
The installation can be either active or inactive at the time; Oracle Database can still detect if the port is used.
Is the port used by a process that is currently running?
This could be any process on the host, including processes other than Oracle Database processes.
If the answer to any of the preceding questions is yes, then Oracle RAC moves to the next highest port in the allotted port range and continues checking until it finds a free port.
In most cases, the Oracle Database component's port number is listed in the tool used to configure the port. In addition, ports for some Oracle Database applications are listed in the portlist.ini
file. This file is located in the directory $
ORACLE_HOME
/install
.
If you change a port number after installation, then it is not updated in the portlist.ini
file, so you can rely on this file only immediately after installation. To find or change a port number, use the methods described in this appendix.
Table F-1 lists the port numbers and protocols used by components that are configured during the installation. By default, the first port in the range is assigned to the component, if it is available.
Table F-1 Ports Used in Oracle Components
Component and Description | Default Port Number | Port Range | Protocol | Used Only On Interconnect |
---|---|---|---|---|
The port number is assigned automatically during installation. You cannot view or modify it afterward. This port is used exclusively with the cluster interconnect, which is a private network that is physically separated from the public network, and uses dedicated hardware for the interconnect. |
Dynamic |
Dynamic |
UDP |
Yes |
Cluster Synchronization Service (CSSD) Node restart advisory messages fixed port. This port is used on all interfaces that have broadcast capability. Broadcast occurs only when a node eviction restart is imminent. |
42424 |
Dynamic |
UDP |
Yes |
The port number is assigned automatically during installation. You cannot view or modify it afterward. This port is used exclusively with the cluster interconnect, which is a private network that is physically separated from the public network, and uses dedicated hardware for the interconnect. |
Dynamic |
Dynamic |
UDP |
Yes |
Grid Plug and Play (GPNPD) GPNPD provides access to the Grid Plug and Play profile, and coordinates updates to the profile among the nodes of the cluster to ensure that all of the nodes node have the most recent profile. |
Dynamic |
Dynamic |
TCP |
No |
Multicast Domain Name Service (MDNSD) Grid Plug and Play uses the mDNS process to locate profiles in the cluster, as well as by GNS to perform name resolution. The mDNS process is a background process on Linux and UNIX, and a service on Windows. |
5353 |
Dynamic |
UDP/TCP |
No |
Shares the Oracle Net port and is configured during installation. To reconfigure this port, use Net Configuration Assistant to reconfigure the Oracle Net listener. See: |
Default 1521, (same value as the listener) |
modifiable manually to any available port |
TCP |
No |
Listening port for Oracle client connections to Oracle Connection Manager. You can configure Oracle Connection Manager after installation using Net Configuration Assistant. |
1630 |
1630 |
TCP |
No |
Oracle Clusterware Daemon (CRSD) CRS daemon (Oracle Cluster Ready Services Daemon) internode connection. The port number is assigned automatically during installation. You cannot view or modify it afterward. This port is used exclusively with the cluster interconnect. |
Dynamic |
Dynamic |
UDP |
Yes |
The port number is assigned automatically during installation. You cannot view or modify it afterward. This port is used exclusively with the cluster interconnect. |
Dynamic |
Dynamic |
UDP |
Yes |
Oracle Enterprise Manager Database Console HTTP port for Oracle Enterprise Manager Database Control. It is configured during installation. "Changing the Oracle Enterprise Manager Database Console Ports" explains how to modify its port number. |
1158 |
5500–5519 |
TCP/HTTP |
No |
Oracle Enterprise Manager Database Console RMI port for Oracle Enterprise Manager Database Control. It is configured during installation."Changing the Oracle Enterprise Manager Database Console Ports" explains how to modify its port number. |
5520 |
5520–5539 |
TCP |
No |
Oracle Enterprise Manager Database Console JMS port for Oracle Enterprise Manager Database Control. It is configured during installation. "Changing the Oracle Enterprise Manager Database Console Ports" explains how to modify its port number. |
5540 |
5540–5559 |
TCP |
No |
Oracle Grid Interprocess Communication (GIPCD) A support daemon that enables Redundant Interconnect Usage. |
42424 |
Dynamic |
UDP |
Yes |
Oracle Grid Naming Service (GNSD) The Oracle Grid Naming Service daemon provides a gateway between the cluster mDNS and external DNS servers. The |
53 (public), Dynamic (interconnect) |
53 (public), Dynamic (interconnect) |
UDP/TCP (public), TCP interconnect |
No |
Oracle High Availability Services (OHASD) The Oracle High Availability Services (OHAS) daemon starts the Oracle Clusterware stack. This port is used exclusively with the cluster interconnect. |
42424 |
Dynamic |
UDP |
Yes |
HTTP port for Oracle Management Agent, which is part of Oracle Enterprise Manager. It is configured during installation. "Changing the Oracle Management Agent Port" explains how to modify its port number |
3938 |
1830–1849 |
HTTP |
No |
Oracle Notification Services (ONS) Port for ONS, used to publish and subscribe service for communicating information about all Fast Application Notification (FAN) events. The FAN notification process uses system events that Oracle publishes when cluster servers become unreachable or if network interfaces fail. Use |
6100 (local) 6200 (remote) |
Configured manually |
TCP |
No |
Oracle Real Application Clusters The port number is assigned automatically during installation. You cannot view or modify it afterward. This port is used exclusively with the cluster interconnect. |
Dynamic |
Dynamic |
UDP |
Yes |
Allows Oracle client connections to the database over the protocol for Oracle Net. You can configure it during installation. To reconfigure this port, use Net Configuration Assistant. |
1521 |
Port number changes to the next available port. Modifiable manually to any available port. |
TCP |
No |
The Oracle XML DB HTTP port is used if Web-based applications need to access an Oracle database from an HTTP listener. It is configured during installation, but you cannot view it afterward. "Changing the Oracle XML DB Ports" explains how to change its port number. |
0 |
Configured manually |
HTTP |
No |
Oracle XML DB The Oracle XML DB FTP is used when applications need to access an Oracle database from an FTP listener. It is configured during installation, but you cannot view it afterward. "Changing the Oracle XML DB Ports"explains how to change its port number. |
0 |
Configured manually |
FTP |
No |
To find the current setting for the Oracle Management Agent port, search for EMD_URL
in the ORACLE_HOME
/
host_sid
/sysman/config/emd.properties
file, where host_sid
is a string that contains the local host name and the SID for the Oracle RAC database.
To change the Oracle Management Agent HTTP port, use the emca -reconfig
ports command:
emca -reconfig ports -AGENT_PORT 1831
To find the current HTTP, RMI, and JMS port settings, search in the following files, where host_sid
is a string that contains the local host name and the SID for the Oracle RAC database:
HTTP port: Search for REPOSITORY_URL
in the ORACLE_HOME
/host_sid
/sysman/config/emd.properties
file.
RMI port: Search for the port
attribute in the rmi-server
tag in the ORACLE_HOME
/
oc4j/j2ee/OC4J_DBConsole_
host_sid
/config/rmi.xml
file.
JMS port: Search for the port
attribute in the jms-server
tag in the ORACLE_HOME
/
oc4j/j2ee/OC4J_DBConsole_
host_sid
/config/jms.xml
file.
To change the Oracle Enterprise Manager Database Control ports, use the emca -reconfig
ports command:
ORACLE_HOME/bin> emca -reconfig ports option setting
In the previous example, option
specifies one or more of the following ports and setting
is the new port value:
Option | Description | Example |
---|---|---|
DBCONTROL_HTTP_PORT | Sets the HTTP port | emca -reconfig ports -DBCONTROL_HTTP_PORT 1820 |
RMI_PORT | Sets the RMI port | emca -reconfig ports -RMI_PORT 5520 |
JMS_PORT | Sets the JMS port | emca -reconfig ports -JMS_PORT 5521 |
You can specify multiple -reconfig port
settings in one line, for example:
emca -reconfig ports -DBCONTROL_HTTP_PORT 1820 -AGENT_PORT 1821 -RMI_PORT 5520
By default, the FTP and HTTP (including HTTPS) ports for Oracle XML DB are set to 0, which disables FTP or HTTP access to Oracle XML DB.
To change the Oracle XML DB ports, you must run the catxdbdbca.sql
script, which in a default installation is located in $ORACLE_HOME
/rdbms/admin
.
To change the Oracle XML DB ports:
Check that the Oracle listener is running.
If you cannot start the listener, then refer to Oracle Database Net Services Administrator's Guide.
log in to SQL*Plus as SYS
or XDB
using the SYSDBA
privileges.
For example, to log in to SQL*Plus as SYS
:
SQL> sqlplus sys/ as sysdba
Run the catxdbdbca.sql
script.
For example, to use 2200
for the FTP port and 8200
for the HTTP port, and assuming your Oracle home is in the location /u01/app/oracle/product/11.2.0/dbhome_1/
, enter the following command:
SQL> /u01/app/oracle/product/11.2.0/dbhome_1/rdbms/admin/catxdbdbca.sql 2200 8200
Exit SQL*Plus.
See Also: Oracle Database Administrator's Guide for information about connecting to the database using SQL*Plus |
Most Oracle Real Application Clusters (Oracle RAC) installation errors are due to a failure to complete all of the steps that are required before starting Oracle Universal Installer.
This chapter is intended for database administrators to use in consultation with system and storage administrators to coordinate the installation and plan the tasks for Oracle Clusterware, in preparation for completing an installation of Oracle RAC. This chapter contains the following topics:
Confirming Cluster Readiness for Oracle RAC Installation with CVU
Stopping Existing Oracle Processes for Upgrades or Coexisting Databases
This section provides a list of tasks that Oracle recommends you complete before starting Oracle Clusterware and Oracle RAC installation. Whether your location is a Tier IV data center with a large project team of system administrators, storage administrators, network administrators, database administrators, and third-party hardware and software vendors, or you are a project team of one, planning is important to help ensure that your installation proceeds smoothly.
It is beyond the scope of this documentation set to advise how to determine hardware sizing or capacity planning for your installation. Note that with Oracle Clusterware and Oracle RAC, you can add nodes and instances as needed in response to testing, or in response to increased workloads.
Review and complete the following steps as part of your installation plan:
Before you decide whether you want to install Oracle Database 11g release 2 (11.2) on existing hardware, or decide what server and storage hardware to purchase for an installation, log on to My Oracle Support:
Click the Certify tab. Check the Certification Matrix for Oracle RAC for the operating system platform on which you intend to install, to ensure that your hardware configuration is supported for use with Oracle Clusterware and Oracle RAC. You can receive guidance about supported hardware options that can assist you with your purchasing decisions.
To access to the Certification Matrix:
Log on to My Oracle Support.
Click the More tab.
Select Certifications.
Provide information about your platform to obtain certification information.
In addition to specific certified hardware configurations, the Certify page provides support and patch information, and general advice about how to proceed with an Oracle Clusterware or Oracle Clusterware with Oracle RAC 11g release 2 (11.2) installation, including important information about vendor clusterware and other configuration issues.
Contact your Oracle Sales Representative if you do not have a My Oracle Support account.
Also, you may want to refer to Oracle.com (http://www.oracle.com
) for additional resources about planning for specific implementation scenarios, best practices, and other information that can help you with your installation plan. In particular, refer to the Oracle Real Application Clusters pages on the Oracle Technology Network:
http://www.oracle.com/technology/products/database/clustering/index.html
The Oracle RAC Technologies Certification Matrix provides details about supported technology that you can use to deploy Oracle RAC.
OUI can install critical patch updates, system requirement updates for supported operating systems, and other significant updates that can ensure that your installation succeeds. Oracle recommends that you enable the Software Updates option during installation.
If you choose to enable the Software Updates option, then you must provide a valid My Oracle Support user name and password, so that OUI can download the latest updates, or you must provide a path to the location of a Software Updates option package that you have downloaded previously.
In addition, if you have a proxy realm, then be prepared to provide the destination proxy realm, and the user authentication (user name and password) required to authenticate access through that realm to the Internet, so that OUI can obtain software updates. Check with your network administrator before installation if you do not have this information.
If you plan to run the installation in a secured data center, then you can download updates before starting the installation by starting OUI on a system that has Internet access. To start OUI to download updates only, enter the following command:
$ ./runInstaller -downloadUpdates
Provide the My Oracle Support user name and password, and provide proxy settings if needed. After you download updates, transfer the update file to a directory on the server where you plan to run the installation.
The Oracle Technology Network (OTN) contains white papers about deployment options, capacity planning, best practices on various NFS platforms, and extended clusters deployments, which are not addressed in this guide. You can review available papers at the following Web site:
http://www.oracle.com/technology/products/database/clustering/index.html
Installing Oracle RAC consists of the following steps:
Prepare servers (system, storage, and network administration):
Install the operating system and install the operating system packages and patches to the required version.
Create the required groups, users, and software homes.
Set up the domain name forwarding for Grid Naming Service (GNS) if you plan to deploy GNS, and set up the network addresses in the DNS and on the server as needed.
Set up the required storage.
(optional) Stage all of the software on one node for installation (the "local node").
See Also: Oracle Clusterware Administration and Deployment Guide for more information about using OIFCFG to modify interconnects |
Install Oracle Grid Infrastructure for a cluster, which includes Oracle Clusterware and Oracle Automatic Storage Management (system and storage administration):
Install Oracle Grid Infrastructure for a cluster. During installation, Fixup scripts perform additional configuration of operating system parameters, secure shell (SSH) for installation and user environment variables.
Patch Oracle Clusterware and Oracle Automatic Storage Management to the latest patchset.
Install Oracle RAC (database administration):
Install Oracle Real Application Clusters.
Patch Oracle RAC to the latest patchset.
Complete the postinstallation configuration of the Oracle RAC database.
Before installing Oracle Clusterware or Oracle RAC, use CVU to ensure that your cluster is prepared for an installation. CVU is incorporated into Oracle Universal Installer, so CVU runs when you start an Oracle RAC installation. However, you can use CVU to ensure that any packages or configuration information that is required for Oracle RAC are in place before you begin your Oracle RAC installation.
Oracle provides CVU to perform system checks in preparation for an installation, patch updates, or other system changes. In addition, CVU can generate fixup scripts, which are scripts run by the root
user that can change many kernel parameters to at least the minimum settings required for a successful installation.
Using CVU can help system administrators, storage administrators, and database administrators to ensure that everyone has completed the system configuration and preinstallation steps, so that installations, updates, or patches complete successfully. You can obtain the latest version of CVU at the following URL:
http://www.oracle.com/technology/products/database/clustering/cvu/cvu_download_homepage.html
If you have a vendor who is performing hardware or operating system configuration steps, then ask the vendor to complete the relevant CVU check of the cluster after they complete their work to ensure that your system is configured correctly. Database administrators should refer to the section "Confirming Cluster Readiness for Oracle RAC Installation with CVU" to confirm that their system is prepared for installation before they start an Oracle RAC installation.
If you have an existing Oracle installation, then document version numbers, patches, and other configuration information, and review upgrade procedures for your existing installation. Review the Oracle upgrade documentation before proceeding with installation, to decide how you want to proceed.
Oracle RAC 11g release 2 (11.2), you must have Oracle Clusterware and Oracle ASM 11g release 2 (11.2) installed on your cluster. The Oracle Clusterware release version must be equal to or greater than the Oracle RAC version that you want to install.
For late-breaking updates and best practices about pre-upgrade, post-upgrade, compatibility, and interoperability discussions, refer to "Oracle Upgrade Companion." "Oracle Upgrade Companion" is available through Note 785351.1 on My Oracle Support:
For upgrades, note the following:
You can have only one version of Oracle Clusterware running on a cluster at a time. The Oracle Clusterware version must be the most recent release of any software (Oracle Clusterware, Oracle Database, Oracle RAC and Oracle Automatic Storage Management) running on the cluster.
You can have multiple Oracle homes of Oracle Databases on your cluster. However, the Oracle RAC database software in these homes must be from a version that is equal to or prior to the version of Oracle Clusterware that is installed; you cannot have a version of Oracle Database running on Oracle Clusterware that was released after the version of Oracle Clusterware that you are running.
For example:
If you have Oracle Clusterware 11g release 2 installed as your clusterware, then you can have an Oracle Database 10g release 1 single-instance database running on one node, and separate Oracle RAC 10g release 1, release 2, and Oracle RAC 11g release 1 or release 2 databases also running on the cluster.
However, you cannot have Oracle Clusterware 10g release 2 installed on your cluster, and install Oracle RAC 11g.
Starting with release 10.1.0.6 and 10.2.0.3, you can use Database Upgrade Assistant (DBUA) for patch set upgrades with Oracle RAC. You can also use DBUA to upgrade between major point releases of Oracle RAC (for example, from 10.1 to 10.2, or 10.2 to 11g).
You cannot change the owner of the Oracle Database home during an upgrade. You must use the same Oracle software owner that owns the existing Oracle Database home.
As with any system change, back up your existing database before installing new software.
Before you start an installation on which you want to support languages other than English, review Oracle Database Globalization Support Guide.
Note the following:
Oracle recommends that you use Unicode AL32UTF8 as the database character set.
Unicode is the universal character set that supports most of the currently spoken languages. Unicode also supports many historical scripts (alphabets). Unicode is the native encoding of many technologies, including Java, XML, XHTML, ECMAScript, and LDAP. Unicode is ideally suited for databases that support the Internet and the global economy.
The locale setting of your operating system session determines the language of the user interface and the globalization behavior for components such as Oracle Universal Installer, Oracle Net Configuration Assistant, and Database Configuration Assistant. The locale setting also determines the globalization behavior of Oracle Database sessions created by a user application through Oracle JDBC driver, unless overridden by the application.
The NLS_LANG
environment variable determines the language of the user interface and the globalization behavior for components such as SQL*Plus, exp
, and imp
. This variable also sets the language and territory used by the client application and the database. The variable also declares the character set for entering and displaying data by the client application.
Note: Oracle Database Installation Guide for your platform contains a fuller discussion of database character sets used with different languages, and provides further information about installing and configuring Oracle Database globalization support. |
Before installing Oracle Clusterware, review the Oracle Grid Infrastructure Installation Guide for your operating system platform, particularly the Preinstallation, and the Storage chapters, to complete all of the required steps needed for an installation.
The Oracle Grid Infrastructure Installation Guide also contains most tasks requiring root privileges or storage administrator privileges that need to be completed before starting an Oracle RAC installation.
In addition, review the Release Notes and My Oracle Support (https://support.oracle.com
) to ensure that you have the most current information about system requirements and other information that can affect your installation. The small time that this review takes can save a much greater amount of time required to track down causes of installation errors later. Also check to make sure that you have the most current version of this document; Oracle Documentation is updated after release.
Oracle recommends that you install a Web browser on your cluster nodes, both to enable Oracle Enterprise Manager and Oracle Application Express, if you install Oracle RAC, and to access online documentation as needed. Online documentation is available in PDF and HTML formats.
Note: Refer to Oracle Database Concepts for an overview of Oracle Database, and Oracle Real Application Clusters Administration and Deployment Guide for additional information about Oracle Clusterware or Oracle RAC configuration and deployment. The Oracle Grid Infrastructure Installation Guide for your platform contains server and storage configuration information for Oracle RAC. |
Oracle Clusterware must be installed successfully before attempting to install Oracle RAC. To complete installations successfully, ensure that required hardware, network, and operating system preinstallation steps for Oracle software are complete. Failure to complete required preinstallation steps is the most common reason for failed installations.
Before Oracle Clusterware is installed as part of an Oracle Grid Infrastructure for a Cluster installation, you should already have completed installing and configuring CPUs, memory, shared storage, network cards, host bus adaptors, interconnects, and any other networking or server hardware. You should also have installed the operating system, and any vendor clusterware. Review your vendor documentation to complete these tasks, and if relevant, work with your vendor to complete the Oracle preinstallation steps that are listed in this document to confirm that the vendor hardware and software is correctly configured.
During Oracle Grid Infrastructure for a Cluster installation, the person completing the installation identifies the planned use for each global interface, identifying each interface as a Public interface type (used with public IP addresses and virtual IP addresses), a Private interface type (used with interconnects between cluster member nodes), or a Do not use interface type, which Oracle Clusterware and Oracle RAC should ignore. For example, an interface used as a dedicated interface for a network file system should be marked as a Do not use interface type.
Additional network configuration is not required during a Oracle RAC configuration.
Server and network preparation for installation include the following:
The following is a summary of server hardware and software configuration.
Oracle Requires
Each node in a cluster requires the following:
Supported server hardware, including processors and system configuration.
Review My Oracle Support before starting an installation on your current hardware, and before purchasing new hardware, to ensure that the hardware is supported with Oracle Clusterware with Oracle RAC 11g release 2 (11.2).
Note: You must use the same operating system on each node in the cluster. In addition, Oracle strongly recommends that you use the same software configurations on each node of the cluster. |
Operating system packages listed in system requirements.
Customized operating system package installations are not supported, because they may not include required packages. You can add additional packages as needed, but if you subtract packages from the default packages set, then you may encounter what are called failed dependencies errors for required software packages.
A supported interconnect software protocol on each node, to support Oracle Clusterware voting disk polling, and to support Cache Fusion with Oracle RAC. Your interconnect must be certified by Oracle for your platform.
Oracle Recommends: System Administrator Tasks
Oracle recommends the following to simplify server installation and maintenance, and to prevent service issues
Utilizing a time protocol that ensures all nodes in the cluster use the same reference time. With Oracle Clusterware 11g release 2 (11.2), if the network time protocol is not enabled at installation, then the Oracle Clusterware installation enables the cluster time synchronization service.
Configuring redundant switches, for all cluster sizes.
Using identical server hardware on each node, to simplify server maintenance.
Additional Options: System Administrators and Vendors
Though you do not need to use vendor clusterware with Oracle Clusterware, Oracle supports the use of many vendor clusterware options. However, you must install Oracle Clusterware to use Oracle RAC. Be aware that when you use vendor clusterware, Oracle Clusterware defers to the vendor clusterware for some tasks, such as node membership decisions.
You may require third-party vendor clusterware if you use a non-ethernet interconnect.
After you have set up server hardware, review "Checking Hardware Requirements" in the preinstallation chapter to ensure that your system has enough RAM, that the TMP
and TEMP_DIR
locations have enough available space for the installation, and that your system meets other hardware requirements.
Configure the Oracle software groups and users, and configure user environments, as described in the The Oracle Grid Infrastructure Installation Guide for your platform preinstallation chapters. These include the following:
Creating and configuring the Oracle RAC installation owner (typically oracle
); this is the user account that you should use for installation.
Creating an OSDBA operating system group whose members are granted the SYSDBA privilege through operating system authentication (typically dba
).
Creating binary and storage installation paths.
The Oracle base location is the location where Oracle Database binaries are stored. During installation, you are prompted for the Oracle base path. Typically, an Oracle base path for the database is created during an Oracle Grid Infrastructure installation.
Every Oracle home can have its own Oracle base. Typically, the Oracle base path is based on the installation owner name. If you have separate installation owners for the Oracle Grid Infrastructure installation and the Oracle RAC installation, then you have two Oracle base paths that are in accordance with Optimal Flexible Architecture (OFA) guidelines. For example, if the user grid
is the installation owner of the Oracle Grid infrastructure installation, and the user oracle
is the installation owner of the Oracle Database installation, then you have two Oracle base directories:
/u01/app/grid
--the Oracle base for the user grid
, which owns the Oracle Grid Infrastructure binaries
/u01/app/oracle
--the Oracle base for the user oracle
, which owns the Oracle Database binaries
In preparation for installation, Oracle recommends that you only set the ORACLE_BASE
environment variable to define paths for Oracle binaries and configuration files. Oracle Universal Installer (OUI) creates other paths and environment variables as necessary, in accordance with the Optimal Flexible Architecture (OFA) rules for well-structured Oracle software environments.
For example, with Oracle Database 11g, Oracle recommends that you do not set an Oracle home environment variable, and instead enable OUI to create it. If the Oracle base path is /u01/app/oracle
, then by default, OUI creates the following Oracle home path:
/u01/app/oracle/product/11.2.0/dbhome_1
Ensure that the paths that you select for Oracle software, such as Oracle home paths and the Oracle base path, use only ASCII characters. Because installation owner names are used by default for some path, this ASCII character restriction applies to user names, file names, and directory names.
The following is an overview of network configuration requirements for Grid Naming Service (GNS) in a Grid Plug and Play configuration, and in a manual network configuration. Network administrators and system administrators can refer to the Preinstallation chapter in the Oracle Grid Infrastructure Installation Guide for your platform for detailed configuration information.
The network configuration for Oracle Clusterware and Oracle RAC requires several addresses. The following is a list of those addresses:
GNS virtual IP address (GNS installations only): A static IP address configured in the GNS. The GNS virtual IP listener forwards queries to nodes in the subdomain on the cluster managed by GNS.
Within the subdomain, the GNS uses multicast Domain Name Service (mDNS) to enable the cluster to map hostnames and IP addresses dynamically as nodes are added and removed from the cluster, without requiring additional host configuration in the DNS.
To enable GNS, you must have your network administrator provide a set of IP addresses for a subdomain that is assigned to the cluster (for example, grid.example.com
), and forward DNS requests for that subdomain to the GNS virtual IP address for the cluster, which GNS will serve.
Single Client Access Name (SCAN): A domain name that resolves to all the addresses allocated for the SCAN. Oracle recommends that you allocate three addresses to the SCAN. During Oracle Grid Infrastructure installation, listeners are created for each of the SCAN addresses, and Oracle Clusterware controls which server responds to a SCAN address request.
For high availability, you should provide at least three IP addresses in the DNS to use for SCAN mapping. A SCAN domain name must be unique within your corporate network.
Virtual IP address: A public internet protocol (IP) address for each node, to be used as the Virtual IP address (VIP) for client connections. If a node fails, then Oracle Clusterware fails over the VIP address to an available node.
During installation, if you do not use Grid Naming Service (which provides the VIP automatically), then you provide VIP addresses. The VIP for each node is associated with the same interface name on every node that is part of your cluster. If you have a domain name server (DNS), then your network administrator should register the host names for the VIP with the DNS, so that the DNS is resolvable from any client, as well as the cluster nodes. The VIP should not be in use at the time of the installation, because this is an IP address that Oracle Clusterware manages.
Public IP address: A public host name address for each node, assigned by GNS, or assigned by the system administrator during initial system configuration for manual configurations. The public IP address name must be resolvable to the hostname. Register both the public IP and the VIP address with the DNS. If you do not have a DNS, then you must make sure that both public IP addresses are in the node /etc/hosts
file (for all cluster nodes).
Private IP address: A private IP address for each node to serve as the private interconnect address, dedicated exclusively to internode cluster communication. GNS configures the address automatically. If you select manual configuration, then you must ensure that the following is true for each private IP address. The private IP address must be:
Separate from the public network
Accessible on the same network interface on each node
Connected to a network switch between the nodes for the private network; crosscable interconnects are not supported
The private interconnect is used for internode communication by both Oracle Clusterware and Oracle RAC. If you use manual configuration, then the private IP address must be available in each node's /etc/hosts
file. Oracle recommends that the private network interconnects ar configured on dedicated switches that are not connected to anything other than the nodes in the same cluster.
Note: All host names must conform to the RFC 952 standard, which permits alphanumeric characters. Host names using underscores ("_") are not enabled. |
Table 1-1 and the content following provides an overview of IP address requirements for manual network connections.
If you choose not to use GNS, then before installation your network administrator must configure SCAN, public, virtual, and private IP addresses. Your network administrator also must reconfigure the network when nodes are added or removed from the cluster.
For example, with a two node cluster where each node has one public and one private interface, and you have defined a SCAN domain address to resolve on your DNS to one of three IP addresses, you might have the configuration shown in the following table for your network interfaces:
Table 1-1 Manual Network Configuration Example
Identity | Home Node | Host Node | Given Name | Type | Address | Address Assigned By | Resolved By |
---|---|---|---|---|---|---|---|
Node 1 Public |
Node 1 |
|
|
Public |
192.0.2.101 |
Fixed |
DNS |
Node 1 VIP |
Node 1 |
Selected by Oracle Clusterware |
|
Virtual |
192.0.2.104 |
Fixed |
DNS and hosts file |
Node 1 Private |
Node 1 |
|
|
Private |
192.168.0.1 |
Fixed |
DNS and hosts file, or none |
Node 2 Public |
Node 2 |
|
|
Public |
192.0.2.102 |
Fixed |
DNS |
Node 2 VIP |
Node 2 |
Selected by Oracle Clusterware |
|
Virtual |
192.0.2.105 |
Fixed |
DNS and hosts file |
Node 2 Private |
Node 2 |
|
|
Private |
192.168.0.2 |
Fixed |
DNS and hosts file, or none |
SCAN VIP 1 |
none |
Selected by Oracle Clusterware |
mycluster-scan |
virtual |
192.0.2.201 |
Fixed |
DNS |
SCAN VIP 2 |
none |
Selected by Oracle Clusterware |
mycluster-scan |
virtual |
192.0.2.202 |
Fixed |
DNS |
SCAN VIP 3 |
none |
Selected by Oracle Clusterware |
mycluster-scan |
virtual |
192.0.2.203 |
Fixed |
DNS |
Footnote 1 Node hostnames may resolve to multiple addresses.
Secure Shell (SSH) configuration is required for Oracle Clusterware and Oracle RAC.
System administrators can follow the procedure described in "Configuring SSH on All Cluster Nodes" in the Preinstallation chapter of Oracle Grid Infrastructure Installation Guide for your operating system, to configure SSH access for each node on the cluster to all other nodes of the cluster.
Oracle recommends that you ask your system administrator to permit SSH connections on all cluster member nodes for the user account that will own the Oracle RAC installation. When this is enabled, you can use a script within the Oracle Universal Installer to set up SSH configuration for installation. If this script is not able to run because of environment restrictions for the user account you use for installation, then you or your system administrator will need to set up SSH manually for installation.
In addition, note that any messages displayed to the terminal, such as a "message of the day," can disable SSH configuration. Ask your system administrator to review the SSH section in the Oracle Grid Infrastructure Installation Guide for your platform.
You need a web browser to access documentation, and to use Oracle Enterprise Manager and Oracle Application Express. Web browsers must support Java Script and the HTML 4.0 and CSS 1.0 standards. The following browsers meet these requirements:
Netscape Navigator 8.1
Netscape Navigator 9.0
Microsoft Internet Explorer 6.0 SP2
Microsoft Internet Explorer 7.0 SP1
Microsoft Internet Explorer 8.0 and later
Firefox 2.0
Firefox 3.5
Firefox 3.6
Safari 3.1
Safari 4.0.x
Google Chrome 3.0
Google Chrome 4.0
Oracle Clusterware and Oracle RAC are tested with specific operating system kernels, and specific operating system components. Oracle requires that you use the operating system kernel and components that are certified for this release.
Oracle recommends that you or your system administrator review the system requirements carefully in Oracle Grid Infrastructure Installation Guide before beginning installation, to ensure that your system meets these requirements. If your system does not meet minimum operating system kernel and component requirements, then your installation may fail to complete, or other errors may develop during Oracle Clusterware or Oracle Database runtime.
In addition to the standard system requirements configuration, deployment on specific server hardware can include additional operating system configuration steps. Review the Preinstallation chapter, and check the My Oracle Support Certify page to ensure that you are aware of any additional requirements or recommendations for your specific hardware platform configuration.
To install Oracle RAC, you need to have configured shared storage for the database files. Review the following sections for storage planning:
Overview of Oracle Automatic Storage Management Cluster File System
Using Logical Volume Managers with Oracle Grid Infrastructure and Oracle RAC
Refer to the "Configuring Storage for Grid Infrastructure for a Cluster and Oracle Real Application Clusters (Oracle RAC)" chapter in Oracle Grid Infrastructure Installation Guide for your operating system to review storage options for installation planning. Storage and system administrators can refer to this chapter to configure storage for Oracle Database files on Oracle RAC. Note that when Database Configuration Assistant (DBCA) configures automatic disk backup, DBCA uses a database recovery area that must be shared.
With Oracle Database and Oracle Grid Infrastructure release 11.2 and later, Oracle Automatic Storage Management (Oracle ASM) has been extended to include a general purpose file system (Oracle Automatic Storage Management File System, or ACFS), which can be used for any files that you cannot place directly on Oracle ASM, such as database homes, or application files. At the time of this release, this feature is available on Linux, but not available on UNIX platforms.
See Also : The Certify page on My Oracle Support for the most current information about storage options:
Oracle Automatic Storage Management Administrator's Guide for an overview of storage configuration administration |
There are two ways of storing Oracle Database and recovery files:
Note: Installing files on raw or block devices is not supported. Install files on a shared file system or on Oracle ASM. Upgrading existing raw or block devices is supported. |
Oracle Automatic Storage Management: Oracle Automatic Storage Management (Oracle ASM) is an integrated, high-performance database file system and disk manager for Oracle Database files. It performs striping and mirroring of database files automatically.
A supported shared file system: Supported file systems include the following:
A supported cluster file system: Note that if you intend to use a cluster file system for your data files, then you should create partitions large enough for the database files when you create partitions for Oracle Clusterware.
See Also: The Certify page on My Oracle Support for supported cluster file systems |
NAS Network File System (NFS) listed on Oracle Certify: Note that if you intend to use NFS for your data files, then you should create partitions large enough for the database files when you create partitions for Oracle Clusterware.
See Also: The Certify page on My Oracle Support for supported Network Attached Storage (NAS) devices, and supported cluster file systems |
Note: Oracle RAC software can be installed on OCFS2. However, it cannot be installed on Oracle Cluster File System (OCFS). Oracle RAC and Oracle Grid Infrastructure software can be installed on network-attached storage (NAS).For OCFS2 certification status, refer to the Certify page on My Oracle Support. |
For all installations, you must choose the storage option that you want to use for Oracle Database files, or for Oracle Clusterware with Oracle RAC (Oracle RAC). If you want to enable automated backups during the installation, then you must also choose the storage option that you want to use for recovery files (the Fast Recovery Area). You do not have to use the same storage option for each file type.
Table 1-2 shows the storage options supported for storing Oracle Database files and Oracle Database recovery files. Oracle Database files include data files, control files, redo log files, the server parameter file, and the password file.
Note: For up-to-date information about supported storage options for Oracle RAC installations, refer to the Certify pages on the My Oracle Support Web site:
|
Table 1-2 Supported Storage Options for Oracle Database and Recovery Files
Storage Option | File Types Supported | |
---|---|---|
Database | Recovery | |
Oracle Automatic Storage Management |
Yes |
Yes |
OCFS2 |
Yes |
Yes |
Local storage |
No |
No |
NFS file system Note: Requires a certified NAS device |
Yes |
Yes |
Shared raw devices being used with existing installation |
Yes for upgrades only, or as a manual configuration |
No |
Yes for upgrades only, or as a manual configuration |
No |
Use the following guidelines when choosing the storage options that you want to use for each file type:
You can choose any combination of the supported storage options for each file type if you satisfy all requirements listed for the chosen storage options.
Oracle recommends that you choose Oracle Automatic Storage Management (Oracle ASM) as the storage option for database and recovery files.
For Standard Edition Oracle RAC installations, Oracle ASM is the only supported storage option for database or recovery files.
If you intend to upgrade an existing Oracle RAC database, or an Oracle RAC database with Oracle ASM instances, then you must ensure that your system meets the following conditions:
Oracle Universal Installer (OUI) and Database Configuration Assistant (DBCA) are run on the node where the Oracle RAC database or Oracle RAC database with Oracle ASM instance is located.
The Oracle RAC database or Oracle RAC database with an Oracle ASM instance is running on the same nodes that you intend to make members of the new cluster installation. For example, if you have an existing Oracle RAC database running on a three-node cluster, then you must install the upgrade on all three nodes. You cannot upgrade only two nodes of the cluster, removing the third node in the upgrade.
See Also: Oracle Database Upgrade Guide for information about how to prepare for upgrading an existing database |
If you do not have a storage option that provides external file redundancy, then you must configure at least three voting disk areas to provide voting disk redundancy.
Oracle Grid Infrastructure and Oracle RAC only support cluster-aware volume managers. To confirm that a volume manager you want to use is supported, look under the Certifications tab on My Oracle Support, which is available at the following URL:
https://support.oracle.com
This section contains additional information about Oracle Clusterware, Oracle Automatic Storage Management (Oracle ASM), and Oracle RAC, that may be helpful for your installation plan team to read to decide how you want to configure your installation. It contains the following topics:
Obtain the Oracle ASM ASMSNMP Password for Oracle Enterprise Manager
Oracle Database 11g Real Application Clusters Release Compatibility
In past releases, Oracle ASM was installed as part of the Oracle Database installation. With Oracle Database 11g release 2 (11.2), Oracle ASM is part of an Oracle Grid Infrastructure installation. If you want to upgrade an existing Oracle ASM installation, then you must upgrade Oracle ASM by running an Oracle Grid Infrastructure upgrade.
If you did not select Oracle ASM as the storage option for the Oracle Clusterware files, then you must first use Oracle ASM Configuration Assistant (ASMCA) to configure Oracle ASM, create an Oracle ASM instance, and create a disk group to use for your Oracle Database storage.
Obtain the ASMSNMP from the Oracle Grid Infrastructure administrator.
During Oracle Grid Infrastructure for a cluster installation, the ASMSNMP account is created and its status is set to OPEN. It is granted the SYSDBA for ASM privilege. Oracle Enterprise Manager requires access to the ASMSNMP account to monitor Oracle ASM instances and retrieve data from ASM-related data dictionary views.
See Also: Oracle Automatic Storage Management Administrator's Guide for additional information about the ASMSNMP account |
Oracle Clusterware provides clustering services. You do not require vendor clusterware when you use Oracle Clusterware. If you intend to install Oracle RAC, then you must install Oracle Clusterware.
For Oracle RAC, you and your system administrator should note that all instances in Oracle RAC environments share the control file, server parameter file, redo log files, and all data files. These files must be placed on a shared file systems, and all the cluster database instances must have access to them. Each instance also has its own set of redo log files. During failures, shared access to redo log files enables surviving instances to perform recovery.
As part of an installation of Oracle Database 11g release 2 (11.2), time zone version files from 1 to 11 are installed in the path $ORACLE_HOME/oracore/zoneinfo/
. You can continue to use the current time zone version or upgrade to the latest version. Oracle recommends that you upgrade the server to the latest time zone version. Upgrading to a new time zone version may cause existing TIMESTAMP WITH TIME ZONE (TSTZ) data to become stale. Using the newly provided DBMS_DST PL/SQL packages, you can update the TSTZ data transparently with minimal manual procedures and system downtime.
All instances of an Oracle RAC database must use the same time zone. The Oracle RAC database time zone defaults to the time zone setting of the Oracle Grid Infrastructure Grid user, except when an instance is started with SQL*Plus. When you use SQL*Plus, you must be sure to use the same time zone setting for the database instance that is used for Oracle Clusterware. You can change the time zone Oracle Clusterware uses for a database by using the following command, where time zone
is the time zone to which you want to change:
srvctl setenv database -T 'TZ=time zone
Time zone version files are also installed on the clients. Starting with Oracle Database 11g release 2, you no longer need to upgrade Client time zone files immediately. Upgrades can be done at a time when it is most convenient to the system administrator. However, there could be a small performance penalty when client and server use different time zone versions.
See Also: Oracle Database Upgrade Guide for information about preparing to upgrade Timestamp with Time Zone data, Oracle Database Globalization Support Guide for information about how to upgrade the Time Zone file and Timestamp with Time Zone data, and Oracle Call Interface Programmer's Guide for information about performance effects of clients and servers operating with different versions of Time Zone files |
If you are installing Oracle Real Application Clusters on HP-UX, then ensure that there is a low-privileged user created to own external jobs, and be aware that you must set this user as the external jobs user by logging in as the root user and modifying $ORACLE_HOME/rdbms/admin/externaljob.ora
after installation. Only the root user can modify this file.
You can install and operate different releases of Oracle Database software on the same computer:
When you have Oracle Clusterware installed with different release versions of Oracle software, the Oracle Clusterware version must be greater than or equal to the Oracle Database software version. Oracle Clusterware and Oracle Automatic Storage Management are both upgraded with an Oracle Grid Infrastructure 11g release 2 (11.2) installation.
If you have an existing Oracle home, then you can create a new Oracle home and install Oracle Database 11g release 2 (11.2) into the new Oracle home. You should ensure that Oracle Clusterware is in a separate Oracle Grid Infrastructure home. Oracle Grid Infrastructure for a Cluster installations cannot be installed in the Oracle base directory for Oracle Database.
If you are running the Oracle9i release of Oracle RAC, and you want to continue to use that release, then you must run cluster software that is compatible with that release, such as Oracle Cluster Manager or a third party cluster software. Oracle Clusterware release 11g can be installed on the same system as Oracle9i database software, but Oracle9i software cannot be supported by Oracle Clusterware 11g.
Note: If you want to remove third party cluster software after upgrading your database to Oracle Database 10g or Oracle Database 11g, then you must first remove the third party cluster software, and then re-install Oracle Clusterware. |
If OUI detects a previous database release, then OUI asks you about your upgrade preferences. You have the option to upgrade one of the previous release databases with DBUA or to create a new database using DBCA. The information collected during this dialog is passed to DBUA or DBCA after the software is installed.
If OUI detects a previous Oracle Clusterware release, then you are asked to upgrade the existing Oracle Clusterware installation. Only one Oracle Clusterware version can exist on a server, and servers must be members of one cluster only.
You cannot install Oracle Grid Infrastructure for a standalone server and then install Oracle Grid Infrastructure for a cluster. If you have Oracle Grid Infrastructure for a standalone server installed, then you must remove that installation before you can install Oracle Grid Infrastructure for a cluster.
Note: Do not move Oracle binaries from the Oracle home to another location. Doing so can cause dynamic link failures. |
Before you start your installation, use CVU to ensure that your system is prepared for Oracle RAC installation. If any checks fail, then fix the errors reported, either manually or by using a generated fixup script, or contact your system or storage administrator to have the cause of the errors addressed.
CVU is available in the Grid home, in the bin
directory. For example, if the Oracle Grid Infrastructure for a cluster home is /u01/crs
, then the path is /u01/crs/bin
. To start CVU, navigate to the Grid home bin
directory, and use a command similar to the following:
cluvfy stage -pre dbinst -fixup -n nodelist -r release -osdba OSDBA -verbose
In the preceding command, -fixup
and -osdba
are optional flags.
For example, for a two node cluster with nodeA and nodeB, where you are testing the cluster to prepare to install Oracle Database 11g with Oracle RAC, and your OSDBA group is dba
, the following command checks for system readiness:
$ ./cluvfy stage -pre dbinst -fixup -n nodea,nodeb -osdba dba -verbose
For more information about CVU commands, enter ./cluvfy -help
.
If you are planning an installation on a system where you have an existing Oracle RAC or Oracle Database installation, then you must perform additional tasks to prepare your system for installation.
Table 1-3 provides an overview of what you need to do if you have an existing Oracle Database installation. Review the table, and perform tasks as required.
See Also: Oracle Database Upgrade Guide for additional information about preparing for and performing upgrades |
Table 1-3 Overview of System Preparation for Upgrades or Co-existing Databases
Installation Scenario | What you need to do |
---|---|
Upgrading from Oracle Database 10g release 1 (10.1) to 11g release 2 (11.2) |
No additional tasks. Refer to Installing Oracle Database 11g on a System with Oracle Database 10g |
Installing Oracle Database 11g release 2 (11.2) on a system to coexist with Oracle Database 10g release 1 (10.1) |
No additional tasks. Refer to Installing Oracle Database 11g on a System with Oracle Database 10g |
Upgrading from Oracle9i release 9.2 to Oracle Database 11g release 2 (11.2) |
Shut down the Global Service Daemon, and shut down a default listener on port 1521, if present. Refer to Installing Oracle 11g Database on a System with Oracle9i Database Release 2 |
Installing Oracle Database 11g release 2 (11.2) on a system to co-exist with Oracle9i release 9.2 |
Shut down a default listener on port 1521, if present, and shut down the Global Service Daemon. Refer to Installing Oracle 11g Database on a System with Oracle9i Database Release 2 |
Installing Oracle Database 11g on a System with Oracle Database 10g
If your system has an Oracle Database 10g installation, and you install Oracle Database 11g release 2 (11.2) either to coexist with or to upgrade the Oracle Database 10g installation, then most installation types configure and start a default Oracle Net listener using TCP/IP port 1521 and the IPC key value EXTPROC
. One of the following occurs:
During a co-existing installation, Database Configuration Assistant (DBCA) automatically migrates the listener and related files from the Oracle Database 10g Oracle home to the Oracle Database 11g Oracle home.
During an upgrade, Oracle Database Upgrade Assistant (DBUA) automatically locates the Oracle Database10g listener, and migrates the listener to Oracle Database 11g.
Installing Oracle 11g Database on a System with Oracle9i Database Release 2
Explanation of Tasks If you are installing an Oracle Database 11g release 2 (11.2) on a system with an existing Oracle9i Database release 2 (9.2), and the Oracle Net listener process is using the same port or key value as the default used with the Oracle Database 11g release 2 (11.2) installation, port 1521, then OUI can only configure the new listener; it cannot start it. To ensure that the new listener process starts during the installation, you must shut down any existing listeners before starting OUI. To do this, refer to "Shutting Down the Listener"
You must shut down the Global Services Daemon (GSD), because otherwise, during Oracle Database 11g installation, the Oracle Database 9i release 9.2 SRVM shared data is upgraded into an Oracle Cluster Registry that the 9.2 GSD will not be able to use. The Oracle Grid Infrastructure installation starts an 11g release 2 (11.2) GSD to serve the Oracle9i 9.2 clients. To do this, refer to "Shutting down the Global Services Daemon"
Shutting Down the Listener To determine if an existing Oracle Database 9i listener process is running and to shut it down if necessary, follow these steps:
Switch user to the software owner user. For example:
# su - oracle
Enter the following command to determine if an Oracle Database 9i listener process is running and to identify its name and the Oracle home directory in which it is installed:
$ ps -ef | grep tnslsnr
This command displays information about the Oracle Net listeners running on the system:
... oracle_home1/bin/tnslsnr LISTENER -inherit
In this example, oracle_home1
is the Oracle home directory where the listener is installed and LISTENER
is the listener name.
Set the ORACLE_HOME environment variable to specify the appropriate Oracle home directory for the listener:
Bourne, Bash, or Korn shell:
$ ORACLE_HOME=oracle_home1
$ export ORACLE_HOME
C or tcsh shell:
% setenv ORACLE_HOME oracle_home1
Enter the following command to identify the TCP/IP port number and IPC key value that the listener is using:
$ $ORACLE_HOME/bin/lsnrctl status listenername
Note: If the listener uses the default name LISTENER, then you do not have to specify the listener name in this command. |
Enter a command similar to the following to stop the listener process:
$ $ORACLE_HOME/bin/lsnrctl stop listenername
Repeat this procedure to stop all listeners running on this system and on all other nodes in the cluster.
Shutting down the Global Services Daemon As the database installation software owner (for example, oracle
), on each node of the cluster, use the following syntax to shut down the GSD:
$ cd 92_Oracle_home
$ bin/gsdctl stop
In the preceding syntax example, the variable 92_Oracle_home
is the Oracle Database 9i release 2 (9.2) home.
Oracle Real Application Clusters Installation Guide for Linux and UNIX explains how to install and configure Oracle Real Application Clusters (Oracle RAC). Before you use this guide, you must first complete an installation of Oracle Clusterware, as described in the Oracle Grid Infrastructure Installation Guide for your platform.
Oracle Real Application Clusters Installation Guide for Linux and UNIX provides database installation information for database administrators (DBAs) who install and configure Oracle RAC.
For information about Oracle's commitment to accessibility, visit the Oracle Accessibility Program website at http://www.oracle.com/pls/topic/lookup?ctx=acc&id=docacc
.
Access to Oracle Support
Oracle customers have access to electronic support through My Oracle Support. For information, visit http://www.oracle.com/pls/topic/lookup?ctx=acc&id=info
or visit http://www.oracle.com/pls/topic/lookup?ctx=acc&id=trs
if you are hearing impaired.
For more information, refer to the following Oracle resources:
Oracle Real Application Clusters Documentation
Most Oracle error message documentation is only available in HTML format. If you only have access to the Oracle Documentation media, then browse the error messages by range. Once you find a range, use your browser's "find in page" feature to locate a specific message. When connected to the Internet, you can search for a specific error message using the error message search feature of the Oracle online documentation.
Installation Guides
Oracle Diagnostics Pack Installation Guide
Oracle Grid Infrastructure Installation Guide for your platform
Operating System-Specific Administrative Guides
Oracle Clusterware Administration and Deployment Guide
Oracle Database Administrator's Reference, 11g Release 2 (11.2) for UNIX Systems
Oracle Database Platform Guide for Microsoft Windows
Oracle Database 11g Release 2 (11.2) Real Application Clusters Management
Oracle Real Application Clusters Administration and Deployment Guide
Oracle Database 2 Day DBA Online Help
Oracle Enterprise Manager Real Application Clusters Guide Online Help
Getting Started with the Oracle Diagnostics Pack
Generic Documentation
Oracle Database New Features Guide
Oracle Database Net Services Administrator's Guide
Oracle Database Concepts
Oracle Database Reference
Printed documentation is available for sale in the Oracle Store at the following Web site:
To download free release notes, installation documentation, white papers, or other collateral, please visit the Oracle Technology Network (OTN). You must register online before using OTN; registration is free and can be done at the following Web site:
http://www.oracle.com/technetwork/index.html
If you already have a username and password for OTN, then you can go directly to the documentation section of the OTN Web site:
http://www.oracle.com/technetwork/indexes/documentation/index.html
Oracle error message documentation is available only in HTML. You can browse the error messages by range in the Documentation directory of the installation media. When you find a range, use your browser's search feature to locate a specific message. When connected to the Internet, you can search for a specific error message using the error message search feature of the Oracle online documentation.
The following text conventions are used in this document:
Convention | Meaning |
---|---|
boldface | Boldface type indicates graphical user interface elements associated with an action, or terms defined in text or the glossary. |
italic | Italic type indicates book titles, emphasis, or placeholder variables for which you supply particular values. |
monospace | Monospace type indicates commands within a paragraph, URLs, code in examples, text that appears on the screen, or text that you enter. |
This appendix describes the directory structures for Oracle Real Application Clusters (Oracle RAC) software environments. This appendix contains the following topics:
When you install Oracle Database 11g release 2 (11.2) with Oracle RAC, all subdirectories except for oraInventory
, the Oracle Automatic Storage Management home (if applicable), and the Oracle Clusterware home, are under a top-level ORACLE_BASE
directory. The ORACLE_HOME
and ADMIN
directories are also located under ORACLE_BASE
.
See Also: The Oracle Real Application Clusters Administration and Deployment Guide for further information about the$ORACLE_HOME and /ADMIN directories |
Table D-1 shows the hierarchical directory tree of a sample optimal flexible architecture (OFA)-compliant database for Oracle RAC:
Table D-1 Directory Structure for A Sample OFA-Compliant Environment
See Also: Oracle Database Administrator's Reference for Linux and UNIX-Based Operating Systems for further information about the$ORACLE_HOME and /admin directories |
This chapter describes server parameter file (SPFILE) placement and configuration in Oracle Real Application Clusters (Oracle RAC) environments. This chapter contains the following topics:
Server Parameter File Errors in Oracle RAC
See Also: The Oracle Real Application Clusters Administration and Deployment Guide for more information about parameters, and for a discussion of parallel execution-related parameters in Oracle RAC data warehouse environments |
Oracle Database uses parameter settings in parameter files to determine how to control various database resources. You can use two types of files for parameter administration: the server parameter file (SPFILE
) or one or more traditional client-side parameter files.
Oracle recommends that you manage parameters using an SPFILE
. If you use client-side parameter files, then Oracle does not preserve parameter changes made for self-tuning after shutdown.
By default, Oracle Database creates the init
parameter file (PFILE
) for each node based on one SPFILE
. You can change parameter settings in the SPFILE
file by using only Oracle Enterprise Manager or ALTER SYSTEM SET
SQL statements, because the server parameter file is a binary file. Do not try to edit this file manually.
Note: Oracle recommends that you avoid modifying the values for self-tuning parameters; overriding these settings can adversely affect performance. |
If you are upgrading from a previous Oracle Database release, then review the following section about the server parameter file for Oracle RAC.
With Oracle Database 11g release 2 and later databases, Database Agent automatically maintains the SPFILE based on the value you specify in the srvctl command srvctl add/modify -p
spfile
, where spfile
is the name of the SPFILE. You do not need to create or maintain init
ORACLE_SID
.ora
files.
The default location of the server parameter file (SPFILE)
is in the following path:
$ORACLE_HOME/dbs/spfile$ORACLE_SID.ora
The default location of the server parameter file is inappropriate for Oracle RAC databases and Oracle ASM instances if you use storage on a node, because all instances must use the same server parameter file.
Instead, Oracle recommends that you use a PFILE
in this directory:
$ORACLE_HOME/dbs/init$ORACLE_SID.ora
This path is valid for each instance and it refers to a single, shared initialization parameter file. For example, if you use NFS storage, then on a Linux or UNIX system, the file should contain an entry similar to the following:
SPFILE='/nfs/db/oracle_dg/dbspfile'
However, if you use a cluster file system, then use the following file location:
On Linux and UNIX-based systems:
SPFILE='/shared_mount/dbname/spfiledbname.ora'
In this syntax example, the variable shared_mount
is the shared mount point, and the variable dbname
is the database name.
To use DBCA to create your database, and to use the server parameter file, start DBCA. When the Initialization Parameters page appears, select Create server parameter file (SPFILE) under the File Locations tab. When you have made this selection, then enter either a shared file system filename, or enter the storage path in the Server Parameters Filename field.
Note: When you use DBCA to create the server parameter file, the defaultSPFILE file name is $ORACLE_HOME/dbs/init$ORACLE_ SID .ora |
Oracle Database searches for your parameter file in the following order:
$ORACLE_HOME/dbs/spfile
sid
.ora
$ORACLE_HOME/dbs/spfile.ora
$ORACLE_HOME/dbs/init
sid
.ora
Migrate to the SPFile
by creating and editing the server parameter file using the procedures described in this section. This section contains the following topics:
For single-node cluster-enabled configurations, or if you are using Oracle ASM disk groups or a cluster file system, place the server parameter file in an Oracle ASM disk group or on a file system. Otherwise, place the server parameter file on a shared raw device that is at least 5 MB in size.
Migrate to the SPFILE
file by completing the following procedures:
Combine the initialization parameter files for all instances into one init
dbname
.ora
file, where dbname
is the name of the database, by copying all shared IFILE
contents as is. All parameters defined in your IFILE
parameter files are global. Therefore, create them in the format parameter=value, without SID
prefixes.
Copy all instance-specific parameter definitions from init
sid
.ora
files using the following syntax, where the variable sid
is the SID
of the instance:
sid.parameter=value
If you are using a cluster file system, then Oracle recommends that you use the following command to define the location of the new SPFILE, where spfile
is the name of the SPFILE:
srvctl modify database -p spfile
You can also create the server parameter file using the CREATE SPFILE
statement as in the following example:
CREATE SPFILE='?/dbs/spfile_dbname.ora' FROM PFILE='?/dbs/initdbname.ora'
If you use Oracle ASM, then you can use the following syntax to create the server parameter file, where disk_group_name
is the disk group name, db_uniquename
is the database name, dbname
is the database name, and Oracle_home
is the Oracle home path:
CREATE SPFILE='+disk_group_name/db_uniquename/spfiledbname.ora' FROM PFILE='Oracle_home/dbs/initdbname.ora'
These statements read your combined init
dbname
.ora
file that you created by merging your IFILEs, and then transfer the settings for the parameters from the merged file into your server parameter file.
Oracle recommends that you use the SPFILE
by using the PFILE
as a pointer to the SPFILE
. Do this by running the STARTUP
command as in the following example, where the SID variable sid
in the file init
sid
.ora
is the SID that you use in the SPFILE
entry from Step 3. For example:
STARTUP PFILE=$ORACLE_HOME/dbs/initsid.ora
If you use this STARTUP
command syntax, then Oracle Database uses the server parameter file entry specified in the file init
sid
.ora
.
Note: With Oracle ASM 11g release 2 and later releases, theSPFILE for the Oracle ASM instance can be stored in Oracle ASM. You can also place an Oracle ASM instance SPFILE on a file system, or you can use a traditional client-side parameter file for each Oracle ASM instance. |
See Also: The Oracle Automatic Storage Management Administrator's Guide for more information about administering an Oracle ASM instance |
Oracle Database reports errors that occur during server parameter file creation, or while reading the file during startup. If an error occurs during a parameter update, then Oracle records the error in your ALERT.LOG
file, and ignores subsequent parameter updates to the file. If this happens, then do either of the following:
Shut down the instance, recover the server parameter file, and restart the instance.
Enable the instance to continue running without regard for subsequent parameter updates.
Oracle Database displays errors for parameter changes that you attempt when you incorrectly use the ALTER SYSTEM SET
statement. Oracle Database does this when an error occurs while reading from or writing to the server parameter file.
See Also: The Oracle Real Application Clusters Administration and Deployment Guide for more information about backing up theSPFILE |
This appendix describes the procedures for converting from Oracle Database 11g release 2 (11.2) single-instance databases to Oracle Real Application Clusters (Oracle RAC) and Oracle RAC One Node databases. This appendix contains the following topics:
Administrative Issues for Converting Databases to Oracle RAC
Preparing to Convert with rconfig and Oracle Enterprise Manager
If you are upgrading from Oracle Parallel Server to Oracle RAC, or from an earlier version of Oracle RAC, then use Oracle Database Upgrade Assistant (DBUA).
The procedures in this chapter assume that your original single-instance database and the target Oracle RAC database are using the same release and running on the same platform.
Note: You must use clustered Oracle ASM instances for Oracle RAC databases. |
See Also: Oracle Database Licensing Information for compliance requirements that apply to the license you have purchased |
Note the following administrative considerations before converting single-instance databases to Oracle RAC:
Backup procedures should be available before converting from a single-instance Oracle Database to Oracle RAC.
For archiving with Oracle RAC environments, the archive file format requires a thread number.
The archived logs from all instances of an Oracle RAC database are required for media recovery. Because of this, if you archive to a file and you do not use a cluster file system, or some other means to provide shared file systems, then you require a method of accessing the archive logs from all nodes on which the cluster database has instances.
By default, all database files are migrated to Oracle-Managed Files (OMF). This feature simplifies tablespace creation, ensures datafile location consistency and compliance with OFA rules, and reduces human error with data file management.
You can use Database Configuration Assistant (DBCA) to convert from single-instance Oracle databases to Oracle RAC, or Oracle RAC One Node. DBCA automates the configuration of the control file attributes, creates the undo tablespaces and the redo logs, and makes the initialization parameter file entries for cluster-enabled environments. DBCA also configures Oracle Net Services, Oracle Clusterware resources, and the configuration for Oracle RAC database management for use by Oracle Enterprise Manager or the SRVCTL utility (SRVCTL).
Before you use DBCA to convert a single-instance database to an Oracle RAC or an Oracle RAC One Node database, ensure that your system meets the following conditions:
It is a supported hardware and operating system software configuration.
It has shared storage: either Oracle Cluster File System or Oracle ASM is available and accessible from all nodes. On Linux on POWER systems, ensure that GPFS is available and accessible from all nodes.
Your applications have no design characteristics that preclude their use with cluster database processing.
If your platform supports a cluster file system, then you can use it for Oracle RAC. You can also convert to Oracle RAC and use a non-shared file system. In either case, Oracle strongly recommends that you use Oracle Universal Installer (OUI) to perform an Oracle Database 11g release 2 (11.2) installation that sets up the Oracle home and inventory in an identical location on each of the selected nodes in your cluster.
This section describes the following scenarios:
Converting Oracle Database 11g Installations to Oracle RAC Using DBCA
Converting Single Instance on a Cluster to Oracle RAC Using DBCA
To convert from a single-instance Oracle Database that is on a non-cluster computer to Oracle RAC, perform the procedures described in the following sections, and in the order shown:
Use DBCA to create a preconfigured image of your single-instance database by using the following procedure:
Navigate to the bin
directory in $ORACLE_HOME
, and start DBCA.
At the Welcome page, click Next.
On the Operations page, select Manage Templates, and click Next.
On the Template Management page, select Create a database template and From an existing database (structure as well as data), and click Next.
On the Source Database page, select the database name in the Database instance list, and click Next.
On the Template Properties page, enter a name for your template in the Name field. Oracle recommends that you use the database name.
By default, the template files are generated in the directory ORACLE_HOME/assistants/dbca/templates
If you choose to do so, you can enter a description of the file in the Description field, and change the template file location in the Template data file field.
When you have completed the entries, click Next.
On the Location of Database Related Files page, select Maintain the file locations, so that you can restore the database to the current directory structure, and click Finish.
DBCA generates two files: a database structure file (template_name
.dbc
), and a database preconfigured image file (template_name
.dfb
).
Complete Oracle Clusterware installation, as described in the Oracle Grid Infrastructure Installation Guide for your platform.
Validate the cluster configuration using the Cluster Verification Utility (CVU).
Copy the preconfigured database image. This includes copying the database structure *.dbc
file and the database preconfigured image *.dfb
file that DBCA created in the previous procedure "Back Up the Original Single-Instance Database" to a temporary location on the node in the cluster from which you plan to run DBCA.
Run OUI to perform an Oracle Database installation with Oracle RAC.
Select Cluster Installation Mode on the Specify Hardware Cluster Installation page of OUI, and select the nodes to include in your Oracle RAC database.
On the OUI Database Configuration Types page, select the Advanced install type.
After installing the Oracle Database software, OUI runs postinstallation configuration tools, such as Network Configuration Assistant (NETCA), DBCA, and so on.
On the DBCA Template Selection page, use the template that you copied to a temporary location in the "Copy the Preconfigured Database Image" procedure. Use the browse option to select the template location.
Select the option that you want to deploy. Your choices are the following: Oracle RAC database; Oracle RAC One Node database; or Oracle single instance database.
If you selected raw storage on the OUI Storage Options page, then on the DBCA File Locations Tab on the Initialization Parameters page, replace the data files, control files, and log files, and so on, with the corresponding raw device files if you have not set up the DBCA_RAW_CONFIG
environment variable. You must also replace default database files with raw devices on the Storage page.
See Also: Chapter 3, "Creating Oracle Real Application Clusters Databases with Database Configuration Assistant" for more details about DBCA |
After creating the Oracle RAC database, DBCA displays the Password Management page on which you must change the passwords for database privileged users who have SYSDBA
and SYSOPER
roles. When DBCA exits, the conversion process is complete.
Use DBCA to convert a single-instance Oracle Database to Oracle RAC One Node by using the following procedure:
Change directory to $ORACLE_HOME/bin
Start DBCA:
$ dbca
From the Welcome window, select Oracle Real Application Clusters database.
Use the template that you selected to deploy in Section C.2.1.5, "Install Oracle Database 11g Software with Oracle RAC."
There are three scenarios in which a single-instance database can exist on a cluster computer:
Scenario 1: The Oracle home from which the single-instance database is running is cluster-installed. The steps to perform a conversion to Oracle RAC in this scenario are described in "Single Instance on a Cluster Running from a Cluster-Enabled Oracle Home".
Scenario 2: The Oracle home from which the single-instance database is running is cluster-installed, but the Oracle RAC feature is disabled. Refer to "Single Instance on a Cluster Running from an Oracle RAC-Disabled Oracle Home" for instructions on how to convert the database to Oracle RAC in this environment.
Scenario 3: The Oracle home from which the single-instance database is running is not cluster-installed. The procedure to convert the single-instance database to Oracle RAC in this scenario is described in "Single Instance on a Cluster Running from a Single-Instance Installed Oracle Home".
Perform the following procedures to convert a single-instance database on a cluster running from a cluster-installed (Oracle Database 11g with Oracle RAC) Oracle home.
Use DBCA to create a preconfigured image of your single-instance database as described in the section "Back Up the Original Single-Instance Database". To perform the conversion manually, shut down the single-instance database.
To add nodes to your cluster, add and connect these nodes to the cluster as described in the Oracle Grid Infrastructure Installation Guide for your platform. Ensure that all of these nodes can access the shared storage. Also, extend the Oracle Clusterware home to the new nodes using the procedures for "Extending Clusterware and Oracle Software to New Nodes" as described in Oracle Clusterware Administration and Deployment Guide.
From the existing Oracle home, extend this home to the new nodes using the procedure "Adding Nodes at the Oracle RAC Database Layer" as described in Oracle Clusterware Administration and Deployment Guide.
From one of the newly added nodes, configure the listeners on the additional nodes using NETCA. Choose the same port number and protocol that you used on the existing node. If NETCA displays the existing node in the node list page, then do not select this node, because the listener is already configured on it.
Convert the database using one of the following procedures:
If you created the preconfigured image of the single instance database as described in the section "Back Up the Original Single-Instance Database", then use DBCA to complete the conversion to an Oracle RAC database.
Start DBCA from the initial node. Select the names of the nodes to include as part of your cluster database. On the Template Selection page, select the preconfigured template that you created in Step 1. Enter the database name and respond to the remaining DBCA prompts.
To use raw devices for the cluster database files: When the Initialization Parameters page appears, enter the raw device name for the SPFILE on the File Locations tab. On the Storage page, replace the default database file names with the raw devices for the control files, redo logs, and data files to create the cluster database. Click Finish, and create the database.
After creating the Oracle RAC database, DBCA displays the Password Management page on which you must change the passwords for the database privileged users who have SYSDBA
and SYSOPER
roles. When DBCA exits, the conversion process is complete.
Because you did not use DBCA to create a preconfigured image of your single-instance database in step 1, perform the following steps to complete the conversion:
Create the Optimal Flexible Architecture (OFA) directory structure on each of the nodes that you have added.
If you are converting single-instance database files on a file system to raw devices, then copy the database data files, control files, redo logs, and server parameter file to their corresponding raw devices using the dd
command. Otherwise, continue to the next step.
Re-create the control files by running the CREATE CONTROLFILE
SQL statement with the REUSE
keyword and specify MAXINSTANCES
and MAXLOGFILES
, and so on, as needed for your Oracle RAC configuration. The MAXINSTANCES
recommended default is 32.
Shut down the database instance.
If your single-instance database was using an SPFILE, then create a temporary PFILE from the SPFILE using the following SQL statement:
CREATE PFILE='pfile_name' from spfile='spfile_name'
Set the CLUSTER_DATABASE
parameter to TRUE
, set the INSTANCE_NUMBER
parameter to a unique value for each instance, using a sid
.
parameter
=
value
syntax.
If you optimized memory usage on your single-instance database, then adjust the size of the system global area (SGA) to avoid swapping and paging when you convert to Oracle RAC. you should make this adjustment because Oracle RAC requires about 350 bytes for each buffer to accommodate the Global Cache Service (GCS). For example, if you have 10,000 buffers, then Oracle RAC requires about 350 multiplied by 10,000 bytes more memory. Therefore, adjust the size of the SGA by changing the DB_CACHE_SIZE
and DB_nK_CACHE_SIZE
parameters accordingly.
Start up the database instance using the PFILE created in step 5.
If your single-instance database was using automatic undo management, then create an undo tablespace for each additional instance using the CREATE UNDO TABLESPACE
SQL statement. If you are using raw devices, then ensure that the data file for the undo tablespace is on the raw device.
Create redo threads that have at least two redo logs for each additional instance. If you are using raw devices, then ensure that the redo log files are on raw devices. Enable the new redo threads by using an ALTER DATABASE
SQL statement. Then, shut down the database instance.
Copy the Oracle password file from the initial node, or from the node from which you are working, to the corresponding location on the additional nodes on which the cluster database will have an instance. Replace the ORACLE_SID
name in each password file appropriately for each additional instance.
Add REMOTE_LISTENER=LISTENER_
DB_NAME
and sid
.LOCAL_LISTENER=LISTENER_
SID
parameters to the PFILE.
Configure the net service entries for the database and instances, and address entries for the LOCAL_LISTENER
for each instance and for the REMOTE_LISTENER
in the tnsnames.ora
file. When you have done this, copy these to all nodes.
Create the SPFILE from the PFILE using the procedures in the section "Procedures for Migrating to the Server Parameter File". If you are not using a cluster file system, then ensure that the SPFILE is on a raw device.
Create the $ORACLE_HOME/dbs/init
sid
.ora
file that contains the following entry:
spfile='spfile_path_name'
The variable spfile_path_name
is the complete path name of the SPFILE.
On the local node, run catclust.sql
. This command creates the dictionary views needed for Oracle RAC databases. For example:
SQL> start ?/rdbms/admin/catclust.sql
Add the configuration for the Oracle RAC or Oracle RAC One Node database and its instance-to-node mapping using SRVCTL.
Start the database using SRVCTL.
After starting the database with SRVCTL, your conversion process is complete and, for example, you can run the following SQL statement to see the status of all the instances in your Oracle RAC database:
SQL> select * from v$active_instances
You can install a single instance on a cluster running from an Oracle RAC Oracle home that is disabled if you performed a one-node cluster (with Oracle RAC) installation, but later disabled the Oracle RAC feature by unlinking it from the oracle
binary before creating the single instance database. (However, you can also select the local and non-cluster selection on the Node Selection Page to create a single-instance home on a cluster with Oracle RAC disabled.) Perform the following procedures to convert this type of single-instance database to an Oracle RAC database:
Use DBCA to create a preconfigured image of your single-instance database as described in the section "Back Up the Original Single-Instance Database". To perform the conversion manually, shut down the single-instance database.
Change the directory to the lib
subdirectory in the rdbms
directory under the Oracle home.
Relink the oracle
binary by running the following commands:
make -f ins_rdbms.mk rac_on make -f ins_rdbms.mk ioracle
Continue with step 2.
You can only install a single instance on a cluster running from a single-instance installed Oracle home if you selected the local installation option on the OUI Specify Hardware Cluster Installation page during the single-instance Oracle Database 11g release 2 (11.2) installation.
To convert this type of database to an Oracle RAC database, perform the procedures described in the following sections:
"Install Oracle Database 11g Software with Oracle RAC". In this step, make sure that you select a new Oracle home other than the one from which the single-instance database was running.
You can convert an Oracle RAC One Node database to an Oracle RAC database by logging in as the Oracle RAC One Node database owner and entering the following SRVCTL command:
$ srvctl convert database -c RAC
After you run this command, you must create server pools for each service that the database has, in addition to the database server pool. The values for SERVER_NAMES
of the service server pools must be set to the node that you converted from an Oracle RAC One Node to an Oracle RAC node.
Converting an administrator-managed Oracle RAC One Node database to an Oracle RAC database sets all database services so that the single instance is preferred. After you convert the database, you can add instances by running the SRVCTL add
instance
command.
Converting a policy-managed Oracle RAC One Node database to an Oracle RAC database sets all database services to UNIFORM cardinality. It also results in reusing the server pool in which the database currently runs. The conversion reconfigures the database to run on all of the nodes in the server pool. The command does not start any additional instances but running the SRVCTL start
database
command starts the database on all of the nodes in the server pool.
See Also: The Oracle Real Application Clusters Administration and Deployment Guide for more information |
You can use rconfig
, or Oracle Enterprise Manager to assist with converting a single-instance database installation to an Oracle RAC database. The first of these, rconfig
, is a command-line utility. Oracle Enterprise Manager Grid Control database administration option, Convert to Cluster Database, provides a GUI conversion tool. The following sections describe how to use these conversion tools:
Configuration Changes during Oracle RAC Conversion Using RCONFIG
Converting Databases to Oracle RAC Using RCONFIG or Oracle Enterprise Manager
Converting Databases to Oracle RAC Using Oracle Enterprise Manager
Note: Before you start the conversion, as is always true before you start any major change to your database, always back up your existing database. |
Before you convert a single-instance database to an Oracle RAC database, ensure that the following conditions are met for each cluster node that you intend to make an Oracle RAC database node:
Oracle Clusterware 11g release 2 (11.2) is installed, configured, and running.
Oracle RAC 11g release 2 (11.2) software is installed.
The Oracle binary is enabled for Oracle RAC.
Shared storag=Y¦e, either Oracle Cluster File System or Oracle ASM, is available and accessible from all nodes.
User equivalence exists for the oracle
account.
If you intend to use Oracle Enterprise Manager, then all Oracle Enterprise Manager agents are configured and running, and are configured with cluster and host information.
You have backed up your existing database.
Note: You must use clustered Oracle ASM instances for Oracle RAC databases. |
Note the following changes that occur when you convert a single-instance database to Oracle RAC using rconfig
:
During the conversion, rconfig
places the target Oracle RAC database into archive log mode, and enables archiving for the database. If you do not plan to use archivelog space, then you can disable archive logging after the conversion has completed.
For the Shared Storage Type value, if you enter CFS, and you use a cluster file system for your single-instance database storage, then rconfig converts the environment to use Oracle Managed Files (OMF) for database storage, and places the datafiles in a subdirectory located under the shared storage location.
During the conversion, rconfig
moves database files to a specified shared location, and configures them using OMF.
If you want to avoid using OMF with your converted database, then the single-instance database files must already be located in a shared file system, and you must indicate that rconfig
should not move the files.
The following is a list of scenarios for converting a single-instance Oracle database to an Oracle RAC database:
Converting a single-instance Oracle Database 11g release 2 (11.2) to an Oracle 11g release 2 (11.2) Oracle RAC database, running out of the same Oracle home and using the same data files as the single-instance database.
In this scenario, run the rconfig
utility from the Oracle RAC database home, or use the Convert to RAC option on the single-instance database target of Oracle Enterprise Manager Grid Control.
Converting a single-instance Oracle Database version previous to Oracle Database 11g release 2 (11.2), running out of the same Oracle home and using the same data files as the single-instance database
In this scenario, use OUI and Database Upgrade Assistant (DBUA) to update the single-instance database to Oracle Database 11g release 2 (11.2). Then use rconfig
or the Oracle Enterprise Manager Convert to RAC option, as described in the preceding scenario.
Converting a single-instance Oracle Database 11g release 2 (11.2) to an Oracle 11g release 2 (11.2) Oracle RAC database, running out of a different Oracle home and using the same data files as the single-instance database.
In this scenario, run the rconfig
utility in the target database home, or use the Convert to RAC option in the single-instance database target of Oracle Enterprise Manager Grid Control. Provide the file storage location when prompted.
Converting a single-instance Oracle Database 11g release 2 (11.2) to an 11g release 2 (11.2) Oracle RAC database, running out of a different Oracle home, and where the host where the single-instance database is running is not one of the nodes of the Oracle RAC database.
In this scenario, create a clone image of the single-instance database, and move the clone image to a host that is one of the nodes of the cluster. Then use rconfig
or the Oracle Enterprise Manager Convert to RAC option, as described in the preceding scenario.
You can use Oracle Enterprise Manager Grid Control to convert a single-instance database to Oracle RAC. To use this feature, complete the following steps:
See Also: The Oracle Database Upgrade Guide for information about upgrading existing Oracle Database installations |
Log in to Grid Control. From the Grid Control Home page, click the Targets tab.
On the Targets page, click the Databases secondary tab, and click the link in the Names column of the database that you want to convert to Oracle RAC.
On the Database Instance Home page, click the Change Database secondary tab.
On the Administration page, in the Database Administration Change Database section, click Convert to Cluster Database.
Log in as the database user SYS
with SYSDBA
privileges to the database you want to convert, and click Next.
On the Convert to Cluster Database: Cluster Credentials page, provide a user name and password for the oracle
user and password of the target database that you want to convert. If the target database is using Oracle ASM, then also provide the ASM SYS
user and password, and click Next.
On the Hosts screen, select the host nodes in the cluster that you want to be cluster members in the Oracle RAC database installed. When you have completed your selection, click Next.
On the Convert to Database: Options page, select whether you want to use the existing listener and port number, or specify a new listener and port number for the cluster. Also provide a prefix for cluster database instances on the cluster.
When you have finished entering information, click Next, or click Help if you need assistance in deciding how to enter information.
On the Convert to Cluster Database: Shared Storage page, either select the option to use your existing shared storage area, or select the option to have your database files copied to a new shared storage location. Also, decide if you want to use your existing Fast Recovery Area, or if you want to copy your recovery files to a new Fast Recovery Area using files managed by Oracle Database.
If you use Oracle ASM, then Oracle recommends that you place the database area and the recovery area in separate failure groups. A failure group is defined by shared hardware, such as a controller shared between two disks, or two disks that are on the same spindle. If two disks share hardware that could fail, making both disks unavailable, then theses disks are said to be in the same failure group. If you do not use Oracle ASM, then Oracle recommends that the data files and the Fast Recovery Area are located on separate locations, as with separate Oracle ASM failure groups, so that a hardware failure does not affect availability.
When you have finished entering information, click Next, or click Help if you need assistance in deciding how to enter information.
On the Convert to Cluster Database: Review page, review the options you have selected. Click Submit Job if you want to proceed to conversion. If you want to change options you have selected, click Back. To cancel the conversion, click Cancel.
On the Confirmation page, click View Job to check the status of the conversion.
You can use the command line utility rconfig
to convert a single-instance database to Oracle RAC, or to convert it to Oracle RAC One Node, depending on the values you provide in the ConvertToRAC.xml
file. To use this feature, complete the following steps:
See Also: The Oracle Database Upgrade Guide for information about upgrading existing Oracle Database installations |
As the oracle
user, navigate to the directory $ORACLE_HOME/assistants/rconfig/sampleXMLs
, and open the file ConvertToRAC.xml
using a text editor, such as vi
.
Review the ConvertToRAC.xml
file, and modify the parameters as required for your system. The XML sample file contains comment lines that provide instructions for how to configure the file.
Caution: Set the convert optionConvert verify="ONLY" to perform a test conversion to ensure that a conversion can be completed successfully. |
When you have completed making changes, save the file with the syntax filename
.xml
. Make a note of the name you select.
Navigate to the directory $ORACLE_HOME/bin
, and use the following syntax to run the command rconfig
:
rconfig input.xml
where input.xml
is the name of the XML input file you configured in step 2.
For example, if you create an input XML file called convert.xml
, then enter the following command
$ ./rconfig convert.xml
Note: TheConvert verify option in the ConvertToRAC.xml file has three options:
|
If performing the conversion fails, then use the following procedure to recover and reattempt the conversion.:
Attempt to delete the database using the DBCA delete database option.
Restore the source database.
Review the conversion log, and fix any problems it reports that may have caused the conversion failure. The rconfig
log files are under the rconfig
directory in $ORACLE_BASE/cfgtoollogs
.
Reattempt the conversion.
The following are two examples of an XML ConvertToRAC
input file for the rconfig
utility. Example C-1 is an XML input file to convert a single-instance database with Oracle ASM to an Oracle RAC database using Policy Managed server pools on Oracle ASM storage. Example C-2 is an XML input file to convert a single-instance database with Oracle ASM to an Oracle RAC database using Administrator Managed server pools.
Example C-1 RConfig ConvertToRAC for Policy Managed Databases
<?xml version="1.0" encoding="UTF-8"?> <n:RConfig xmlns:n="http://www.oracle.com/rconfig" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.oracle.com/rconfig"> <n:ConvertToRAC> <!-- Verify does a precheck to ensure all pre-requisites are met, before the conversion is attempted. Allowable values are: YES|NO|ONLY --> <n:Convert verify="YES"> <!--Specify current OracleHome of non-rac database for SourceDBHome --> <n:SourceDBHome>/oracle/product/11.2.0/db_1</n:SourceDBHome> <!--Specify OracleHome where the rac database should be configured. It can be same as SourceDBHome --> <n:TargetDBHome>/oracle/product/11.2.0/db_1</n:TargetDBHome> <!--Specify SID of non-rac database and credential. User with sysdba role is required to perform conversion --> <n:SourceDBInfo SID="sales"> <n:Credentials> <n:User>sys</n:User> <n:Password>oracle</n:Password> <n:Role>sysdba</n:Role> </n:Credentials> </n:SourceDBInfo> <!--Specify the list of existing or new server pools which will be used by the Policy Managed Cluster Database. --> <n:ServerPoolList> <n:ExistingServerPool name="custom"/> <n:NewServerPool name="newpool" cardinality="2"/> </n:ServerPoolList> <!--Specify RacOneNode along with servicename to convert database to RACOne Node --> <!--n:RacOneNode servicename="salesrac1service"/--> <!--InstancePrefix is not required for Policy Managed database. If specified, it will be ignored. Instance names are generated automatically based on db_unique_name for Policy Managed dababase.--> <!-- Listener details are no longer needed starting 11.2. Database is registered with default listener and SCAN listener running from Oracle Grid Infrastructure home. --> <!--Specify the type of storage to be used by rac database. Allowable values are CFS|ASM. The non-rac database should have same storage type. ASM credentials are no needed for conversion. --> <n:SharedStorage type="ASM"> <!--Specify Database Area Location to be configured for rac database.If this field is left empty, current storage will be used for rac database. For CFS, this field will have directory path. --> <n:TargetDatabaseArea>+ASMDG</n:TargetDatabaseArea> <!--Specify Fast Recovery Area to be configured for rac database. If this field is left empty, current recovery area of non-rac database will be configured for rac database. If current database is not using recovery Area, the resulting rac database will not have a recovery area. --> <n:TargetFlashRecoveryArea>+ASMDG</n:TargetFlashRecoveryArea> </n:SharedStorage> </n:Convert> </n:ConvertToRAC> </n:RConfig>
Example C-2 Rconfig ConvertToRAC for Administrator Managed Databases
<?xml version="1.0" encoding="UTF-8"?> <n:RConfig xmlns:n="http://www.oracle.com/rconfig" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.oracle.com/rconfig rconfig.xsd"> <n:ConvertToRAC> <!-- Verify does a precheck to ensure all pre-requisites are met, before the conversion is attempted. Allowable values are: YES|NO|ONLY --> <n:Convert verify="YES"> <!--Specify current OracleHome of non-rac database for SourceDBHome --> <n:SourceDBHome>/oracle/product/11.2.0/db_1</n:SourceDBHome> <!--Specify OracleHome where the rac database should be configured. It can be same as SourceDBHome --> <n:TargetDBHome>/oracle/product/11.2.0/db_1</n:TargetDBHome> <!--Specify SID of non-rac database and credential. User with sysdba role is required to perform conversion --> <n:SourceDBInfo SID="sales"> <n:Credentials> <n:User>sys</n:User> <n:Password>oracle</n:Password> <n:Role>sysdba</n:Role> </n:Credentials> </n:SourceDBInfo> <!--Specify the list of nodes that should have rac instances running for the Admin Managed Cluster Database. LocalNode should be the first node in this nodelist. --> <n:NodeList> <n:Node name="node1"/> <n:Node name="node2"/> </n:NodeList> <!--Specify RacOneNode along with servicename to convert database to RACOne Node --> <!--n:RacOneNode servicename="salesrac1service"/--> <!--Instance Prefix tag is optional starting with 11.2. If left empty, it is derived from db_unique_name.--> <n:InstancePrefix>sales</n:InstancePrefix> <!-- Listener details are no longer needed starting 11.2. Database is registered with default listener and SCAN listener running from Oracle Grid Infrastructure home. --> <!--Specify the type of storage to be used by rac database. Allowable values are CFS|ASM. The non-rac database should have same storage type. ASM credentials are no needed for conversion. --> <n:SharedStorage type="ASM"> <!--Specify Database Area Location to be configured for rac database.If this field is left empty, current storage will be used for rac database. For CFS, this field will have directory path. --> <n:TargetDatabaseArea>+ASMDG</n:TargetDatabaseArea> <!--Specify Fast Recovery Area to be configured for rac database. If this field is left empty, current recovery area of non-rac database will be configured for rac database. If current database is not using recovery Area, the resulting rac database will not have a recovery area. --> <n:TargetFlashRecoveryArea>+ASMDG</n:TargetFlashRecoveryArea> </n:SharedStorage> </n:Convert> </n:ConvertToRAC> </n:RConfig>
After completing the conversion, note the following recommendations for Oracle RAC environments, as described in the Oracle RAC documentation:
Follow the recommendations for using load balancing and transparent application failover, as described in the Oracle Real Application Clusters Administration and Deployment Guide
Use locally managed tablespaces instead of dictionary managed tablespaces to reduce contention and manage sequences in Oracle RAC as described in the Oracle Real Application Clusters Administration and Deployment Guide
Follow the guidelines for using automatic segment space management as described in the Oracle Real Application Clusters Administration and Deployment Guide
The buffer cache and shared pool capacity requirements in Oracle RAC are slightly greater than the requirements for single-instance Oracle databases. Therefore, you should increase the size of the buffer cache by about 10%, and the size of the shared pool by about 15%.
This chapter describes server pool and Enterprise Manager Database Control configuration in Oracle Real Application Clusters (Oracle RAC) environments. This chapter contains the following topics:
Server pools are logical divisions of a cluster into pools of servers, which are allocated to host databases or other applications. Server pools are managed using crsctl
and srvctl
commands.
Caution: By default, any named user may create a server pool. To restrict the operating system users that have this privilege, Oracle strongly recommends that you add specific users to the CRS Administrators list. |
See Also: Oracle Clusterware Administration and Deployment Guide for more information about adding users to the CRS Administrator's list. |
Each server pool name must be unique within the cluster. Two server pools cannot use the same name.
There are two types of server pool management:
Administrator-managed: Database administrators define the servers on which databases resource run, and place resources manually as needed. This is the management strategy used in previous releases.
Policy managed: Database administrators specify in which server pool (excluding generic or free) the database resource will run. Oracle Clusterware is responsible for placing the database resource on a server.
The server pool name is a required attribute. You can also provide values for the following attributes; if you do not specify them, then they are set to the default value:
MIN_SIZE: Minimum number of servers on which you want a resource to run. The default is 0
.
MAX_SIZE: Maximum number of servers on which you want a resource to run. The default is -1
, which indicates that resources can run on all available nodes in the cluster.
IMPORTANCE: The relative importance of the resource pool, used to determine how to reconfigure servers when a node joins or leaves the cluster. The default is 0
.
Note the following about Oracle RAC One Node and server pools:
Oracle RAC One Node runs only in one server pool. This server pool is treated the same as any other server pool.
Online relocation of an Oracle RAC One Node database instances permits planned migrations of an Oracle RAC One Node database from one node to another node. Relocations must always be within a server pool.
With WLM integration, online relocation of Oracle RAC One Node database instance operations do not take a server away from a server pool with its importance set to a lower value, because the target server must already be in the server pool, or in Free.
An Oracle RAC One Node database can be a Data Guard physical standby.
Using policy-based server pool management, database administrators define resource requirements for the workload, including the following:
Defining the number of servers or instances preferred for the application. Oracle Clusterware automatically allocates resources up to the number specified, so long as they are physically available, and relocates a resource to another node as necessary.
Defining the number of instances required to support the application.
Resources are no longer defined as belonging to a specific instance or node. Instead, the priority of resource requirements is defined.
The following sections describe how Oracle Enterprise Manager Database Control is configured during the Oracle RAC installation, and how you can configure Database Control after installation:
Each EMCA command can be used in Oracle RAC environments, and certain commands are only applicable in cluster configurations. To perform EMCA cluster commands, use the -cluster
flag that is available in almost every EMCA operational mode.
When you use EMCA to configure Database Control for Oracle RAC, you configure the Database Control for each instance in the cluster. However, by default, the Database Control Console will only start on the local node. On every other node of the cluster, only the Enterprise Manager agent will start. This is because the Database Control Console opens a number of connections to the database. If an instance of the console is running on every host in the cluster, then you can easily exceed the maximum number of permitted open connections on a 32-node or 64-node environment.
When Database Control Console is started on the local node, on every other node, the commands emctl start dbconsole
and emctl stop dbconsole
only start and stop the agent. Each of the remote agents upload their respective data to the console running on the local node, where you can monitor and manage all the targets in the cluster. On each instance of the Oracle RAC database, the following subdirectories will be created, where nodename
is the name of a node in the cluster, and DBUniqueName
is the database unique name.
$ORACLE_HOME/nodename_DBUniqueName $ORACLE_HOME/nodename_DBUniqueName
However, note that if you upgrade an existing Oracle RAC 10g release 1 database configured with Database Control to the current release, then the existing Database Control configuration is retained. The existing Database Control has a Database Console running on each cluster node. The console will still be started on each individual node.
If you want to modify the existing configuration, then use the following command:
emca -reconfig dbcontrol –cluster –EM_NODE nodename -EM_NODE_LIST NODE_list
where nodename
is the public name of the node, and NODE_list
is a comma-deliminated list of nodes. This command reconfigures the current Database Control setup, and performs the following actions in sequence:
Starts a Database Control Console on nodename
, if one has not been started yet.
Redirects the agents monitoring the database instances in NODE_list so that they upload their data to the console running on nodename
. Also, agents monitoring database instances on nodename
will also upload their data to the local console. Note that if you do not enter the command options -EM_NODE or -EM_NODE_LIST at the command line, you will be prompted for them.
-EM_NODE defaults to the local node if not specified when prompted. If not specified, -EM_NODE_LIST defaults to all nodes where the database is configured.
Use this command to start the console on more than one node. For instance, on an eight-node cluster with nodes node1
, node2
, node3
, node4
, node5
, node6
, node7
, and node8
, you can run the following commands in succession:
$ emca -reconfig dbcontrol –cluster –EM_NODE node1 -EM_NODE_LIST node2,node3,node4 $ emca -reconfig dbcontrol –cluster –EM_NODE node5 -EM_NODE_LIST node6,node7,node8
In this scenario, there are two Database Control consoles running, one on node1
and the other on node5
. From either of these consoles, you can manage and monitor all of the targets in the cluster.
For information about the current cluster configuration, you can run the following command:
$ emca -displayConfig dbcontrol –cluster
Caution: If you use the commandemca -cluster to configure Database Control for Oracle RAC, then check the TNS_ADMIN environment variable on all cluster nodes. If you have a different TNS_ADMIN set for each node, then the listener for the target cannot be configured correctly. In this case, set the same TNS_ADMIN on all cluster nodes before running the emca -cluster command. |
To configure Enterprise Manager for a newly added node of a cluster, use the following command:
$ emca -addNode db
On policy managed databases, another common operation is the addition and deletion of cluster nodes. After you add a new node, you can run EMCA to configure Database Control or central agent management for that node using the command emca -addNode db
. Running EMCA only configures Enterprise Manager so that you can manage the instance in a way that is consistent with the rest of the cluster database instances.
When configuring Enterprise Manager for a new node, run the EMCA command only after you have added the node. Also, run the command from a node in the cluster that already has Enterprise Manager configured for it, because the Enterprise Manager configuration settings on the existing node are propagated to the new node when you run the command. Because the -addNode
option can be used only in an Oracle RAC environment, you do not need to use the -cluster
option on the command line.
After running the command emca -addNode db
, enter the node name and database unique name for the node and database. For example:
Node name: node2 Database Unique Name: EM102
To deconfigure Enterprise Manager for a specific node you are deleting from the cluster, log into a different node from the node that you plan to delete, and use the command emca -deleteNode db
. Running this command removes the Enterprise Manager configuration so that you will no longer be able to manage the instance with Enterprise Manager on that node. Ensure that you run the EMCA command before you delete the cluster node. Because this option can be used only in an Oracle RAC environment, you do not need to use the -cluster
option on the command line.
If you create a database while installing Oracle RAC 11g release 2 (11.2), you have the option of configuring your database so it can be managed by Oracle Enterprise Manager 10g Grid Control Console or by Oracle Enterprise Manager 11g Database Control Console.
To select Grid Control Console as your management option, the Oracle Management Service must be installed on a network host. In addition, the Oracle Management Agent must be installed on the host where you are installing the database. Otherwise, the Grid Control Console option is unavailable and you must instead manage your database with Database Control.
For most of the Oracle RAC 11g release 2 (11.2) installation types, you must choose either Database Control or Grid Control as your management option when you create a database during the installation.
However, if you create a database using one of the following methods, then you can choose not to configure Database Control:
Choosing the Advanced database configuration option during an Enterprise or Standard Edition installation
Running Database Configuration Assistant (DBCA) after the installation
If you do not configure Database Control during the Oracle RAC 11g release 2 (11.2) installation, no hostname_dbuniquename directory is created in the resulting Oracle home directory.
The primary method for configuring an existing Oracle Database 11g release 2 (11.2) database so it can be managed with Database Control is to use DBCA. You can use DBCA to create a new database or to reconfigure an existing database.
See Also: "Installing Oracle Software and Building the Database" in Oracle Database 2 Day DBA Online Help for more information about using DBCA to create a new database instance, and Oracle Enterprise Manager Real Application Clusters Guide Online Help for Oracle RAC examples using DBCA |
To use DBCA to reconfigure your database so it can be managed with Database Control:
Log into the database host as a member of the administrative group that is authorized to install Oracle software and create and run the database.
Start DBCA and change directory to the ORACLE_HOME/bin
directory and enter the following command:
$ ./dbca
The DBCA Welcome page appears.
Advance to the Operations page and select Configure Database Options.
Advance to the Database page and select the database you want to configure.
Advance to the Management Options page and select the following options:
Configure the Database with Enterprise Manager
Use Database Control for Database Management
Optionally, select the options for enabling e-mail notifications and enabling daily backups.
For more information about Enterprise Manager notifications and daily backups, click Help on the Management Options page.
Advance until the Finish button is available.
Click Finish to reconfigure the database so it uses Database Control.
After DBCA reconfigures the database, a new subdirectory appears in the Oracle home. This directory is named using the following format and contains Database Control configuration and state files specific to the database you just configured:
hostname_dbuniquename
Note that for cluster databases, the directories are named nodename_dbuniquename
. For example:
mycluster1.example.com_myNewDB
When you use DBCA to configure Oracle RAC 11g release 2 (11.2), DBCA provides a graphical user interface to help you select Database Control options and to configure other aspects of your database. However, if you want to use the operating system command line to configure Database Control, then you can use the Enterprise Manager Configuration Assistant (EMCA).
Caution: During the configuration of Database Control using EMCA, the database may become unavailable, and users may not be able to connect to the database or perform operations on the database. |
To configure Database Control with EMCA:
Set the ORACLE_HOME
environment variable to identify the Oracle home for the database you want to manage:
Change directory to the ORACLE_HOME/bin
directory.
Start EMCA by entering the following command with any of the optional command-line arguments shown in Table 6-1:
$ ./emca
Depending upon the arguments you include on the EMCA command line, EMCA prompts you for the information required to configure Database Control.
For example, enter the following command to configure Database Control so it will perform automatic daily backups of your database:
$ ./emca -config dbcontrol db -backup
EMCA commands are of the form:
emca [operation] [mode] [flags] [parameters]
Note: To configure Database Console for a single-instance database that uses Oracle ASM, you do not need to provide extra parameters with the EMCA command. Run the following command to configure the Database Console which will automatically detect the Oracle ASM instance:emca -config dbcontrol db -repos create |
Table 6-1 describes the valid execution operations and modes, and lists the optional parameters in brackets. Table 6-2 discusses the flags and their behavior, while Table 6-3 defines the optional parameters in detail. EMCA parameters are of the form [-parameterName parameterValue]. Multiple parameters can be used in combination at the command line.
Table 6-1 EMCA Command-Line Operations
Command | Description |
---|---|
emca -h | --h | -help | --help |
Use this option to display the Help message for the EMCA utility. The options described in Table 6-1, Table 6-2, andTable 6-3, and the valid parameters you may include are listed. |
emca –version |
Prints the version information associated with EMCA. |
emca -config dbcontrol db [-repos (create | recreate)] [-cluster] [-silent] [-backup] [parameters] |
Configures Database Control for a database. Options include creating (or recreating) Database Control repository, configuring automatic backups, and performing these operations on a cluster database. |
emca -config centralAgent (db | asm) [-cluster] [-silent] [parameters] |
Configures central agent management for a database or an Oracle Automatic Storage Management (ASM) instance. Options include performing this operation on a cluster environment.This operation will configure the database so that it can be centrally managed by the Oracle Enterprise Manager 10g Grid Control Console. To use this option, you must have previously installed the Oracle Management Service component of Enterprise Manager on a network host. In addition, the Oracle Management Agent must be installed on the host where you are running the database. |
emca -config all db [-repos (create | recreate)] [-cluster] [-silent] [-backup] [parameters] |
Configures both Database Control and central agent management for a database. The possible configuration options are similar to those described above. |
emca -deconfig dbcontrol db [-repos drop] [-cluster] [-silent] [parameters] |
Deconfigures Database Control for a database. Options include dropping the Database Control repository and performing these operations on a cluster database. For example, you might use this command to remove the Database Control configuration from a database you are planning to delete. In such a scenario, remove the Database Control configuration before physically deleting the database. This operation does not remove the actual database or its data files. |
emca -deconfig centralAgent (db | asm) [-cluster] [-silent] [parameters] |
Deconfigures central agent management for a database or an Oracle ASM instance. Options include performing this operation on a cluster environment. For example, you might use this command to remove the central agent management configuration from a database you are planning to delete. In such a scenario, remove the central agent management configuration before physically deleting the database. This operation does not remove the actual database or its data files. |
emca -deconfig all db [-repos drop] [-cluster] [-silent] [parameters] |
Deconfigures both Database Control and central agent management for a database. The possible deconfiguration options are similar to those described above. |
emca -addNode (db | asm) [-silent] [parameters] |
Configures Enterprise Manager for a new cluster instance of a database or ASM storage. For more information, refer to "Understanding Using EMCA with Oracle RAC". |
emca -deleteNode (db | asm) [-silent] [parameters] |
Deconfigures Enterprise Manager for a node getting deleted. This is discussed further below, in "Understanding Using EMCA with Oracle RAC". For more information, refer to "Understanding Using EMCA with Oracle RAC". |
emca -addInst (db | asm) [-silent] [parameters] |
Configures Enterprise Manager for a new cluster instance of a database or Oracle ASM storage. For more information, refer to "Understanding Using EMCA with Oracle RAC". |
emca -deleteInst (db | asm) [-silent] [parameters] |
Deconfigures Enterprise Manager for a specific instance of a cluster database or Oracle ASM storage. This is discussed further below, in "Understanding Using EMCA with Oracle RAC". |
emca -reconfig ports [-cluster] [parameters] |
Explicitly reassigns Database Control ports. Options include performing this operation on a cluster environment. For more information, refer to "Specifying the Ports Used by Database Control". |
emca -reconfig dbcontrol -cluster [-silent] [parameters] |
Reconfigures Database Control deployment for a cluster database. Note that this command must be used with the "-cluster" option. For more information, refer to "Understanding Using EMCA with Oracle RAC". |
emca -displayConfig dbcontrol -cluster [-silent] [parameters] |
Displays information about the current deployment configuration of Database Control in a cluster environment. Note that this command must be used with the "-cluster" option. For more information, refer to "Understanding Using EMCA with Oracle RAC". |
emca -upgrade (db | asm | db_asm) [-cluster] [-silent] [parameters] |
Upgrades the configuration of an earlier version of Enterprise Manager to the current version. This operation can be performed for database, Oracle ASM, or database and Oracle ASM instances together simultaneously. This does not upgrade the actual database or Oracle ASM instances, nor does it upgrade the Enterprise Manager software. Instead, it upgrades the configuration files for the specified instance so that they are compatible with the current version of the Enterprise Manager software. EMCA will attempt to upgrade all instances of the specified database and/or Oracle ASM target on the host, across all Oracle Homes (since it is likely that certain target properties, such as listener port or Oracle Home, have changed). |
emca -restore (db | asm | db_asm) [-cluster] [-silent] [parameters] |
Restores the current version of Enterprise Manager configuration to an earlier version. This is the inverse of the "-upgrade" option (and will reverse any changes that result from an "-upgrade" operation), and as such, the options are similar. |
Table 6-2 EMCA Command-Line Flags
Flag | Description |
---|---|
db |
Performs the operation for a database (including cluster databases). Use this option for databases that use Oracle Automatic Storage Management (ASM) to store the data files. If a database is using Oracle ASM, then all of the configuration operations and modes described here (except for |
asm |
Performs the operation only for Oracle ASM instances (including cluster Oracle ASM instances). |
db_asm |
This flag can only be used in |
-repos create |
Creates a new Database Control management repository. |
-repos drop |
Drops the current Database Control management repository. |
-repos recreate |
Drops the current Database Control management repository and then recreates a new one. |
-cluster |
Performs the operation for a cluster database or Oracle ASM instance. |
-silent |
Performs the operation without prompting for additional information. If this mode is specified, then all the required parameters must be entered at the command line or specified in an input file using the –respFile argument. You can view a list of the available parameters by entering emca -help at the command line. |
-backup |
Configures automatic backup for a database. EMCA will prompt for daily automatic backup options. The default Enterprise Manager settings will be used to backup the database files. Note: If you use this option, EMCA will use the value of the db_recovery_file_dest initialization parameter to identify the recovery area for the automated backups. If that parameter is not set, EMCA will generate an error. You can modify these settings later using the Maintenance page in Database Control. For more information, see the Database Control online Help. |
Table 6-3 EMCA Command-Line Parameters
Parameter | Description |
---|---|
-respFile |
Specifies the path of an input file listing parameters for EMCA to use while performing its configuration operation. |
-SID |
Database system identifier |
-PORT |
Port number for the listener servicing the database |
-ORACLE_HOME |
Database Oracle Home, as an absolute path |
-LISTENER_OH |
Oracle Home from where the listener is running. If the listener is running from an Oracle Home other than the one on which the database is running, the parameter LISTENER_OH must be specified. |
-HOST_USER |
Host machine user name (for automatic backup) |
-HOST_USER_PWD |
Host machine user password (for automatic backup) |
-BACKUP_SCHEDULE |
Schedule in the form of "HH:MM" (for daily automatic backups) |
-EMAIL_ADDRESS |
E-mail address for notifications |
-MAIL_SERVER_NAME |
Outgoing Mail (SMTP) server for notifications |
-ASM_OH |
Oracle Automatic Storage Management Oracle Home |
-ASM_SID |
System identifier for Oracle ASM instance |
-ASM_PORT |
Port number for the listener servicing the Oracle ASM instance |
-ASM_USER_ROLE |
User role for connecting to the Oracle ASM instance |
-ASM_USER_NAME |
User name for connecting to the Oracle ASM instance |
-ASM_USER_PWD |
Password for connecting to the Oracle ASM instance |
-DBSNMP_PWD |
Password for DBSNMP user |
-SYSMAN_PWD |
Password for SYSMAN user |
-SYS_PWD |
Password for SYS user |
-SRC_OH |
Oracle Home of the database with Enterprise Manager configuration to be upgraded/restored |
-DBCONTROL_HTTP_PORT |
Use this parameter to specify the port you use to display the Database Control Console in your Web browser. For more information, refer to "Specifying the Ports Used by Database Control". |
-AGENT_PORT |
Use this parameter to specify the Management Agent port for Database Control. For more information, refer to "Specifying the Ports Used by Database Control". |
-RMI_PORT |
Use this parameter to specify the RMI port for Database Control. For more information, refer to "Specifying the Ports Used by Database Control". |
-JMS_PORT |
Use this parameter to specify the JMS port for Database Control. For more information, refer to "Specifying the Ports Used by Database Control". |
-CLUSTER_NAME |
Cluster name (for cluster databases) |
-DB_UNIQUE_NAME |
Database unique name (for cluster databases) |
-SERVICE_NAME |
Database service name (for cluster databases) |
-EM_NODE |
Node from which Database Control console is to be run (for cluster databases). For more information, refer to "Understanding Using EMCA with Oracle RAC". |
-EM_NODE_LIST |
Comma-separated list of nodes for agent-only configurations, uploading data to –EM_NODE. For more information, refer to "Understanding Using EMCA with Oracle RAC". |
Instead of answering a series of prompts when you run EMCA, you can use the -respFile
argument to specify an input file. The input file that you create must be in a format similar to the following example:
PORT=1521 SID=DB DBSNMP_PWD=xpE234D SYSMAN_PWD=KDOdk432
After you create an EMCA input file, you can use it on the command line as follows:
$ ./emca -config dbcontrol db -respFile input_file_path
For example, to configure the Database Control to perform daily backups and create the Database Control Management Repository, create an input file similar to the one shown in the following example:
PORT=1521 SID=DB DBSNMP_PWD=dow3l224 SYSMAN_PWD=squN3243 HOST_USER=johnson HOST_USER_PWD=diTf32of SYS_PWD=qlKj4352 BACKUP_SCHEDULE=06:30
When you have created the input file, then enter the following command at the operating system prompt:
$ ./emca -config dbcontrol db -repos create -backup -respFile input_file_path
When you initially install Oracle RAC 11g release 2 (11.2) or configure Database Control with EMCA, Database Control uses a set of default system ports. For example, by default, you access Database Control using port 1158 in 11g Release 2, as in:
http://host.domain:1158/em
This is the default port assigned to Database Control by the Internet Assigned Numbers Authority (IANA). Likewise, the default Database Control Agent port, as assigned by the IANA, is 3938.
To use ports other than the default ports, use the following EMCA command-line arguments when you initially configure the Database Control with EMCA. Alternatively, you can explicitly assign ports after configuring Database Control using the following command:
emca -reconfig ports [-cluster]
Note: You can also use the following EMCA command-line arguments to configure Database Control after you have installed and configured Oracle RAC 11g release 2 (11.2). |
The following list summarizes the EMCA command-line arguments that control the standard Database Control port assignments:
-DBCONTROL_HTTP_PORT port_number
This port number is used in the Database Control Console URL. For example, if you set this port to 5570, you can then display the Database Control Console using the following URL:
http://host.domain:5570/em
-RMI_PORT port_number
This port number is used by the Remote Method Invocation (RMI) system, which is part of the J2EE software required by Database Control. The default port can be changed if you want to configure a specific port for Database Console. When a port other than the default port (1521) is used, use the -RMI_PORT or -JMS_PORT options along with the emca reconfig command.
-JMS_PORT port_number
This port is used by the OC4J Java Message Service (JMS), which is part of the J2EE software required by Database Control. The default port can be changed if you want to configure a specific port for Database Console. When a port other than the default port (1521) is used, use the -RMI_PORT or -JMS_PORT options along with the emca reconfig command.
-AGENT_PORT port_number
This port is used by the Database Control Management Agent, which is monitoring and administering the database for the Database Control.
The following section describes some troubleshooting tips to consider when using EMCA to configure the Database Control:
If you change the listener port of the database after you have configured Database Control, then the database status appears as down. To reconfigure Database Control so it uses the new listener port, run the EMCA command using the -config dbcontrol db [-cluster] command-line arguments.
When upgrading an Oracle Database 10g Release 1 database and Oracle ASM instance that was configured for Oracle Enterprise Manager (either Database Control or a Grid Control central agent) to an Oracle Database 11g release 2 (11.2) database, all Enterprise Manager targets on the relevant hosts referring to the upgraded instances are updated automatically. This is because the upgrade involves altering the instance Oracle Home, port, or other target-associated properties. However, some of these targets on the hosts are not updated successfully during the upgrade if they are managed by an Oracle Database 11g release 2 (11.2) Grid Control Agent. To update these targets, in the Home page for the upgraded database (or Oracle ASM) target, click the Monitoring Configuration. On the Monitoring Configuration page, you can update properties such as Oracle Home, listener port and so on to the correct values.
When the database host name (including the domain name) or the IP address changes, deconfigure and then reconfigure the Database Console with the -repos create
flag for the emca
command. For example:
emca -deconfig dbcontrol db -repos drop emca -config dbcontrol db -repos create
or
emca -deconfig dbcontrol db emca -config dbcontrol db -repos recreate
When the TNS configuration is changed, set the environment variable and then run the following command:
emca -config dbcontrol db