IEEE 1394 is an interface standard for a serial bus for highspeed communications and isochronous realtime data transfer. It was developed in the late 1980s and. V7. 7. x Supported Hardware List, Device Driver, Firmware and Recommended Software Levels for SAN Volume Controller. KernSafe TotalMounter can be easily used on Windows 7, including 64bit edition allowing for comfortable mounting almost any image file types. DAEMON Tools Lite free for noncommercial usage product is a wellknown solution that allows you to mount, copy and create an image. It works with the most. Windows 7 Command Prompt Commands A Complete List of CMD Commands Available in Windows 7. Build Your Own Oracle RAC 11g Cluster on Oracle Linux and iSCSI by Jeffrey Hunter. Learn how to set up and configure an Oracle RAC 11g Release 2 development cluster. Oracle VM VirtualBox User Manual. Oracle Corporation. Build Your Own Oracle RAC 1. Cluster on Oracle Enterprise Linux and i. SCSIby Jeffrey Hunter Learn how to set up and configure an Oracle RAC 1. Release 2 development cluster on Oracle Linux for less than US2,7. The information in this guide is not validated by Oracle, is not supported by Oracle, and should only be used at your own risk it is for educational purposes only. Updated November 2. Contents. Introduction. Oracle RAC 1. 1g Overview Shared Storage Overviewi. SCSI Technology. Hardware and Costs. Install the Linux Operating System. Install Required Linux Packages for Oracle RACNetwork Configuration. Cluster Time Synchronization Service. Install Openfiler. Configure i. SCSI Volumes using Openfiler. Configure i. SCSI Volumes on Oracle RAC Nodes. Create Job Role Separation Operating System Privileges Groups, Users, and Directories. Logging In to a Remote System Using X Terminal. Configure the Linux Servers for Oracle. Configure RAC Nodes for Remote Access using SSH OptionalAll Startup Commands for Both Oracle RAC Nodes. Install and Configure ASMLib 2. Download Oracle RAC 1. Release 2 Software Preinstallation Tasks for Oracle Grid Infrastructure for a Cluster. Install Oracle Grid Infrastructure for a Cluster. Postinstallation Tasks for Oracle Grid Infrastructure for a Cluster. Create ASM Disk Groups for Data and Fast Recovery Area. Install Oracle Database 1. Oracle Real Application Clusters Install Oracle Database 1. Examples formerly Companion Create the Oracle Cluster Database. Post Database Creation Tasks OptionalCreate Alter Tablespaces. Verify Oracle Grid Infrastructure and Database Configuration. Starting Stopping the Cluster. Troubleshooting. Conclusion. Acknowledgements. Introduction. One of the most efficient ways to become familiar with Oracle Real Application Clusters RAC 1. Oracle RAC 1. 1g cluster. Theres no better way to understand its benefitsincluding fault tolerance, security, load balancing, and scalabilitythan to experience them directly. Unfortunately, for many shops, the price of the hardware required for a typical production RAC configuration makes this goal impossible. A small two node cluster can cost from US1. US2. 0,0. 00. This cost would not even include the heart of a production RAC environment, the shared storage. In most cases, this would be a Storage Area Network SAN, which generally start at US1. For those who want to become familiar with Oracle RAC 1. Oracle RAC 1. 1g Release 2 system using commercial off the shelf components and downloadable software at an estimated cost of US2,2. US2,7. 00. The system will consist of a two node cluster, both running Oracle Enterprise Linux OEL Release 5 Update 4 for x. Oracle RAC 1. 1g Release 2 for Linux x. ASMLib 2. 0. All shared disk storage for Oracle RAC will be based on i. SCSI using Openfiler release 2. Network Storage Server. Although this article should work with Red Hat Enterprise Linux, Oracle Enterprise Linux available for free will provide the same if not better stability and will already include the ASMLib software packages with the exception of the ASMLib userspace libraries which is a separate download. This guide is provided for educational purposes only, so the setup is kept simple to demonstrate ideas and concepts. For example, the shared Oracle Clusterware files OCR and voting files and all physical database files in this article will be set up on only one physical disk, while in practice that should be configured on multiple physical drives. In addition, each Linux node will only be configured with two network interfaces one for the public network eth. Oracle RAC private interconnect and the network storage server for shared i. SCSI access eth. For a production RAC implementation, the private interconnect should be at least Gigabit or more with redundant paths and only be used by Oracle to transfer Cluster Manager and Cache Fusion related data. A third dedicated network interface eth. Gigabit network for access to the network storage server Openfiler. Oracle Documentation. While this guide provides detailed instructions for successfully installing a complete Oracle RAC 1. Oracle documentation see list below. In addition to this guide, users should also consult the following Oracle documents to gain a full understanding of alternative configuration options, installation, and administration with Oracle RAC 1. Oracles official documentation site is docs. Network Storage Server. Powered by r. Path Linux, Openfiler is a free browser based network storage management utility that delivers file based Network Attached Storage NAS and block based Storage Area Networking SAN in a single framework. The entire software stack interfaces with open source applications such as Apache, Samba, LVM2, ext. Linux NFS and i. SCSI Enterprise Target. Openfiler combines these ubiquitous technologies into a small, easy to manage solution fronted by a powerful web based management interface. Openfiler supports CIFS, NFS, HTTPDAV, FTP, however, we will only be making use of its i. SCSI capabilities to implement an inexpensive SAN for the shared storage components required by Oracle RAC 1. The operating system and Openfiler application will be installed on one internal SATA disk. A second internal 7. GB 1. 5K SCSI hard disk will be configured as a single Volume Group that will be used for all shared disk storage requirements. The Openfiler server will be configured to use this volume group for i. SCSI based storage and will be used in our Oracle RAC 1. Oracle grid infrastructure and the Oracle RAC database. Oracle Grid Infrastructure 1. Release 2 1. 1. 2 With Oracle grid infrastructure 1. Release 2 1. 1. 2, the Automatic Storage Management ASM and Oracle Clusterware software is packaged together in a single binary distribution and installed into a single home directory, which is referred to as the Grid Infrastructure home. You must install the grid infrastructure in order to use Oracle RAC 1. Release 2. Configuration assistants start after the installer interview process that configure ASM and Oracle Clusterware. While the installation of the combined products is called Oracle grid infrastructure, Oracle Clusterware and Automatic Storage Manager remain separate products. After Oracle grid infrastructure is installed and configured on both nodes in the cluster, the next step will be to install the Oracle RAC software on both Oracle RAC nodes. In this article, the Oracle grid infrastructure and Oracle RAC software will be installed on both nodes using the optional Job Role Separation configuration. One OS user will be created to own each Oracle software product grid for the Oracle grid infrastructure owner and oracle for the Oracle RAC software. Throughout this article, a user created to own the Oracle grid infrastructure binaries is called the grid user. This user will own both the Oracle Clusterware and Oracle Automatic Storage Management binaries. The user created to own the Oracle database binaries Oracle RAC will be called the oracle user. Both Oracle software owners must have the Oracle Inventory group oinstall as their primary group, so that each Oracle software installation owner can write to the central inventory ora. Flex. Pod Datacenter with Red Hat Enterprise Linux Open. Stack Platform. Flex. Pod is a pre validated datacenter architecture followed by best practices that is built on the Cisco Unified Computing System UCS, the Cisco Nexus family of switches, and Net. App unified storage systems. Flex. Pod has been a trusted platform for running a variety of virtualization hypervisors as well as bare metal operating systems. The Flex. Pod architecture is highly modular, delivers a baseline configuration, and also has the flexibility to be sized and optimized to accommodate many different use cases and requirements. The Flex. Pod architecture can both scale up adding additional resources within a Flex. Pod unit and scale out adding additional Flex. Pod units. Flex. Pod with Red Hat Enterprise Linux Open. Stack Platform 6. Flex. Pods already wide range of validated and supported design portfolio entries. The audience of this document includes, but is not limited to, sales engineers, field consultants, professional services, IT managers, partner engineers, and customers who want to take advantage of an infrastructure that is built to deliver IT efficiency and enable IT innovation. It is expected from the audience of this document to have the necessary training and background to install and configure Red Hat Enterprise Linux, Cisco Unified Computing System UCS, Cisco Nexus switches, and Net. App storage as well as high level understanding of Open. Stack components. External references are provided where applicable and it is recommended that the audience be familiar with these documents. This document describes the steps required to deploy and configure Red Hat Enterprise Linux Open. Stack Platform 6 on Flex. Pod. The architecture can be very easily expanded with predictable linear performance. While readers of this document are expected to have sufficient knowledge to install and configure the products used, configuration details that are important to this solutions deployments are specifically mentioned. This solution is based on Open. Stack Juno release hardened and streamlined by Red Hat in Red Hat Enterprise Linux Open. Stack Platform 6. In Flex. Pod with Red Hat Enterprise Linux Open. Stack Platform 6, Cisco Unified Computing System, Net. App, and Red Hat Open. Stack Platform are combined to deliver Open. Stack Infrastructure as a Service Iaa. S deployment that is quick and easy to deploy. Flex. Pod with Red Hat Enterprise Linux Open. Stack Platform helps IT organizations accelerate cloud deployments while retaining control and choice over their environments with open and inter operable cloud solutions. Flex. Pod with Red Hat Enterprise Linux Open. Stack Platform 6. Furthermore, it includes Open. Stack HA through redundant controller nodes. In this solution, Open. Stack block, file, and object storage is provided by highly available Net. App storage systems. Flex. Pod is a best practice datacenter architecture that includes these components Cisco Unified Computing System Cisco UCS Cisco Nexus switches Net. App fabric attached storage FAS andor Net. App E Series storage systems. These components are connected and configured according to best practices of both Cisco and Net. App, and provide the ideal platform for running a variety of enterprise workloads with confidence. As previously mentioned, the reference architecture covered in this document leverages the Cisco Nexus 9. Series switch. One of the key benefits of Flex. Pod is the ability to maintain consistency at scaling, including scale up and scale out. Each of the component families shown in Figure 1Cisco Unified Computing System, Cisco Nexus, and Net. App storage systems offers platform and resource options to scale the infrastructure up or down, while supporting the same features and functionality that are required under the configuration and connectivity best practices of Flex. Pod. As customers transition toward shared infrastructure or cloud computing they face a number of challenges such as initial transition hiccups, return on investment ROI analysis, infrastructure management and future growth plan. The Flex. Pod architecture is designed to help with proven guidance and measurable value. By introducing standardization, Flex. Pod helps customers mitigate the risk and uncertainty involved in planning, designing, and implementing a new datacenter infrastructure. The result is a more predictive and adaptable architecture capable of meeting and exceeding customers IT demands. Cisco and Net. App have thoroughly validated and verified the Flex. Pod solution architecture and its many use cases while creating a portfolio of detailed documentation, information, and references to assist customers in transforming their datacenters to this shared infrastructure model. This portfolio includes, but is not limited to the following items Best practice architectural design Workload sizing and scaling guidance Implementation and deployment instructions Technical specifications rules for Flex. Pod configuration dos and donts Frequently asked questions FAQs Cisco Validated Designs CVDs and Net. App Verified Architectures NVAs focused on a variety of use cases. Cisco and Net. App have also built a robust and experienced support team focused on Flex. Pod solutions, from customer account and technical sales representatives to professional services and technical support engineers. The Co operative Support Program extended by Net. App, Cisco and Red Hat provides customers and channel service partners with direct access to technical experts who collaborate with cross vendors and have access to shared lab resources to resolve potential issues. Flex. Pod supports tight integration with virtualized and cloud infrastructures, making it a logical choice for long term investment. The following IT initiatives are addressed by the Flex. Pod solution. Flex. Pod is a pre validated infrastructure that brings together compute, storage, and network to simplify, accelerate, and minimize the risk associated with datacenter builds and application rollouts. These integrated systems provide a standardized approach in the datacenter that facilitates staff expertise, application onboarding, and automation as well as operational efficiencies relating to compliance and certification. Flex. Podis a highly available and scalable infrastructure that IT can evolve over time to support multiple physical and virtual application workloads. Flex. Pod has no single point of failure at any level, from the server through the network, to the storage. The fabric is fully redundant and scalable, and provides seamless traffic failover, should any individual component fail at the physical or virtual layer. Flex. Pod addresses four primary design principles Application availability Makes sure that services are accessible and ready to use. Scalability Addresses increasing demands with appropriate resources. Flexibility Provides new services or recovers resources without requiring infrastructure modifications. Manageability Facilitates efficient infrastructure operations through open standards and APIs. Performance and comprehensive security are key design criteria that are not directly addressed in this solution but have been addressed in other collateral, benchmarking, and solution testing efforts. This design guide validates the functionality and basic security elements.