- Create Virtual IP, done in earlier steps on node1, extend, done in earlier steps
- Create EC2 resource and create as dependency for virtual IP, done in earlier steps
- Install SAP on node1 on “virtual hostname” based on “virtual IP”
- Stopsap on node 1
- Use the LifeKeeper GUI to “In-service” the virtual IP to node 2, and Install SAP on node 2 on “virtual hostname” based on “virtual IP”
- Stopsap on node 2 and modify profile files on both nodes (see below)
- Use the LifeKeeper GUI to “In-service” the virtual IP back to node 1
- Create replication resource for the mount points needed for SAP, done in earlier steps, as advised by SAP consultants
- Startsap on node1 and ensure SAP is working properly
- In /etc/default/LifeKeeper on both nodes add the follow entries to the end:
SAP_EXPERTMODE=1
SAP_DB_CHECK_IGNORE=1
- Re-run the SPS setup program to add the SAP recovery kit
Mount the sps.img file (downloaded as per earlier steps) using the following command:
mount <PATH/IMAGE_NAME> <MOUNT_POINT> -t iso9660 -o loop
Where PATH is the path to the image
IMAGE_NAME is the name of the image
MOUNT_POINT is the path to mount location
Change to the sps.img mounted directory and type the following: ./setup -k
You will now be shown a menu of recovery kits available. Select the recovery kit for SAP by using the arrow keys and pressing the <spacebar> to select, press <enter> to continue and complete the installation.
- Setup NFS servers
- Copy file systems onto the SAP server and create replication resources on the file systems for redundancy and failover
- Create NFS resources following the NFS Recovery Kit guide
Simplified steps are given here
- Create SAP resources following the SAP Recovery Kit guide
Simplified steps are given here
このトピックへフィードバック