Included below are the restrictions or known issues open against LifeKeeper for Linux, broken down by functional area.
| Description |
|---|
|
In Release 7.4 and forward, relocation of the SIOS product RPM packages is no longer supported. |
|
Linux Dependencies Successful completion of the installation of SIOS Protection Suite for Linux, including the optional Recovery Kits, requires the installation of a number of prerequisite packages. Note: Unsuccessful installation of these dependent packages will also impact the ability to start SIOS Protection Suite for Linux along with the ability to load the SIOS Protection Suite for Linux GUI. To prevent script failures, these packages should be installed prior to attempting to run the installation setup script. For a complete list of dependencies, see the Linux Dependencies topic. |
|
The multipathd daemon will log errors in the error log when the nbd driver is loaded as it tries to scan the new devices Solution: To avoid these errors in the log, add devnode "^nbd" to the blacklist in /etc/multipath.conf. |
|
Incomplete NFS Setup Logging When running the Installation setup script from the ISO image ( |
|
If the Workaround: On RHEL, CentOS or Oracle Linux, remove the Example:
|
|
Unexpected termination of daemons Daemons using IPC terminate unexpectedly after update to Red Hat Enterprise Linux 7.2 and Red Hat 7.2 derivative systems. A new systemd feature was introduced in Red Hat Enterprise Linux 7.2 related to the cleanup of all allocated inter-process communication (IPC) resources when the last user session finishes. A session can be an administrative cron job or an interactive session. This behavior can cause daemons running under the same user, and using the same resources, to terminate unexpectedly.
To work around this problem, edit the file /etc/systemd/logind.conf and add the following line:
Then, execute the following command, so that the change is put into effect:
After performing these steps, daemons no longer crash in the described situation. Applications (such as MQ, Oracle, SAP, etc) using shared memory and semaphores may be affected by this issue and therefore require this change. |
| Description |
|---|
|
File system labels should not be used in large configurations The use of file system labels can cause performance problems during boot-up with large clusters. The problems are generally the result of the requirement that to use labels all devices connected to a system must be scanned. For systems connected to a SAN, especially those with LifeKeeper where accessing a device is blocked, this scanning can be very slow. To avoid this performance problem on Red Hat systems, edit /etc/fstab and replace the labels with the path names. |
|
When |
|
DataKeeper Create Resource fails When using DataKeeper in certain environments (e.g., virtualized environments with IDE disk emulation, or servers with HP CCISS storage), an error may occur when a mirror is created:
This is because LifeKeeper does not recognize the disk in question and cannot get a unique ID to associate with the device. Workaround: Add a pattern for our disk(s) to the DEVNAME device_pattern file, e.g.:
|
|
Specifying hostnames for API access The key name used to store LifeKeeper server credentials must match exactly the hostname of the other LifeKeeper server (as displayed by the hostname command on that server). If the hostname is an FQDN, then the credential key must also be the FQDN. If the hostname is a short name, then the key must also be the short name. Workaround: Make sure that the hostname(s) stored by credstore match the hostname exactly. |
|
The use of In the current version of LifeKeeper/SPS, there have been significant enhancements to the logging and other major core components. These enhancements affect the tunables in the Solution: Prior to restoring from an When using lkbackups taken from versions previous to 8.0.0 and restoring on 8.0.0 or later: LKSYSLOGTAG=LifeKeeper LKSYSLOGSELECTOR=local6 Also, when using lkbackups taken from versions previous to 9.0.1 and restoring on 9.0.1 or later: PATH=/opt/LifeKeeper/bin:/usr/java/jre1.8.0_51/bin:/usr/java/bin:/usr/java/jdk1.8.0_51/bin:/bin:/usr/bin:/usr/sbin:/sbin See section on Logging With |
|
Restore of an The configuration files for created resources are saved during an Solution: Restore from |
|
Resources removed in the wrong order during failover In cases where a hierarchy shares a common resource instance with another root hierarchy, resources are sometimes removed in the wrong order during a cascading failover or resource failover. Solution: Creating a common root will ensure that resource removals in the hierarchy occur from the top down.
Note: Using |
|
LifeKeeper syslog EMERG severity messages do not display to a SLES11 host's console which has AppArmor enabled LifeKeeper is accessing Solution: To allow LifeKeeper syslog EMERG severity messages to appear on a SLES11 console with AppArmor enabled, add the following entry to /var/run/utmp kr If added to apparmor_parser -r /etc/apparmor.d/sbin.syslog-ng Verify that the AppArmor update was successful by sending an EMERG syslog entry via: logger -p local6.emerg "This is a syslog/lk/apparmor test." |
|
SIOS strongly discourages the use of RHEL 6.0. If RHEL 6.0 is used, please understand that an OS update may be required to fix certain issues including, but not limited to:
Note: In DataKeeper configurations, if the operating system is updated, a reinstall/upgrade of SIOS Protection Suite for Linux is required. |
|
Delete of nested file system hierarchy generates "Object does not exist" message Solution: This message can be disregarded as it does not create any issues. |
|
When a database has nested file system resources, the file system kit will create the file system for both the parent and the nested child. However, Solution: When multiple file systems are nested within a single mount point, it may be necessary to manually create the additional dependencies to the parent application tag using |
|
DataKeeper: Nested file system create will fail with DataKeeper When creating a DataKeeper mirror for replicating an existing File System, if a file system is nested within this structure, you must unmount it first before creating the File System resource. Workaround: Manually unmount the nested file systems and remount / create each nested mount. |
|
When lkstart is run on SLES 11 SP2, the following insserv message is generated:
LifeKeeper and steeleye-runit scripts are configured by default to start in run level 4 where dependent init script syslog is not. If system run level is changed to 4, syslog will be terminated and LifeKeeper will be unable to log. Solution: Make sure system run level is NOT changed to run level 4. |
|
Changing the mount point of the device protected by Filesystem resource may lead data corruption The mount point of the device protected by LifeKeeper via the File System resource (filesys) must not be changed. Doing so may lead to the device being mounted on multiple nodes and if a switchover is done and this could lead to data corruption. |
|
XFS file system usage may cause quickCheck to fail. In the case CHECK_FS_QUOTAS setting is enabled for LifeKeeper installed on Red Hat Enterprise Linux 7 / Oracle Linux 7 / CentOS 7, quickCheck fails if uquota, gquota option is set to the XFS file system resource, which is to be protected. Solution: Use usrquota, grpquota instead of uquota, gquota for mount options of XFS file system, or, disable CHECK_FS_QUOTAS setting. |
|
Btrfs is not supported Btrfs cannot be used for LifeKeeper files (/opt/LifeKeeper), bitmap files if they are not in /opt/LifeKeeper, lkbackupfiles, or any other LifeKeeper related files. In addition, LifeKeeper does not support protecting Btrfs within a resource hierarchy. |
|
SLES12 SP1 or later on AWS The following restrictions apply with SLES12 SP1 or later on AWS:
|
|
Shutdown Strategy set to "Switchover Resources" may fail when using Quorum/Witness Kit in Witness mode
Hierarchy switchover during LifeKeeper shutdown may fail to occur when using the Quorum/Witness Kit in Witness mode.
|
|
Edit /etc/service If the following entry in /etc/service is deleted, LifeKeeper cannot start up. lcm_server 7365/tcpDon't delete this entry when editing the file. |
|
Any storage unit which returns a string including a space for the SCSI ID cannot be protected by LifeKeeper. |
|
Automatic recovery of network may fail when link status goes down on Red Hat 5.x and Red Hat 6.x systems when using bonded interfaces. The network doesn't recover automatically when using bonded interfaces when the link status is lost and then restored on Red Hat 5.x or Red Hat 6.x. Loss of the link status can occur via a bad network cable, bad switch or hub or a cable reconnection and ifdown -> ifup. To recover from this status, restart the network manually by executing the following command by a user with root authorization.
# service network restart Please note that this problem is already corrected for RHEL7. Also, this problem doesn't occur with SLES. |
| Description |
|---|
|
/etc/hosts settings dependency /etc/hosts settings:
and the listed internet hostid is correct, then the configuration of /etc/hosts may be the cause. To correctly match /etc/hosts entries, IPv4 entries must be listed before any IPv6 entries. To verify if the /etc/hosts configuration is the cause, run the following command: /opt/LifeKeeper/bin/lmutil lmhostid -internet -n If the IPv4 address listed does not match the IPv4 address in the installed license file, then /etc/hosts must be modified to place IPv4 entries before IPv6 entries to return the correct address. |
| Description |
|---|
|
SIOS has migrated to the use of the ip command and away from the ifconfig command. Because of this change, customers with external scripts are advised to make a similar change. Instead of issuing the ifconfig command and parsing the output looking for a specific interface, scripts should instead use "ip -o addr show" and parse the output looking for lines that contain the words "inet" and "secondary".
So for the above output from the ip command, the following lines contain virtual IP addresses for the eth0 interface:
|
|
'IPV6_AUTOCONF = No' for /etc/sysconfig/network-scripts/ifcfg-<nicName> is not being honored on reboot or boot On boot, a stateless, auto-configured IPv6 address is assigned to the network interface. If a comm path is created with a stateless IPv6 address of an interface that has IPV6_AUTOCONF=No set, the address will be removed if any system resources manage the interface, e.g. ifdown <nicName>;ifup <nicName>. Comm path using auto-configured IPv6 addresses did not recover and remained dead after rebooting primary server because IPV6_AUTOCONF was set to No. Solution: Use Static IPv6 addresses only. The use of auto-configured IPv6 addresses could cause a comm loss after a reboot, change NIC, etc. While IPv6 auto-configured addresses may be used for comm path creation, it is incumbent upon the system administrator to be aware of the following conditions:
|
|
IP: Modify Source Address Setting for IPv6 doesn't set source address When attempting to set the source address for an IPv6 IP resource, it will report success when nothing was changed. Workaround: Currently no workaround is available. This will be addressed in a future release. |
|
IP: Invalid IPv6 addressing allowed in IP resource creation Entering IPv6 addresses of the format 2001:5c0:110e:3368:000000:000000001:61:14 is accepted when the octets contain more than four characters. Workaround: Enter correctly formatted IPv6 addresses. |
|
Can't connect to host via IPv6 addressing lkGUIapp will fail connecting to a host via IPv6 hex addressing, either via resolvable host name or IP address. lkGUIapp requires an IPv4 configured node for connection. IPv6 comm paths are fully supported. |
|
IPv6 resource reported as ISP when address assigned to bonded NIC but in 'tentative' state IPv6 protected resources in LifeKeeper will incorrectly be identified as 'In Service Protected' (ISP) on SLES systems where the IPv6 resource is on a bonded interface, a mode other than 'active-backup' (1) and Linux kernel 2.6.21 or lower. The IPv6 bonded link will remain in the 'tentative' state with the address unresolvable. Workaround: Set the bonded interface mode to 'active-backup' (1) or operate with an updated kernel which will set the link state from 'tentative' to 'valid' for modes other than 'active-backup' (1). |
| Description |
|---|
|
Apache Kit does not support IPv6; doesn't indentify IPv6 in httpd.conf Any IPv6 addresses assigned to the 'Listen' directive entry in the httpd.conf file will cause problems. Solution: Until there is support for IPv6 in the Apache Recovery Kit, there can be no IPv6 address in the httpd.conf file after the resource has been created. |
| Description |
|---|
|
The Oracle Recovery Kit does not include support for Connection Manager and Oracle Names features The LifeKeeper Oracle Recovery Kit does not include support for the following Oracle Net features of Oracle: Oracle Connection Manager, a routing process that manages a large number of connections that need to access the same service; and Oracle Names, the Oracle-specific name service that maintains a central store of service addresses. The LifeKeeper Oracle Recovery Kit does protect the Oracle Net Listener process that listens for incoming client connection requests and manages traffic to the server. Refer to the LifeKeeper for Linux Oracle Recovery Kit Administration Guide for LifeKeeper configuration specific information regarding the Oracle Listener. |
|
The Oracle Recovery Kit does not support the ASM or grid component features of Oracle 10g The following information applies to Oracle 10g database instances only. The Oracle Automatic Storage Manager (ASM) feature provided in Oracle 10g is not currently supported with LifeKeeper. In addition, the grid components of 10g are not protected by the LifeKeeper Oracle Recovery Kit. Support for raw devices, file systems, and logical volumes are included in the current LifeKeeper for Linux Oracle Recovery Kit. The support for the grid components can be added to LifeKeeper protection using the gen/app recovery kit. |
|
The Oracle Recovery Kit does not support NFS Version 4 The Oracle Recovery Kit supports NFS Version 3 for shared database storage. NFS Version 4 is not supported at this time due to NFSv4 file locking mechanisms. |
|
Oracle listener stays in service on primary server after failover Network failures may result in the listener process remaining active on the primary server after an application failover to the backup server. Though connections to the correct database are unaffected, you may still want to kill that listener process. |
| The Oracle Recovery Kit does not support Oracle Database Standard Edition 2 (SE2) on AWS EC2 system
During Oracle Database Standard Edition 2 (SE2) test, an unknown behavior was reported on AWS EC2 system. However, we confirmed that other services except EC2 did not have the same issue, meaning that Oracle Recovery Kit supports SE2 on systems except EC2. |
| Description |
|---|
|
The "include" directive is not supported The "include" directive is not supported. All the setup configuration information must be described in a single my.cnf file. |
|
Crash Recovery Restarting MySQL after an abnormal termination initiates a MySQL crash recovery. While in this recovery state MySQL client connections are denied. This will prevent LifeKeeper from checking the state of MySQL causing a possible failover to the standby node. |
| Description |
|---|
|
Top level NFS resource hierarchy uses the switchback type of the hanfs resource The switchback type, which dictates whether the NFS resource hierarchy will automatically switch back to the primary server when it comes back into service after a failure, is defined by the hanfs resource. |
|
Some clients are unable to reacquire nfs file locks When acting as NFS clients, some Linux kernels do not respond correctly to notifications from an NFS server that an NFS lock has been dropped and needs to be reacquired. As a result, when these systems are the clients of an NFS file share that is protected by LifeKeeper, the NFS locks held by these clients are lost during a failover or switchover. When using storage applications with locking and following recommendations for the NFS mount options, SPS requires the additional |
|
NFS v4 changes not compatible with SLES 11 nfs subsystem operation The mounting of a non-NFS v4 remote export on SLES 11 starts rpc.statd. The start up of rpc.statd on the out of service node in a cluster protecting an NFS v4 root export will fail. Solution: Do not mix NFS v2/v3 with a cluster protecting an NFS v4 root export. |
|
NFS v4 cannot be configured with IPv6 IPv6 virtual IP gets rolled up into the NFSv4 heirarchy. Solution: Do not use an IPv6 virtual IP resource when creating an NFSv4 resource. |
|
NFS v4: Unable to re-extend hierarchy after unextend Extend fails because export point is already exported on the target server. A re-extend to server A of a NFS v4 hierarchy will fail if a hierarchy is created on server A and extended to server B, brought in service on server B and then unextended from server A. Solution: On server A run the command "exportfs -ra" to clean up the extra export information left behind. |
|
File Lock switchover with NFSv3 fails on some operating systems Failover file locks with NFSv3 during resources switchover/failover does not work on the operating systems listed below. Lock failover with NFSv3 is currently not supported on these OS versions.
Solution: Use the lock failover features available with NFSv4. |
|
The Oracle Recovery Kit does not support NFSv4 The Oracle Recovery Kit supports NFSv3 for shared database storage. NFSv4 is not supported at this time due to NFSv4 file locking mechanisms. |
|
Stopping and starting NFS subsystem adversely impacts SIOS Protection Suite protected NFS exports. If the NFS subsystem ( Workaround: Do not stop the NFS subsystem while the SIOS Protection Suite NFS ARK is actively protecting exported directories on the system. If the NFS subsystem must be stopped, all NFS resources should be switched to the standby node before stopping the NFS subsystem. Use of the |
| Description |
|---|
|
Failed delete or unextend of a SAP hierarchy Deleting or unextending a SAP hierarchy that contains the same IP resource in multiple locations within the hierarchy can sometimes cause a core dump that results in resources not being deleted. To correct the problem, after the failed unextend or delete operation, manually remove any remaining resources using the LifeKeeper GUI. You may also want to remove the core file from the server. |
|
Handle Warnings gives a syntax error at -e line 1 When changing the default behavior of No in Handle Warnings to Yes, an error is received. Solution: Leave this option at the default setting of No. Note: It is highly recommended that this setting be left on the default selection of No as Yellow is a transient state that most often does not indicate a failure. |
|
Choosing same setting causes missing button on Update Wizard If user attempts to update the Handle Warning without changing the current setting, the next screen, which indicates that they must go back, is missing the Done button. |
|
When changes are made to res_state, monitoring is disabled If Protection Level is set to BASIC and SAP is taken down manually (i.e. for maintenance), it will be marked as FAILED and monitoring will stop. Solution: In order for monitoring to resume, LifeKeeper will need to start up the resource instead of starting it up manually. |
|
ERS in-service fails on remote host if ERS is not parent of Core/CI Creating an ERS resource without any additional SAP resource dependents will cause initial in-service to fail on switchover. Solution: Create ERS as parent of CI/Core instance (SCS or ASCS), then retry in-service. |
| Description |
|---|
|
Important reminder about DataKeeper for Linux asynchronous mode in an LVM over DataKeepr configuration Kernel panics may occur in configurations were LVM resources sit above multiple asynchronous mirrors. In these configurations data consistency may be an issue if a panic occurs. Therefore the required configurations are a single DataKeeper mirror or multiple synchronous DataKeeper mirrors. |
|
Use of lkID incompatible with LVM overwritten on entire disk When lkID is used to generate unique disk IDs on disks that are configured as LVM physical volumes, there is a conflict in the locations in which the lkID and LVM information is stored on the disk. This causes either the lkID or LVM information to be overwritten depending on the order in which lkID and pvcreate are used. Workaround: When it is necessary to use lkID in conjunction with LVM, partition the disk and use the disk partition(s) as the LVM physical volume(s) rather than the entire disk. |
|
LVM actions slow on RHEL 6 When running certain LVM commands on RHEL 6, performance is sometimes slower than in previous releases. This can be seen in slightly longer restore and remove times for hierarchies with LVM resources. |
|
The configuration of Raw and LVM Recovery Kits together is not supported in RHEL 6 environment When creating a Raw resource, the Raw Recovery Kit is looking for a device file based on major # and minor # of Raw device. As the result, |
| Description |
|---|
|
Multipath Recovery Kits (DMMP / HDLM / PPATH / NECSPS): Registration conflict error occurs on lkstop when resource OSF The multipath recovery kits (DMMP, HDLM, PPATH, NECSPS) can have a system halt occur on the active (ISP) node when LifeKeeper is stopped on the standby (OSU) node if the Multipath resource state is OSF. Workarounds: a) Switch the hierarchy to the standby node before LifeKeeper is stopped OR b) Run ins_setstate on the standby node and set the Multipath resource state to OSU before LifeKeeper is stopped |
| Description |
|---|
|
DMMP: Write issued on standby server can hang If a write is issued to a DMMP device that is reserved on another server, then the IO can hang indefinitely (or until the device is no longer reserved on the other server). If/when the device is released on the other server and the write is issued, this can cause data corruption. The problem is due to the way the path checking is done along with the IO retries in DMMP. When "no_path_retry" is set to 0 (fail), this hang will not occur. When the path_checker for a device fails when the path is reserved by another server (MSA1000), then this also will not occur. Workaround: Set "no_path_retry" to 0 (fail). However, this can cause IO failures due to transient path failures. |
|
DMMP: Multiple initiators are not registered properly for SAS arrays that support ATP_C LifeKeeper does not natively support configurations where there are multiple SAS initiators connected to a SAS array. In these configurations, LifeKeeper will not register each initiator correctly, so only one initiator will be able to issue IOs. Errors will occur if the multipath driver (DMMP for example) tries to issue IOs to an unregistered initiator. Solution: Set the following tunable in
|
|
LifeKeeper on RHEL 6.0 cannot support reservations connected to an EMC Clariion |
|
Two or more different storage can not be used concurrently in case of the parameter configuration of DMMP recovery kit is required for some storage model. |
|
DMMP RK doesn't function correctly if the disk name ends with "p<number>". The DMMP RK doesn't function correctly if the disk name ends with "p<number>". Workaround: Do not create disk names ending in "p<number>". |
| Description |
|---|
|
DB2 Recovery Kit reports unnecessary error If DB2 is installed on a shared disk, the following message may be seen when extending a DB2 resource.
This message will not adversely affect the behavior of the DB2 resource extend. |
| Description |
|---|
|
MD Kit does not support mirrors created with “homehost” The LifeKeeper MD Recovery Kit will not work properly with a mirror created with the "homehost" feature. Where "homehost" is configured, LifeKeeper will use a unique ID that is improperly formatted such that in-service operations will fail. On SLES 11 systems, the “homehost” will be set by default when a mirror is created. The version of mdadm that supports “homehost” is expected to be available on other distributions and versions as well. When creating a mirror, specify --homehost="" on the command line to disable this feature. If a mirror already exists that has been created with the “homehost” setting, the mirror must be recreated to disable the setting. If a LifeKeeper hierarchy has already been built for a mirror created with “homehost”, the hierarchy must be deleted and recreated after the mirror has been built with the “homehost” disabled. |
|
MD Kit does not support MD devices created on LVM devices The LifeKeeper MD Recovery Kit will not work properly with an MD device created on an LVM device. When the MD device is created, it is given a name that LifeKeeper does not recognize. |
|
MD Kit configuration file entries in /etc/mdadm.conf not commented out The LifeKeeper configuration file entries in /etc/mdadm.conf should be commented out after a reboot. These file entries are not commented out. |
|
Local recovery not performed in large configurations In some cases with large configurations (6 or more hierarchies), if a local recovery is triggered ( |
|
Mirrors automatically started during boot On some systems (for example those running RHEL 6), there is an AUTO entry in the configuration file (/etc/mdadm.conf) that will automatically start mirrors during boot (example: AUTO +imsm +1.x –all). Solution: Since LifeKeeper requires that mirrors not be automatically started, this entry will need to be edited to make sure that LifeKeeper mirrors will not be automatically started during boot. The previous example (AUTO +imsm +1.x –all) is telling the system to automatically start mirrors created using imsm metadata and 1.x metadata minus all others. This entry should be changed to "AUTO -all", telling the system to automatically start everything “minus” all; therefore, nothing will be automatically started. Important: If system critical resources (such as root) are using MD, make sure that those mirrors are started by other means while the LifeKeeper protected mirrors are not. |
| Description |
|---|
|
Extend fails when MaxDB is installed and configured on shared storage. Workaround: Install a local copy of MaxDB on the backup node(s) in the same directories as the primary system. |
| Description |
|---|
|
User Name/Password Issues:
|
|
Resource Create Issue:
|
|
Extend Issues:
|
|
Properties Page Issues:
|
|
Sybase Monitor server is not supported in 15.7 or later with SIOS Protection Suite. If the Sybase Monitor server process is configured in Sybase 15.7 or later, you must use a Generic Application (gen/app) resource to protect this server process. |
|
Remove command not recognized by Sybase on SLES 11 If bringing Sybase in service on the backup server, it must first be taken out of service on the primary server. In order for Sybase to recognize this command, you must add a line in " Example:
|
|
Unable to detect that Sybase ARK is running Symptom: Unable to detect that Sybase ARK is running. |
| Description |
|---|
|
Error when lksupport command is executed: The following error can be output when lksupport command is executed in the case MQ queue manager protected by MQ RK is set on the disk shared by NFS.
This happens because the root access to NFS area is prohibited. This error output doesn't cause any problem. |
|
Quickcheck fails if queue has long messages if a message is in the test queue of size > 101 characters the put/get fails and the queue fills up |
|
Install fails if the only installed MQ is a relocated install (non-standard and not likely) |
|
Package install fails if the MQ package does not have the default name |
|
Compile samples fails if the software is not installed under /opt/mqm |
|
If two listeners are defined for a single instance and one is set to manual and the other is automatic failures can occur in create and quickCheck |
| Description |
|---|
|
Unexpected failover may occur when the operating system shutting down. Workaround: Stop LifeKeeper by lkstop before stopping the OS. Or perform out-of-service for all resources, then you can confirm there is no ISP resources, then do shutdown the OS. |
© 2017 SIOS Technology Corp., the industry's leading provider of business continuity solutions, data replication for continuous data protection.