Version 11.5, 2010-08-11
These release notes apply to all SUSE Linux Enterprise Server 11 SP1 High Availability Extension based products (e.g. for
x86, x86_64, Itanium, Power and System z). Some
sections may not apply to a particular architecture/product. Where this is
not obvious, the respective architectures are listed explicitly in these notes.
Instructions for installing SUSE Linux Enterprise Server 11 SP1 High Availability Extension can be found in the
README file on the CD.
Manuals can be found in the
docu directory of the
installation media. Any documentation (if installed) can be found in the
/usr/share/doc/ directory of the installed system.
This Novell product includes materials licensed to Novell under the GNU General Public License (GPL). The GPL requires that Novell make available certain source code that corresponds to those GPL-licensed materials. The source code is available for download at http://www.novell.com/linux/source. Also, for up to three years from Novell's distribution of the Novell product, Novell will mail a copy of the source code upon request. Requests should be sent by e-mail to firstname.lastname@example.org or as otherwise instructed at http://www.novell.com/linux/source. Novell may charge a fee to recover its reasonable costs of distribution.
SUSE Linux Enterprise Server 11 SP1 High Availability Extension is an affordable, integrated suite of robust open source clustering technologies that enable enterprises to implement highly available Linux clusters and eliminate single points of failure.
Used with SUSE Linux Enterprise Server 11 SP1, it helps firms maintain business continuity, protect data integrity, and reduce unplanned downtime for their mission-critical Linux workloads.
SUSE Linux Enterprise Server 11 SP1 High Availability Extension provides all of the essential monitoring, messaging, and cluster resource management functionality of proprietary third-party solutions, but at a more affordable price, making it accessible to a wider range of enterprises.
It is optimized to work with SUSE Linux Enterprise Server 11 SP1, and its tight integration ensures customers have the most robust, secure, and up to date, high availability solution. Based on an innovative, highly flexible policy engine, it supports a wide range of clustering scenarios.
With static or stateless content, the High Availability cluster can be used without a cluster file system. This includes web-services with static content as well as printing systems or communication systems like proxies that do not need to recover data.
Finally, it's open source license minimizes the risk of vendor lock-in, and it's adherence to open standards encourages interoperability with industry standard tools and technologies.
In Service Pack 1, a large number of improvements have been added, some of which are called out explicitly here. For the full list of changes and bugfixes, please refer to the change logs of the RPM packages.
This section includes an overview of some of the major features and new functionality provided by SUSE Linux Enterprise Server 11 SP1 High Availability Extension.
Cluster File System - Oracle Cluster File System 2 (OCFS2)
Cluster file systems are used to provide scalable, high performance, and highly available file access across multiple instances of SUSE Linux Enterprise Server 11 SP1 High Availability Extension servers. Oracle Cluster File System 2 (OCFS2) is a POSIX-compliant shared-disk cluster file system for Linux. OCFS2 is developed under a GPL open source license.
New features included in OCFS2 with this product release are:
copy-on-write clones of files (reflink), particularly useful for cloning virtual machine images and/or taking consistent backups
indexed directories, delivering high performance regardless of number of files per directory
Meta data checksumming, detects all on-disk corruption and capable of correcting some errors transparently
Improved performance for deletion
Improved allocation algorithms reduce fragmentation for large files
Beyond this, OCFS2 continues to deliver the functionality provided in the previous releases:
Access Control Lists (ACL)
POSIX conforming file locking
expand file system during operation
With these features OCFS2 can be used as generic file system for common use without previous limitations to specific workloads. Workloads for OCFS2 included in this product are, but are not limited to:
central storage area for virtual machine images
central storage area for file servers
shared file system for High Availability
all applications using a cluster file system (e.g. Tibco)
The full functionality of OCFS2 is only available in combination with the OpenAIS, corosync, and pacemaker-based cluster stack.
Clustered Logical Volume Manager2 - cLVM2
The Clustered Volume Manager allows multiple nodes to read and write volumes on a single storage device at the block layer level. It features creation and reallocation of volumes on a shared storage infrastructure like SAN or iSCSI, and allows moving volumes to a different storage device during operation. It can be used for volume snapshots for later recovery if needed.
New features included in cLVM2 with this product release are:
cluster-concurrent mirrored logical volumes
For clustered volume groups consisting at least two physical volumes, the administrator can specify that a logical volume should be mirrored across them for added redundancy. Please see the documentation for more details.
Please also see the known issues section.
pvmove for clustered volume groups
To ease migration from old to new storage, physical volumes can now also be migrated transparently. The new physical volume can be larger or smaller (if the origin is not fully utilized) than the origin.
High Availability Cluster Resource Manager (Pacemaker)
Pacemaker orchestrates the cluster's response to change events such as node failures, resource monitoring failures, permanent or transient administrative changes, and ensures that service availability is recovered.
New features introduced by Pacemaker and included with this product release are:
Utilization-based resource placement
Full support for cloned groups
Ability to explore effects of cluster events and configuration changes in the CRM shell
High-Availability Web Konsole (hawk) provides web-based cluster status and management
Support for restricting access to the cluster configuration (ACLs).
With unified command line support system setup, managing and integration is made easier. To extend High Availability to all types of applications, resource agent templates and templates for configuration examples are provided for customization.
Cluster and Distributed Systems infrastructure (corosync and OpenAIS
The Corosync Cluster Engine is an OSI certified implementation of a complete cluster engine. This component provides membership, ordered messaging with virtual synchrony guarantees, closed process communication groups, and an extensible framework.
The OpenAIS Standards Based Cluster Framework is an OSI Certified implementation of the Service Availability Forum Application Interface Specification (AIS), built on top of Corosync. These components are in turn used by Pacemaker, OCFS2, DLM and others.
Both components have been updated in SUSE Linux Enterprise Server 11 SP1 High Availability Extension and now provide the following AIS levels: AMF B.01.02, TMR A.01.01, CKPT B.01.01, CLM B.01.01, EVT B.01.01, LCK B.03.01, MSG B.03.01.
Data replication - Distributed Remote Block Device (DRBD)
Data replication is part of a disaster prevention strategy in most large enterprises. Using network connections data is replicated between different nodes to ensure consistent data storages in case of a site failure.
Data replication is provided in SUSE Linux Enterprise Server 11 SP1 High Availability Extension with DRBD. This software based data replication allows customers to use different types of storage systems and communication layers without vendor lock-in. At the same time, data replication is deeply integrated into the operating system and thus provide ease-of-use. Features related to data replication and included with this product release are:
YaST2 setup tools to assist initial setup
Fully synchronous, memory synchronous or asynchronous modes of operation
Differential storage resynchronization after failure
Bandwidth of background resynchronization tunable
Shared secret to authenticate the peer upon connect
Configurable handler scripts for various DRBD events
Online data verification
With these features data replication can be easier configured and used. And with improved storage resynchronization recovery times will be decreased significantly.
The distributed replicated block device (DRBD) version included supports active/active mirroring, enabling the use of services such as cLVM2 or OCFS2 on top.
IP Load Balancing - Linux Virtual Server (LVS)
Linux Virtual Server (LVS) is an advanced IP load balancing solution for Linux. IP load balancing provides a high-performance, scalable network infrastructure. Such infrastructure is typically used by enterprise customers for webservers or other network related service workloads.
With LVS network requests can be spread over multiple nodes to scale the available resources and balance the resulting workload. By monitoring the compute nodes, LVS can handle node failures and redirect requests to other nodes maintaining the availability of the service.
Relax and Recover (ReaR)
New in SUSE Linux Enterprise Server 11 SP1 High Availability Extension
On the x86 and x86-64 architectures, a disaster recovery framework is included. ReaR allows the administrator to take a full snaphot of the system and restore this snapshot after a disaster on recovery hardware.
Distributed Lock Manager (DLM)
The DLM in SUSE Linux Enterprise Server 11 SP1 High Availability Extension supports both TCP and SCTP for network communications, allowing for improved cluster redundancy in scenarios where network interface bonding is not feasible.
pacemaker-pygui rename and split
In response to customer demand, the Python-based GUI component (formerly packaged as pacemaker-pygui) has been split into a server and client package, allowing server installs without client software.
The new packages are called pacemaker-mgmt for the server, and pacemaker-mgmt-client for the client.
After an update from GA, the client package may not
automatically be installed, depending on installer settings, and thus
The resolution is to install the package manually.
New packages from SP1 not installed automatically on update
New functionality provided via new packages is not automatically installed by an update, which strives to preserve the existing functionality. It is recommended to install the HA pattern manually to take advantage of all new functionality in SUSE Linux Enterprise Server 11 SP1 High Availability Extension.
Non-production fencing/STONITH agents moved
null STONITH agents have been moved to
libglue-devel package, and are no longer
installed by default.
These fencing agents are not suitable for production environments and should only be used for limited functionality demo setups. This move clarifies their intended use case.
OCFS2's O2CB stack.
The legacy O2CB in-kernel stack of OCFS2 is only supported in combination with Oracle RAC. Oracle RAC, due to its technical limitations, cannot be combined with the pacemaker-based cluster stack.
Samba Clustered Trivial Database (ctdb)
SUSE Linux Enterprise Server 11 SP1 High Availability Extension includes the Samba CTDB extension, including an OCF-compliant resource agent to orchestrate fail-over. This is fully supported, together with exporting Samba CTDB from OCFS2.
Due to technical limitations, this also includes the CTDB internal fail-over functionality for IP address take-over. Please note that this part is not supported by Novell. Only Pacemaker clusters are fully supported.
The smb_private_dir parameter for the CTDB resource agent is now deprecated and has been made optional. Existing installations using CTDB should remove this parameter from their configuration at their next convenience.
Several new parameters have been added to the CTDB resource agent in this release - run "crm ra info CTDB" for details. Two of these parameters, ctdb_manages_samba and ctdb_manages_winbind, default to "yes" for compatibility with the previous releases. Existing installations should update their configuration to explicitly set these parameters to "yes", as the defaults will be changed to "no" in a future release.
DRBD resource agent
The new version of DRBD included in SUSE Linux Enterprise Server 11 SP1 High Availability Extension also supplies a new, updated Open Clustering Framework resource agent from the provider linbit.
It is recommended that setups are converted from ocf:heartbeat:drbd to use the new ocf:linbit:drbd agent. Some new features, such as dual-primary support for master resources, is only available in the new version.
Whereas SUSE Linux Enterprise Server 10 clusters utilized heartbeat as the cluster infrastructure layer, providing messaging and membership services, SUSE Linux Enterprise 11 High-Availability Extension uses corosync and openais. heartbeat is no longer included with the product.
Please use the
hb2openais.sh tool for migrating
your SUSE Linux Enterprise Server 10 environment to SUSE Linux Enterprise Server 11 SP1 High Availability Extension.
EVMS2 replaced with LVM2
Since EVMS2 has been depreciated in SUSE Linux Enterprise Server 11 SP1, the clustered extensions are also no longer available in SUSE Linux Enterprise Server 11 SP1 High Availability Extension. A conversion tool is supplied as part of the lvm2-clvm package. After the conversion, the former C-EVMS2 segments can be used as regular, full-featured LVM2 logical volumes.
SUSE Linux Enterprise Server 11 SP1 High Availability Extension is supported in the following environments and deployment scenarios. This list is not exhaustive, but should serve as an initial guide. If your setup is not explicitly covered, please contact Novell for assistance.
Local Data Center setups
A local data center setup is characterized by all of the upto 32 cluster nodes being physically connected to the same two or more switches, with a fencing mechanism available.
SUSE Linux Enterprise Server 11 SP1 High Availability Extension further supports stretched cluster setups, where a single cluster spawns a network topology of no more than 5ms round-trip latency total, insignificant message loss, redundant communication between all peers.
SUSE Linux Enterprise Server 11 SP1 High Availability Extension does not currently support mixed big and little endian clusters.
KVM/qemu and VNC
It is recommended to assign and configure the VNC port
statically and not to use
This will also ensure that guests keep the same port number on
restart, migration, or fail-over, which greatly simplifies
Linux Virtual Server tunnelling support
The LVS TCP/UDP load balancer currently only works with Direct Routing and NAT setups. IP-over-IP tunnelling forwarding to the real servers does not currently work.
Cluster-concurrent RAID1 resynchronization
To ensure data integrity, a full RAID1 resync is triggered when a device is re-added to the mirror group. This can impact performance, and it is thus advised to use multipath IO to reduce exposure to mirror loss.
Due to the need of the cluster to keep both mirrors uptodate and consistent on all nodes, a mirror failure on one node is treated as if the failure had been observed cluster-wide, evicting the mirror on all nodes. Again, multipath IO is recommended to reduce this risk.
In situations where the primary focus is on redundancy and
not on scale-out, building a storage target node (using
md raid1 in a fail-over configuration or
drbd) and reexporting via iSCSI, NFS,
or CIFS could be a viable option.
Quotas on OCFS2 filesystem
To use quotas on ocfs2 filesystem, the filesystem has to be created with appropriate quota features: 'usrquota' filesystem feature is needed for accounting quotas for individual users, 'grpquota' filesystem feature is needed for accounting of quotas for groups. These features can be also enabled later on an unmounted filesystem using tunefs.ocfs2.
For quota-tools to operate on the filesystem, you have to mount the filesystem with 'usrquota' (and/or 'grpquota') mount option.
When a filesystem has appropriate quota feature enabled, it maintains in its metadata how much space and files each user (group) uses. Since ocfs2 treats quota information as a filesystem internal metadata, there is no need to ever run quotacheck(8) program. Instead, all the needed functionality is built into fsck.ocfs2 and the filesystem driver itself.
To enable enforcement of limits imposed on each user / group, run quotaon(8) program similarly as for any other filesystem.
Commands quota(1), setquota(8), edquota(8) work as usual with ocfs2 filesystem. Commands repquota(8) and warnquota(8) do not work with ocfs2 because of a limitation in the current kernel interface.
For performance reasons each cluster node performs quota accounting locally and synchronizes this information with a common central storage once per 10 seconds (this interval is tunable by tunefs.ocfs2 using options 'usrquota-sync-interval' and 'grpquota-sync-interval'). Thus quota information need not be exact at all times and as a consequence user / group can slightly exceed their quota limit when operating on several cluster nodes in parallel.
Resource sets in the Pacemaker CIB
Resource sets can be used to express M:N relations in dependencies, thereby greatly compacting and thus simplifying the configuration.
Note that at this time, resource sets may not include clones or master/slave resource types, but only work for primitive resources.
Support requires an appropriate subscription from Novell; for more information, please see: http://www.novell.com/products/server/services_support.html.
General Support Statement
The following definitions apply:
L1: Installation and problem determination - technical support designed to provide compatibility information, installation and configuration assistance, usage support, on-going maintenance and basic troubleshooting. Level 1 Support is not intended to correct product defect errors.
L2: Reproduction of problem isolation - technical support designed to duplicate customer problems, isolate problem areas and potential issues, and provide resolution for problems not resolved by Level 1 Support.
L3: Code Debugging and problem resolution - technical support designed to resolve complex problems by engaging engineering in patch provision, resolution of product defects which have been identified by Level 2 Support.
Novell will only support the usage of original (unchanged or not recompiled) packages.
Read the READMEs on the CDs.
Get detailed changelog information about a particular package from the RPM:
rpm --changelog -qp <FILENAME>.rpm
<FILENAME>. is the name of the RPM.
ChangeLog file in the top level of CD1 for
a chronological log of all changes made to the updated packages.
Find more information in the
of CD1 of the SUSE Linux Enterprise Server 11 SP1 High Availability Extension CDs. This directory includes PDF versions
of the SUSE Linux Enterprise Server 11 SP1 High Availability Extension startup and preparation guides.
http://www.novell.com/documentation/sles11/ contains additional or updated documentation for SUSE Linux Enterprise Server 11 SP1 High Availability Extension.
Visit http://www.novell.com/linux/ for the latest Linux product news from SUSE/Novell and http://www.novell.com/linux/source/ for additional information on the source code of SUSE Linux Enterprise products.
Thanks for using SUSE Linux Enterprise Server in your business.
The SUSE Linux Enterprise 11 Team.