VMware® Fault Tolerance (FT) provides continuous availability to virtual machines, eliminating downtime and disruption — even in the event of a complete host failure. This whitepaper gives a brief description of the VMware FT architecture and discusses the performance implication of this feature with data from a wide variety of workloads.
VMware white paper Figure 2. High Level Architecture of VMware Fault Tolerance Primary Secondary FT Logging Tra c ACKs Record Replay Client Shared Storage The communication channel between the primary and the secondary host is established by the hypervisor using a standard TCP/IP socket connection and the traffic flowing between them is called FT logging traffic.
2. Performance aspects and Best Practice recommendations This section describes the performance aspects of Fault Tolerance with best practices recommendations to maximize performance. For operational best practices please refer to the VMware Fault tolerance recommendations and Considerations on VMware vSphere 4 white paper.
VMware white paper To ensure that the secondary virtual machine runs as fast as the primary, it is recommended that: • The hosts in the FT cluster are homogenous, with similar CPU make, model, and frequency. The CPU frequency difference should not exceed 400 MHz. • Both the primary and secondary hosts use the same power management policy. • CPU reservation is set to full for cases where the secondary host could be overloaded. The CPU reservation setting on the primary applies to the secondary as well, so setting full CPU reservation ensures that the secondary gets CPU cycles even when there is CPU contention.
VMware white paper It is recommended that FT primary virtual machines be distributed across multiple hosts and, as a general rule of thumb, the number of FT virtual machines be limited to four per host. In addition to avoiding the possibility of saturating the network link, it also reduces the number of simultaneous live migrations required to create new secondary virtual machines in the event of a host failure.
VMware white paper Figure 4. SPECjbb2005 Performance FT tra c: SPECjbb2005 1.4 Mbits/sec FT Disabled FT Enabled RHEL 5 64-bit, 4GB 3.2. Kernel Compile This experiment shows the time taken to do kernel compile, which is both CPU and MMU intensive workload due to forking of many parallel processes.
VMware white paper 3.3. Netperf Throughput Netperf is a micro-benchmark that measures the throughput of sending and receiving network packets. In this experiment netperf was configured so packets could be sent continuously without having to wait for acknowledgements. Since all the receive traffic needs to be recorded and then transmitted to the secondary, netperf Rx represents a workload with significant FT logging traffic.
VMware white paper Figure 7. Netperf Latency Comparison FT tra c Netperf - latency sensitive case Rx: 500 Mbits/sec 1000 Tx: 36 Mbits/sec FT Disabled FT Enabled Receives Transmits 3.5. Filebench random Disk read/write Filebench is a benchmark designed to simulate different I/O workload profiles. In this experiment, filebench was used to generate random I/Os using 200 worker threads. This workload saturates available disk bandwidth for the given block size. Enabling FT did not...
VMware white paper 3.6. Oracle 11g In this experiment, an Oracle 11g database was driven using the Swingbench Order Entry OLTP (online transaction processing) workload. This workload has a mixture of CPU, memory, disk, and network resource requirements. 80 simultaneous database sessions were used in this experiment. Enabling FT had negligible impact on throughput as well as latency of transactions. Figure 9. Oracle 11g Database Performance (throughput) FT tra c: Oracle Swingbench - Throughput 11 –...
VMware white paper 3.7. Microsoft SQL Server 2005 In this experiment, the DVD Store benchmark was used to drive the Microsoft SQL Server® 2005 database. This benchmark simulates online transaction processing of a DVD store. Sixteen simultaneous user sessions were used to drive the workload. As with the previous benchmark, this workload has a mixture of CPU, memory, disk, and networking resource requirements.
VMware white paper 3.8. Microsoft exchange Server 2007 In this experiment, the Loadgen workload was used to generate load against Microsoft Exchange Server 2007. A heavy user profile with 1600 users was used. This benchmark measures latency of operations as seen from the client machine. The performance charts below report both average latency and 95th percentile latency for various Exchange operations. The generally accepted threshold for acceptable latency is 500 ms for the Send Mail operation. While FT caused a slight increase, the observed SendMail latency was well under 500 ms with and without FT.
Fault Tolerance enabled. 5. Conclusion VMware Fault Tolerance is a revolutionary new technology that VMware is introducing with vSphere. The architecture and design of VMware vLockstep technology allows hardware-style Fault Tolerance on single-CPU virtual machines with minimal impact to performance. Experiments with a wide variety of synthetic and real-life workloads show that the performance impact on throughput...
VMware white paper Swingbench configuration: Swingbench version: 2.2, Calling Circle Database No of orders: 23550492 No of Customers: 864967 Runtime: 30 mins ODBC driver: ojdbc6.jar Driver Type: Thin No of Users: 80 Pooled: 1 LogonDelay: 0 Transaction MinDelay: 50 Transaction MaxDelay: 250 QueryTimeout: 60 Workload Weightage: NewCustomerProcess – 20, BrowseProducts – 50, ProcessOrders – 10, BrowseAndUpdateOrders – 50 Note: Database was restored from backup before every run MSSQL 2005 —...
Total Number of tasks: 107192 (1.24 tasks per second) Notes: • Exchange mailbox database was restored from backup before every run • Microsoft Exchange Search Indexer Service was disabled when the benchmark was run VMware vSphere 4 Fault Tolerance: Architecture and Performance Source: Technical Marketing, SD Revision: 20090811...