![]() | Sun System Handbook - ISO 4.1 October 2012 Internal/Partner Edition | ||
|
|
![]() |
||||||||||||||
Solution Type Technical Instruction Sure Solution 1472716.1 : Sun ZFS Storage Appliance: How to set up and use Analytics for benchmarking applications
This document describes which ZFS Storage Appliance Analytics to enable in performance-related benchmarks or proofs of concept (POC) and how to interpret the results of those benchmarks. This document also includes references to the client-side accounting tools that should also be used to monitor the client system during benchmarking or performance testing. In this Document
Applies to:Sun ZFS Storage 7420 - Version Not Applicable to Not Applicable [Release N/A]Sun ZFS Storage 7320 - Version Not Applicable to Not Applicable [Release N/A] Sun ZFS Storage 7120 - Version Not Applicable to Not Applicable [Release N/A] Sun ZFS Backup Appliance - Version Not Applicable to Not Applicable [Release N/A] Information in this document applies to any platform. GoalThis document provides the reader with the minimum level of instrumentation to execute performance test and benchmarks with the ZFS Storage Appliance or ZFS Backup Appliance. FixDetailed performance monitoring of the ZFS Storage Appliance for general use cases can be accomplished by first enabling Advanced Analytics (Configuration->Preferences, check the Enable Advanced Analytics box) and then enabling the following Analytics:
Details for interpreting each accounting statistics are available from the ZFS Storage Appliance and ZFS Backup Appliance browser interface (BUI) help menu: Help->Analytis->Statistics. The client operating system accessing the ZFS Storage Appliance or ZFS Backup Appliance instrumentation should also be enabled and recorded during performance evaluations or benchmarking exercises. For Solaris begin with the following instrumentation (please note the 5 second timing is a good first suggestion, but specific circumstances may dictact a different value):
For the Linux operating system begin with the following instrumentation:
For the Windows operating system begin the Windows PerfMon tool and enable the following instrumentation:
In the specific use case of Oracle Database Access to the ZFS Storage Appliance or ZFS Backup Appliance the Automatic Workload Repository (AWR) report should be enabled and used during any benchmarking or performance testing effort. AWR snapshots should be triggered at the beginning and end of the test workload and the AWR report should be generated from these snapshots. Review the ZFS Storage Appliance Analytics and and operating system accounting statistics in the context of the workload shown in the AWR load profile and tablespace I/O statistics and the storage related wait events. Pay specific attention to the following details:
The Oracle Database documenation specific to your release, available at docs.oracle.com, contains descriptions of all of the wait events shown in the AWR report. In the case of Oracle 11g, you can find the descriptions at this link: http://docs.oracle.com/cd/B28359_01/server.111/b28320/waitevents003.htm#BGGIBDJI. By comparing workload and response times reported by the test application, the database (if used), the operating system, and the storage system bottlnecks can be quickly and accurately identified. In practical systems, bottlenecks can be created by software bugs, misconfigured software, hardware faults, and misconfigured hardware. By carefully instrumenting the system and analyzing the results performance problems can be classified and solved. In general, system throughput may be limited by several factors
Hardware saturation is an important goal of any benchmarking activity. Network, CPU, and storage media all have well-defined limits as to how much throughput they can process, and hitting these limits usually indicates a well-tuned system. If the hardware is not saturated, then the problem limiting throughput may be related to software processing on the hardware. Troubleshooting this type of issue usually requires drawing a software block diagram showing the software components executed during the test, instrumenting those components, and identifying which software component is holding up the work, and finally reconfiguring or fixing that software component. This exercise is most easily undertaken by starting with the software/hardware block diagram in the system. Supply side bottlenecks show up when we double the workload at the application, throughput does not change, application response time increases, and storage response time stays the same. If there were no supply side bottlneck, doubling the number of I/O the application queues to the storage should double the time it takes for the storage to respond to the I/O. If the storage response time does not increase, then something between the application and the storage bottleneckd the I/O. This is a common problem when benchmarking systems that requires expert-level attention to the details of the system. Hardware faults, while rare, are a possibility in any real system. Always confirm there are no errors in the operating system log, the network driver status, and there are no faults reported in the storage system's status pages. Software bugs that are not present in lightly loaded conditions can show up under extereme load conditions. This includes race conditions that cause I/O to hang and processing problems that cause I/O to appear to stall. In cases where there are no obvious hardware faults, saturation, or bottlenecks pay careful attention to application code. In the context of the ZFSSA, if you hit a case where doubling the application workload does not cause a doubling of the front-end protocol response time and the operating system reports the same response time as the ZFSSA then you may have an application-level software bug. Attachments This solution has no attachment |
||||||||||||||
|