Since version 4.0 was released, we have focused on performance testing using standard topologies. For more information related to the system requirements, see System Requirements in the Installing, Upgrading and Migrating section.
The information in this document is distributed AS IS. The use of this information or the implementation of any of these techniques is a customer responsibility and depends on the customer’s ability to evaluate and integrate them into the customer’s operational environment. While each item may have been reviewed by IBM for accuracy in a specific situation, there is no guarantee that the same or similar results will be obtained elsewhere. Customers attempting to adapt these techniques to their own environments do so at their own risk. Any pointers in this publication to external Web sites are provided for convenience only and do not in any manner serve as an endorsement of these Web sites. Any performance data contained in this document was determined in a controlled environment, and therefore, the results that may be obtained in other operating environments may vary significantly. Users of this document should verify the applicable data for their specific environment.
Performance is based on measurements and projections using standard IBM benchmarks in a controlled environment. The actual throughput or performance that any user will experience will vary depending upon many factors, including considerations such as the amount of multi-programming in the user’s job stream, the I/O configuration, the storage configuration, and the workload processed. Therefore, no assurance can be given that an individual user will achieve results similar to those stated here.
This testing was done as a way to compare and characterize the differences in performance between different versions of the product. The results shown here should thus be looked at as a comparison of the contrasting performance between different versions, and not as an absolute benchmark of performance.
We use predominantly automated tooling such as Rational Performance Tester (RPT) to simulate a workload normally generated by client software such as the Eclipse client or web browsers. All response times listed are those measured by our automated tooling and not a client.
The diagram below describes at a very high level which aspects of the entire end-to-end experience (human end-user to server and back again) that our performance tests simulate. The tests described in this article simulate a segment of the end-to-end transaction as indicated in the middle of the diagram. Performance tests are server-side and capture response times for this segment of the transaction.
Application server and database - Linux for System z (single-tier configuration) | |
---|---|
Disk | DS8000 with PAV enabled |
Operating System | SUSE Linux Enterprise Server 10 (S390x) |
Application server | WebSphere Application Server Network Deployment v8.0.0.3 |
Database | DB2 9.7.0.5 |
Application server and database - Linux for System z (single-tier configuration) | |
---|---|
CPU | 1700 mips or higher |
Memory | 6 GB or higher |
Application server - Linux for System z (dual-tier configuration) | |
---|---|
Disk | DS8000 with PAV enabled |
Operating System | SUSE Linux Enterprise Server 10 (S390x) |
Application server | WebSphere Application Server Network Deployment v8.0.0.3 |
Database - IBM System z10 (dual-tier configuration) | |
---|---|
Disk | DS8000 with PAV enabled |
Operating System | z/OS 1.12 |
Database | DB2 on z/OS v10.1 |
Application server - Linux for System z (dual-tier configuration) | |
---|---|
CPU | 1700 mips or higher |
Memory | 6 GB or higher |
Database - IBM System z10 (dual-tier configuration) | |
---|---|
CPU | 200 mips or higher |
Memory | 6 GB or higher |
Application server and database - Linux for System z (single-tier configuration) | |
---|---|
Disk | DS8000 with PAV enabled |
Operating System | SUSE Linux Enterprise Server 10 (S390x) |
Application server | WebSphere Application Server Network Deployment v8.0.0.3 |
Database | DB2 9.7.0.5 |
Application server and database - Linux for System z (single-tier configuration) | |
---|---|
CPU | 2700 mips or higher |
Memory | 10 GB or higher |
Application server - Linux for System z (dual-tier configuration) | |
---|---|
Disk | DS8000 with PAV enabled |
Operating System | SUSE Linux Enterprise Server 10 (S390x) |
Application server | WebSphere Application Server Network Deployment v8.0.0.3 |
Database - IBM System z10 (dual-tier configuration) | |
---|---|
Disk | DS8000 with PAV enabled |
Operating System | z/OS 1.12 |
Database | DB2 on z/OS v10.1 |
Application server - Linux for System z (dual-tier configuration) | |
---|---|
CPU | 2700 mips or higher |
Memory | 10 GB or higher |
Database - IBM System z10 (dual-tier configuration) | |
---|---|
CPU | 300 mips or higher |
Memory | 6 GB or higher |
Rational Build Agent on IBM System z/OS | |
---|---|
Disk | DS8000 with PAV enabled |
Operating System | z/OS 1.12 |
Database | DB2 on z/OS v10.1 |
Rational Build Agent on IBM System z/OS | |
---|---|
CPU | 100 mips or higher |
Memory | 2 GB or higher |
IBM WebSphere Application Server Network Deployment | 8.0.0.3 | JVM settings:
* Small business scale: -Xmx4g -Xms4g -Xmn512m -Xgcpolicy:gencon -Xcompressedrefs -Xgc:preferredHeapBase=0x100000000 * Large business scale: -Xmx6g -Xms6g -Xmn512m -Xgcpolicy:gencon -Xcompressedrefs -Xgc:preferredHeapBase=0x100000000 Thread pools:
OS Configuration: Increase system wide resources for the app server process owner:
|
DB2 | DB2 on z/OS v10.1 | |
License server | Same as CLM version | Hosted locally by JTS server |
RPT workbench | 8.2.1.5 | Defaults |
RPT agents | 8.2.1.5 | Defaults |
Network | Shared subnet within test lab |
LuYuliang is has been a Software Engineer at IBM for over 3 years. He is part of the Rational Performance Engineering group.
Warning: Can't find topic Deployment.PerformanceDatasheetReaderComments
Status icon key: