Thursday Apr 19, 2012

Sun ZFS Storage 7420 Appliance Delivers 2-Node World Record SPECsfs2008 NFS Benchmark

Oracle's Sun ZFS Storage 7420 appliance delivered world record two-node performance on the SPECsfs2008 NFS benchmark, beating results published on NetApp's dual-controller and 4-node high-end FAS6240 storage systems.

  • The Sun ZFS Storage 7420 appliance delivered a world record two-node result of 267,928 SPECsfs2008_nfs.v3 Ops/sec with an Overall Response Time (ORT) of 1.31 msec on the SPECsfs2008 NFS benchmark.

  • The Sun ZFS Storage 7420 appliance delivered 1.4x higher throughput than the dual-controller NetApp FAS6240 and 2.6x higher throughput than the dual-controller NetApp FAS3270 on the SPECsfs2008_nfs.v3 benchmark at less than half the list price of either result.

  • The Sun ZFS Storage 7420 appliance required 10 percent less rack space than the dual-controller NetApp FAS6240.

  • The Sun ZFS Storage 7420 appliance had 3 percent higher throughput than the 4-node NetApp FAS6240 on the SPECsfs2008_nfs.v3 benchmark.

  • The Sun ZFS Storage 7420 appliance required 25 percent less rack space than the 4-node NetApp FAS6240.

  • The Sun ZFS Storage 7420 appliance has 14 percent better Overall Response Time than the 4-node NetApp FAS6240 on the SPECsfs2008_nfs.v3 benchmark.

Performance Landscape

SPECsfs2008_nfs.v3 Performance Chart (in decreasing SPECsfs2008_nfs.v3 Ops/sec order)

Sponsor System Throughput
(Ops/sec)
Overall Response
Time (msec)
Nodes Memory (GB)
Including Flash
Disks Rack Units –
Controllers
+Disks
Oracle 7420 267,928 1.31 2 6,728 280 54
NetApp FAS6240 260,388 1.53 4 2,256 288 72
NetApp FAS6240 190,675 1.17 2 1,128 288 60
EMC VG8 135,521 1.92 280 312
Oracle 7320 134,140 1.51 2 4,968 136 26
EMC NS-G8 110,621 2.32 264 100
NetApp FAS3270 101,183 1.66 2 40 360 66

Throughput SPECsfs2008_nfs.v3 Ops/sec — the Performance Metric
Overall Response Time — the corresponding Response Time Metric
Nodes — Nodes and Controllers are being used interchangeably

Complete SPECsfs2008 benchmark results may be found at http://www.spec.org/sfs2008/results/sfs2008.html.

Configuration Summary

Storage Configuration:

Sun ZFS Storage 7420 appliance in clustered configuration
2 x Sun ZFS Storage 7420 controllers, each with
4 x 2.4 GHz Intel Xeon E7-4870 processors
1 TB memory
4 x 512 GB SSD flash-enabled read-cache
2 x 10GbE NICs
12 x Sun Disk shelves
10 x shelves with 24 x 300 GB 15K RPM SAS-2 drives
2 x shelves with 20 x 300 GB 15K RPM SAS-2 drives and 4 x 73 GB SAS-2 flash-enabled write-cache

Server Configuration:

4 x Sun Fire X4270 M2 servers, each with
2 x 3.3 GHz Intel Xeon E5680 processors
144 GB memory
1 x 10 GbE NIC
Oracle Solaris 10 9/10

Switches:

1 x 24-port 10Gb Ethernet Switch

Benchmark Description

SPECsfs2008 is the latest version of the Standard Performance Evaluation Corporation (SPEC) benchmark suite measuring file server throughput and response time, providing a standardized method for comparing performance across different vendor platforms. SPECsfs2008 results summarize the server's capabilities with respect to the number of operations that can be handled per second, as well as the overall latency of the operations. The suite is a follow-on to the SFS97_R1 benchmark, adding a CIFS workload, an updated NFSv3 workload, support for additional client platforms, and a new test harness and reporting/submission framework.

See Also

Disclosure Statement

SPEC and SPECsfs are registered trademarks of Standard Performance Evaluation Corporation (SPEC). Results as of April 18, 2012, for more information see www.spec.org. Sun ZFS Storage 7420 Appliance 267,928 SPECsfs2008_nfs.v3 Ops/sec, 1.31 msec ORT, NetApp Data ONTAP 8.1 Cluster-Mode (4-node FAS6240) 260,388 SPECsfs2008_nfs.v3 Ops/Sec, 1.53 msec ORT, NetApp FAS6240 190,675 SPECsfs2008_nfs.v3 Ops/Sec, 1.17 msec ORT. NetApp FAS3270 101,183 SPECsfs2008_nfs.v3 Ops/Sec, 1.66 msec ORT.

Nodes refer to the item in the SPECsfs2008 disclosed Configuration Bill of Materials that have the Processing Elements that perform the NFS Processing Function. These are the first item listed in each of disclosed Configuration Bill of Materials except for EMC where it is both the first and third items listed, and HP, where it is the second item listed as Blade Servers. The number of nodes is from the QTY disclosed in the Configuration Bill of Materials as described above. Configuration Bill of Materials list price for Oracle result of US$ 423,644. Configuration Bill of Materials list price for NetApp FAS3270 result of US$ 1,215,290. Configuration Bill of Materials list price for NetApp FAS6240 result of US$ 1,028,118. Oracle pricing from https://shop.oracle.com/pls/ostore/f?p=dstore:home:0, traverse to "Storage and Tape" and then to "NAS Storage". NetApp's pricing from http://www.netapp.com/us/media/na-list-usd-netapp-custom-state-new-discounts.html.

Tuesday Apr 10, 2012

World Record Oracle E-Business Suite 12.1.3 Standard Extra-Large Payroll (Batch) Benchmark on Sun Server X3-2L

Oracle's Sun Server X3-2L (formerly Sun Fire X4270 M3) server set a world record running the Oracle E-Business Suite 12.1.3 Standard Extra-Large Payroll (Batch) benchmark.

  • This is the first published result using Oracle E-Business 12.1.3.

  • The Sun Server X3-2L result ran the Extra-Large Payroll workload in 19 minutes.

Performance Landscape

This is the first published result for the Payroll Extra-Large model using Oracle E-Business 12.1.3 benchmark.

Batch Workload: Payroll Extra-Large Model
System Employees/Hr Elapsed Time
Sun Server X3-2L 789,515 19 minutes

Configuration Summary

Hardware Configuration:

Sun Server X3-2L
2 x Intel Xeon E5-2690, 2.9 GHz
128 GB memory
8 x 100 GB SSD for data
1 x 300 GB SSD for log

Software Configuration:

Oracle Linux 5.7
Oracle E-Business Suite R12 (12.1.3)
Oracle Database 11g (11.2.0.3)

Benchmark Description

The Oracle E-Business Suite Standard R12 Benchmark combines online transaction execution by simulated users with concurrent batch processing to model a typical scenario for a global enterprise. This benchmark ran one Batch component, Payroll, in the Extra-Large size. The goal of the benchmark proposal is to execute and achieve best batch-payroll performance using X-Large configuragion.

Results can be published in four sizes and use one or more online/batch modules

  • X-large: Maximum online users running all business flows between 10,000 to 20,000; 750,000 order to cash lines per hour and 250,000 payroll checks per hour.
    • Order to Cash Online -- 2400 users
      • The percentage across the 5 transactions in Order Management module is:
        • Insert Manual Invoice -- 16.66%
        • Insert Order -- 32.33%
        • Order Pick Release -- 16.66%
        • Ship Confirm -- 16.66%
        • Order Summary Report -- 16.66%
    • HR Self-Service -- 4000 users
    • Customer Support Flow -- 8000 users
    • Procure to Pay -- 2000 users
  • Large: 10,000 online users; 100,000 order to cash lines per hour and 100,000 payroll checks per hour.
  • Medium: up to 3000 online users; 50,000 order to cash lines per hour and 10,000 payroll checks per hour.
  • Small: up to 1000 online users; 10,000 order to cash lines per hour and 5,000 payroll checks per hour.

See Also

Disclosure Statement

Oracle E-Business X-Large Batch-Payroll benchmark, Sun Server X3-2L, 2.90 GHz, 2 chips, 16 cores, 32 threads, 128 GB memory, elapsed time 19.0 minutes, 789,515 Employees/HR, Oracle Linux 5.7, Oracle E-Business Suite 12.1.3, Oracle Database 11g Release 2, Results as of 7/10/2012.

SPEC CPU2006 Results on Oracle's Sun x86 Servers

Oracle's new Sun x86 servers delivered world records on the benchmarks SPECfp2006 and SPECint_rate2006 for two processor servers. This was accomplished with Oracle Solaris 11 and Oracle Solaris Studio 12.3 software.

  • The Sun Fire X4170 M3 (now known as Sun Server X3-2) server achieved a world record result in for SPECfp2006 benchmark with a score of 96.8.

  • The Sun Blade X6270 M3 server module (now known as Sun Blade X3-2B) produced best integer throughput performance for all 2-socket servers with a SPECint_rate2006 score of 705.

  • The Sun x86 servers with Intel Xeon E5-2690 2.9 GHz processors produced a cross-generational performance improvement up to 1.8x over the previous generation, Sun x86 M2 servers.

Performance Landscape

Complete benchmark results are at the SPEC website, SPEC CPU2006 Results. The tables below provide the new Oracle results, as well as, select results from other vendors.

SPECint2006
System Processor c/c/t * Peak Base O/S Compiler
Fujitsu PRIMERGY BX924 S3 Intel E5-2690, 2.9 GHz 2/16/16 60.8 56.0 RHEL 6.2 Intel 12.1.2.273
Sun Fire X4170 M3 Intel E5-2690, 2.9 GHz 2/16/32 58.5 54.3 Oracle Linux 6.1 Intel 12.1.0.225
Sun Fire X4270 M2 Intel X5690, 3.47 GHz 2/12/12 46.2 43.9 Oracle Linux 5.5 Intel 12.0.1.116

SPECfp2006
System Processor c/c/t * Peak Base O/S Compiler
Sun Fire X4170 M3 Intel E5-2690, 2.9 GHz 2/16/32 96.8 86.4 Oracle Solaris 11 Studio 12.3
Sun Blade X6270 M3 Intel E5-2690, 2.9 GHz 2/16/32 96.0 85.2 Oracle Solaris 11 Studio 12.3
Sun Fire X4270 M3 Intel E5-2690, 2.9 GHz 2/16/32 95.9 85.1 Oracle Solaris 11 Studio 12.3
Fujitsu CELSIUS R920 Intel E5-2687, 2.9 GHz 2/16/16 93.8 87.6 RHEL 6.1 Intel 12.1.2.273
Sun Fire X4270 M2 Intel X5690, 3.47 GHz 2/12/24 64.2 59.2 Oracle Solaris 10 Studio 12.2

Only 2-chip server systems listed below, excludes workstations.

SPECint_rate2006
System Processor Base
Copies
c/c/t * Peak Base O/S Compiler
Sun Blade X6270 M3 Intel E5-2690, 2.9 GHz 32 2/16/32 705 632 Oracle Solaris 11 Studio 12.3
Sun Fire X4270 M3 Intel E5-2690, 2.9 GHz 32 2/16/32 705 630 Oracle Solaris 11 Studio 12.3
Sun Fire X4170 M3 Intel E5-2690, 2.9 GHz 32 2/16/32 702 628 Oracle Solaris 11 Studio 12.3
Cisco UCS C220 M3 Intel E5-2690, 2.9 GHz 32 2/16/32 697 671 RHEL 6.2 Intel 12.1.0.225
Sun Blade X6270 M2 Intel X5690, 3.47 GHz 24 2/12/24 410 386 Oracle Linux 5.5 Intel 12.0.1.116

SPECfp_rate2006
System Processor Base
Copies
c/c/t * Peak Base O/S Compiler
Cisco UCS C240 M3 Intel E5-2690, 2.9 GHz 32 2/16/32 510 496 RHEL 6.2 Intel 12.1.2.273
Sun Fire X4270 M3 Intel E5-2690, 2.9 GHz 64 2/16/32 497 461 Oracle Solaris 11 Studio 12.3
Sun Blade X6270 M3 Intel E5-2690, 2.9 GHz 32 2/16/32 497 460 Oracle Solaris 11 Studio 12.3
Sun Fire X4170 M3 Intel E5-2690, 2.9 GHz 64 2/16/32 495 464 Oracle Solaris 11 Studio 12.3
Sun Fire X4270 M2 Intel E5690, 3.47 GHz 24 2/12/24 273 265 Oracle Linux 5.5 Intel 12.0.1.116

* c/c/t — chips / cores / threads enabled

Configuration Summary and Results

Hardware Configuration:

Sun Fire X4170 M3 server
2 x 2.90 GHz Intel Xeon E5-2690 processors
128 GB memory (16 x 8 GB 2Rx4 PC3-12800R-11, ECC)

Sun Fire X4270 M3 server
2 x 2.90 GHz Intel Xeon E5-2690 processors
128 GB memory (16 x 8 GB 2Rx4 PC3-12800R-11, ECC)

Sun Blade X6270 M3 server module
2 x 2.90 GHz Intel Xeon E5-2690 processors
128 GB memory (16 x 8 GB 2Rx4 PC3-12800R-11, ECC)

Software Configuration:

Oracle Solaris 11 11/11 (SRU2)
Oracle Solaris Studio 12.3 (patch update 1 nightly build 120313) Oracle Linux Server Release 6.1
Intel C++ Studio XE 12.1.0.225
SPEC CPU2006 V1.2

Benchmark Description

SPEC CPU2006 is SPEC's most popular benchmark. It measures:

  • Speed — single copy performance of chip, memory, compiler
  • Rate — multiple copy (throughput)

The benchmark is also divided into integer intensive applications and floating point intensive applications:

  • integer: 12 benchmarks derived from real applications such as perl, gcc, XML processing, and pathfinding
  • floating point: 17 benchmarks derived from real applications, including chemistry, physics, genetics, and weather.

It is also divided depending upon the amount of optimization allowed:

  • base: optimization is consistent per compiled language, all benchmarks must be compiled with the same flags per language.
  • peak: specific compiler optimization is allowed per application.

The overall metrics for the benchmark which are commonly used are:

  • SPECint_rate2006, SPECint_rate_base2006: integer, rate
  • SPECfp_rate2006, SPECfp_rate_base2006: floating point, rate
  • SPECint2006, SPECint_base2006: integer, speed
  • SPECfp2006, SPECfp_base2006: floating point, speed

See here for additional information.

See Also

Disclosure Statement

SPEC and the benchmark names SPECfp and SPECint are registered trademarks of the Standard Performance Evaluation Corporation. Results as of 10 April 2012 from www.spec.org and this report.

Thursday Mar 29, 2012

Sun Server X2-8 (formerly Sun Fire X4800 M2) Delivers World Record TPC-C for x86 Systems

Oracle's Sun Server X2-8 (formerly Sun Fire X4800 M2 server) equipped with eight 2.4 GHz Intel Xeon Processor E7-8870 chips obtained a result of 5,055,888 tpmC on the TPC-C benchmark. This result is a world record for x86 servers. Oracle demonstrated this world record database performance running Oracle Database 11g Release 2 Enterprise Edition with Partitioning.

  • The Sun Server X2-8 delivered a new x86 TPC-C world record of 5,055,888 tpmC with a price performance of $0.89/tpmC using Oracle Database 11g Release 2. This configuration is available 7/10/12.

  • The Sun Server X2-8 delivers 3.0x times better performance than the next 8-processor result, an IBM System p 570 equipped with POWER6 processors.

  • The Sun Server X2-8 has 3.1x times better price/performance than the 8-processor 4.7GHz POWER6 IBM System p 570.

  • The Sun Server X2-8 has 1.6x times better performance than the 4-processor IBM x3850 X5 system equipped with Intel Xeon processors.

  • This is the first TPC-C result on any system using eight Intel Xeon Processor E7-8800 Series chips.

  • The Sun Server X2-8 is the first x86 system to get over 5 million tpmC.

  • The Oracle solution utilized Oracle Linux operating system and Oracle Database 11g Enterprise Edition Release 2 with Partitioning to produce the x86 world record TPC-C benchmark performance.

Performance Landscape

Select TPC-C results (sorted by tpmC, bigger is better)

System p/c/t tpmC Price
/tpmC
Avail Database Memory
Size
Sun Server X2-8 8/80/160 5,055,888 0.89 USD 7/10/2012 Oracle 11g R2 4 TB
IBM x3850 X5 4/40/80 3,014,684 0.59 USD 7/11/2011 DB2 ESE 9.7 3 TB
IBM x3850 X5 4/32/64 2,308,099 0.60 USD 5/20/2011 DB2 ESE 9.7 1.5 TB
IBM System p 570 8/16/32 1,616,162 3.54 USD 11/21/2007 DB2 9.0 2 TB

p/c/t - processors, cores, threads
Avail - availability date

Oracle and IBM TPC-C Response times

System tpmC Response Time (sec)
New Order 90th%
Response Time (sec)
New Order Average

Sun Server X2-8 5,055,888 0.210 0.166
IBM x3850 X5 3,014,684 0.500 0.272
Ratios - Oracle Better 1.6x 1.4x 1.3x

Oracle uses average new order response time for comparison between Oracle and IBM.

Graphs of Oracle's and IBM's response times for New-Order can be found in the full disclosure reports on TPC's website TPC-C Official Result Page.

Configuration Summary and Results

Hardware Configuration:

Server
Sun Server X2-8
8 x 2.4 GHz Intel Xeon Processor E7-8870
4 TB memory
8 x 300 GB 10K RPM SAS internal disks
8 x Dual port 8 Gbs FC HBA

Data Storage
10 x Sun Fire X4270 M2 servers configured as COMSTAR heads, each with
1 x 3.06 GHz Intel Xeon X5675 processor
8 GB memory
10 x 2 TB 7.2K RPM 3.5" SAS disks
2 x Sun Storage F5100 Flash Array storage (1.92 TB each)
1 x Brocade 5300 switches

Redo Storage
2 x Sun Fire X4270 M2 servers configured as COMSTAR heads, each with
1 x 3.06 GHz Intel Xeon X5675 processor
8 GB memory
11 x 2 TB 7.2K RPM 3.5" SAS disks

Clients
8 x Sun Fire X4170 M2 servers, each with
2 x 3.06 GHz Intel Xeon X5675 processors
48 GB memory
2 x 300 GB 10K RPM SAS disks

Software Configuration:

Oracle Linux (Sun Fire 4800 M2)
Oracle Solaris 11 Express (COMSTAR for Sun Fire X4270 M2)
Oracle Solaris 10 9/10 (Sun Fire X4170 M2)
Oracle Database 11g Release 2 Enterprise Edition with Partitioning
Oracle iPlanet Web Server 7.0 U5
Tuxedo CFS-R Tier 1

Results:

System: Sun Server X2-8
tpmC: 5,055,888
Price/tpmC: 0.89 USD
Available: 7/10/2012
Database: Oracle Database 11g
Cluster: no
New Order Average Response: 0.166 seconds

Benchmark Description

TPC-C is an OLTP system benchmark. It simulates a complete environment where a population of terminal operators executes transactions against a database. The benchmark is centered around the principal activities (transactions) of an order-entry environment. These transactions include entering and delivering orders, recording payments, checking the status of orders, and monitoring the level of stock at the warehouses.

Key Points and Best Practices

  • Oracle Database 11g Release 2 Enterprise Edition with Partitioning scales easily to this high level of performance.

  • COMSTAR (Common Multiprotocol SCSI Target) is the software framework that enables an Oracle Solaris host to serve as a SCSI Target platform. COMSTAR uses a modular approach to break the huge task of handling all the different pieces in a SCSI target subsystem into independent functional modules which are glued together by the SCSI Target Mode Framework (STMF). The modules implementing functionality at SCSI level (disk, tape, medium changer etc.) are not required to know about the underlying transport. And the modules implementing the transport protocol (FC, iSCSI, etc.) are not aware of the SCSI-level functionality of the packets they are transporting. The framework hides the details of allocation providing execution context and cleanup of SCSI commands and associated resources and simplifies the task of writing the SCSI or transport modules.

  • Oracle iPlanet Web Server middleware is used for the client tier of the benchmark. Each web server instance supports more than a quarter-million users while satisfying the response time requirement from the TPC-C benchmark.

See Also

Disclosure Statement

TPC Benchmark C, tpmC, and TPC-C are trademarks of the Transaction Processing Performance Council (TPC). Sun Server X2-8 (8/80/160) with Oracle Database 11g Release 2 Enterprise Edition with Partitioning, 5,055,888 tpmC, $0.89 USD/tpmC, available 7/10/2012. IBM x3850 X5 (4/40/80) with DB2 ESE 9.7, 3,014,684 tpmC, $0.59 USD/tpmC, available 7/11/2011. IBM x3850 X5 (4/32/64) with DB2 ESE 9.7, 2,308,099 tpmC, $0.60 USD/tpmC, available 5/20/2011. IBM System p 570 (8/16/32) with DB2 9.0, 1,616,162 tpmC, $3.54 USD/tpmC, available 11/21/2007. Source: http://www.tpc.org/tpcc, results as of 7/15/2011.

Sun Server X2-8 (formerly Sun Fire X4800 M2) Posts World Record x86 SPECjEnterprise2010 Result

Oracle's Sun Server X2-8 (formerly Sun Fire X4800 M2) using the Intel Xeon E7-8870 processor and Sun Server X2-4 using the Intel Xeon E7-4870 processor, produced a world record single application server SPECjEnterprise2010 benchmark result of 27,150.05 SPECjEnterprise2010 EjOPS. The Sun Server X2-8 ran the application tier and the Sun Server X2-4 was used for the database tier.

  • The Sun Server X2-8 demonstrated 63% better performance compared to IBM P780 server result of 16,646.34 SPECjEnterprise2010 EjOPS.

  • The Sun Server X2-8 demonstrated 4% better performance than the Cisco UCS B440 M2 result, both results used the same number of processors.

  • This result used Oracle WebLogic Server 12c, Java HotSpot(TM) 64-Bit Server 1.7.0_02, and Oracle Database 11g.

  • This result was produced using Oracle Linux.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results. The table below compares against the best results from IBM and Cisco.

SPECjEnterprise2010 Performance Chart
as of 7/11/2012
Submitter EjOPS* Application Server Database Server
Oracle 27,150.05 1x Sun Server X2-8
8x 2.4 GHz Intel Xeon E7-8870
Oracle WebLogic 12c
1x Sun Server X2-4
4x 2.4 GHz Intel Xeon E7-4870
Oracle Database 11g (11.2.0.2)
Cisco 26,118.67 2x UCS B440 M2 Blade Server
4x 2.4 GHz Intel Xeon E7-4870
Oracle WebLogic 11g (10.3.5)
1x UCS C460 M2 Blade Server
4x 2.4 GHz Intel Xeon E7-4870
Oracle Database 11g (11.2.0.2)
IBM 16,646.34 1x IBM Power 780
8x 3.86 GHz POWER 7
WebSphere Application Server V7
1x IBM Power 750 Express
4x 3.55 GHz POWER 7
IBM DB2 9.7 Workgroup Server Edition FP3a

* SPECjEnterprise2010 EjOPS, bigger is better.

Configuration Summary

Application Server:

1 x Sun Server X2-8

8 x 2.4 GHz Intel Xeon processor E7-8870
256 GB memory
4 x 10 GbE NIC
2 x FC HBA
Oracle Linux 5 Update 6
Oracle WebLogic Server Standard Edition Release 12.1.1
Java HotSpot(TM) 64-Bit Server VM on Linux, version 1.7.0_02 (Java SE 7 Update 2)

Database Server:

1 x Sun Server X2-4
4 x 2.4 GHz Intel Xeon E7-4870
512 GB memory
4 x 10 GbE NIC
2 x FC HBA
2 x Sun StorageTek 2540 M2
4 x Sun Fire X4270 M2
4 x Sun Storage F5100 Flash Array
Oracle Linux 5 Update 6
Oracle Database 11g Enterprise Edition Release 11.2.0.2

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The SPECjEnterprise2010 benchmark has been designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems.

The workload consists of an end to end web based order processing domain, an RMI and Web Services driven manufacturing domain and a supply chain model utilizing document based Web Services. The application is a collection of Java classes, Java Servlets, Java Server Pages, Enterprise Java Beans, Java Persistence Entities (pojo's) and Message Driven Beans.

The SPECjEnterprise2010 benchmark heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second ("SPECjEnterprise2010 EjOPS"). This metric is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is no price/performance metric in this benchmark.

Key Points and Best Practices

  • Sixteen Oracle WebLogic server instances were started using numactl, binding 2 instances per chip.
  • Eight Oracle database listener processes were started, binding 2 instances per chip using taskset.
  • Additional tuning information is in the report at http://spec.org.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Sun Server X2-8, 27,150.05 SPECjEnterprise2010 EjOPS; IBM Power 780, 16,646.34 SPECjEnterprise2010 EjOPS; Cisco UCS B440 M2, 26,118.67 SPECjEnterprise2010 EjOPS. Results from www.spec.org as of 7/11/2012.

Thursday Jan 12, 2012

Netra SPARC T4-2 SPECjvm2008 World Record Performance

Oracle's Netra SPARC T4-2 server equipped with two SPARC T4 processors running at 2.85 GHz delivered a World Record result of 454.52 SPECjvm2008 Peak ops/m on the SPECjvm2008 benchmark. This result just eclipsed the previous record which was run on a similar product, Oracle's SPARC T4-2 server, which is also a two SPARC T4 processor based system.

  • The Netra SPARC T4-2 server demonstrates 41% better performance than the SPARC T3-2 server and similar performance to Oracle's SPARC T4-2 server.

  • The Netra SPARC T4-2 server running the SPECjvm2008 benchmark achieved a score of 454.52 SPECjvm2008 Peak ops/m while the Sun Blade X6270 server module achieved 317.13 SPECjvm2008 Base ops/m.

  • The Netra SPARC T4-2 server with hardware cryptography acceleration greatly increases performance with subtests using AES and RSA encryption ciphers.

  • This result was produced using Oracle Solaris 11 and Oracle JDK 7 Update 2.

  • There are no SPECjvm2008 results published by IBM on POWER7 based systems.

  • The Netra SPARC T4-2 server demonstrates Oracle's position of leadership in Java-based computing by publishing world record results for the SPECjvm2008 benchmark.

Performance Landscape

Complete benchmark results are at the SPECjvm2008 website.

SPECjvm2008 Performance Chart
(ordered by performance)
System Processors Performance
base peak
Netra SPARC T4-2 2 x 2.85 GHz SPARC T4 - 454.52
SPARC T4-2 2 x 2.85 GHz SPARC T4 - 454.25
SPARC T3-2 2 x 1.65 GHz SPARC T3 - 320.52
Sun Blade X6270 2 x 2.93 GHz Intel X5570 317.13 -

base: SPECjvm2008 Base ops/m (bigger is better)
peak: SPECjvm2008 Peak ops/m (bigger is better)

SPEC allows base and peak to be submitted separately. The base metric does not allow any optimization of the JVM, the peak metric allows optimization.

Configuration Summary

Hardware Configuration:

Netra SPARC T4-2 server
2 x 2.85 GHz SPARC T4 processors
256 GB memory

Software Configuration:

Oracle Solaris 11 11/11
Java Platform, Standard Edition, JDK 7 Update 2

Benchmark Description

SPECjvm2008 (Java Virtual Machine Benchmark) is a benchmark suite for measuring the performance of a Java Runtime Environment (JRE), containing several real life applications and benchmarks focusing on core Java functionality. The suite focuses on the performance of the JRE executing a single application; it reflects the performance of the hardware processor and memory subsystem, but has low dependence on file I/O and includes no network I/O across machines.

The SPECjvm2008 workload mimics a variety of common general purpose application computations. These characteristics reflect the intent that this benchmark will be applicable to measuring basic Java performance on a wide variety of both client and server systems.

SPECjvm2008 benchmark highlights:

  • Leverages real life applications (like derby, sunflow, and javac) and area-focused benchmarks (like xml, serialization, crypto, and scimark).
  • Also measures the performance of the operating system and hardware in the context of executing the JRE.

The current rules for the benchmark allow either base or peak to be run. The base run is done without any tuning of the JVM to improve the out of the box performance. The peak run allows tuning of the JVM.

Key Points and Best Practices

  • Enhancements to the JVM had a major impact on performance, especially for the security tests.

See Also

Disclosure Statement

SPEC and SPECjvm are registered trademarks of Standard Performance Evaluation Corporation. Results from www.spec.org and this report as of 1/9/2012. Netra SPARC T4-2 454.52 SPECjvm2008 Peak ops/m submitted for review, SPARC T4-2 454.25 SPECjvm2008 Peak ops/m, SPARC T3-2 320.52 SPECjvm2008 Peak ops/m, Sun Blade X6270 317.13 SPECjvm2008 Base ops/m.

Wednesday Nov 30, 2011

SPARC T4-4 Beats 8-CPU IBM POWER7 on TPC-H @3000GB Benchmark

Oracle's SPARC T4-4 server delivered a world record TPC-H @3000GB benchmark result for systems with four processors. This result beats eight processor results from IBM (POWER7) and HP (x86). The SPARC T4-4 server also delivered better performance per core than these eight processor systems from IBM and HP. Comparisons below are based upon system to system comparisons, highlighting Oracle's complete software and hardware solution.

This database world record result used Oracle's Sun Storage 2540-M2 arrays (rotating disk) connected to a SPARC T4-4 server running Oracle Solaris 11 and Oracle Database 11g Release 2 demonstrating the power of Oracle's integrated hardware and software solution.

  • The SPARC T4-4 server based configuration achieved a TPC-H scale factor 3000 world record for four processor systems of 205,792 QphH@3000GB with price/performance of $4.10/QphH@3000GB.

  • The SPARC T4-4 server with four SPARC T4 processors (total of 32 cores) is 7% faster than the IBM Power 780 server with eight POWER7 processors (total of 32 cores) on the TPC-H @3000GB benchmark.

  • The SPARC T4-4 server is 36% better in price performance compared to the IBM Power 780 server on the TPC-H @3000GB Benchmark.

  • The SPARC T4-4 server is 29% faster than the IBM Power 780 for data loading.

  • The SPARC T4-4 server is up to 3.4 times faster than the IBM Power 780 server for the Refresh Function.

  • The SPARC T4-4 server with four SPARC T4 processors is 27% faster than the HP ProLiant DL980 G7 server with eight x86 processors on the TPC-H @3000GB benchmark.

  • The SPARC T4-4 server is 52% faster than the HP ProLiant DL980 G7 server for data loading.

  • The SPARC T4-4 server is up to 3.2 times faster than the HP ProLiant DL980 G7 for the Refresh Function.

  • The SPARC T4-4 server achieved a peak IO rate from the Oracle database of 17 GB/sec. This rate was independent of the storage used, as demonstrated by the TPC-H @3000TB benchmark which used twelve Sun Storage 2540-M2 arrays (rotating disk) and the TPC-H @1000TB benchmark which used four Sun Storage F5100 Flash Array devices (flash storage). [*]

  • The SPARC T4-4 server showed linear scaling from TPC-H @1000GB to TPC-H @3000GB. This demonstrates that the SPARC T4-4 server can handle the increasingly larger databases required of DSS systems. [*]

  • The SPARC T4-4 server benchmark results demonstrate a complete solution of building Decision Support Systems including data loading, business questions and refreshing data. Each phase usually has a time constraint and the SPARC T4-4 server shows superior performance during each phase.

[*] The TPC believes that comparisons of results published with different scale factors are misleading and discourages such comparisons.

Performance Landscape

The table lists the leading TPC-H @3000GB results for non-clustered systems.

TPC-H @3000GB, Non-Clustered Systems
System
Processor
P/C/T – Memory
Composite
(QphH)
$/perf
($/QphH)
Power
(QppH)
Throughput
(QthH)
Database Available
SPARC Enterprise M9000
3.0 GHz SPARC64 VII+
64/256/256 – 1024 GB
386,478.3 $18.19 316,835.8 471,428.6 Oracle 11g R2 09/22/11
SPARC T4-4
3.0 GHz SPARC T4
4/32/256 – 1024 GB
205,792.0 $4.10 190,325.1 222,515.9 Oracle 11g R2 05/31/12
SPARC Enterprise M9000
2.88 GHz SPARC64 VII
32/128/256 – 512 GB
198,907.5 $15.27 182,350.7 216,967.7 Oracle 11g R2 12/09/10
IBM Power 780
4.1 GHz POWER7
8/32/128 – 1024 GB
192,001.1 $6.37 210,368.4 175,237.4 Sybase 15.4 11/30/11
HP ProLiant DL980 G7
2.27 GHz Intel Xeon X7560
8/64/128 – 512 GB
162,601.7 $2.68 185,297.7 142,685.6 SQL Server 2008 10/13/10

P/C/T = Processors, Cores, Threads
QphH = the Composite Metric (bigger is better)
$/QphH = the Price/Performance metric in USD (smaller is better)
QppH = the Power Numerical Quantity
QthH = the Throughput Numerical Quantity

The following table lists data load times and refresh function times during the power run.

TPC-H @3000GB, Non-Clustered Systems
Database Load & Database Refresh
System
Processor
Data Loading
(h:m:s)
T4
Advan
RF1
(sec)
T4
Advan
RF2
(sec)
T4
Advan
SPARC T4-4
3.0 GHz SPARC T4
04:08:29 1.0x 67.1 1.0x 39.5 1.0x
IBM Power 780
4.1 GHz POWER7
05:51:50 1.5x 147.3 2.2x 133.2 3.4x
HP ProLiant DL980 G7
2.27 GHz Intel Xeon X7560
08:35:17 2.1x 173.0 2.6x 126.3 3.2x

Data Loading = database load time
RF1 = power test first refresh transaction
RF2 = power test second refresh transaction
T4 Advan = the ratio of time to T4 time

Complete benchmark results found at the TPC benchmark website http://www.tpc.org.

Configuration Summary and Results

Hardware Configuration:

SPARC T4-4 server
4 x SPARC T4 3.0 GHz processors (total of 32 cores, 128 threads)
1024 GB memory
8 x internal SAS (8 x 300 GB) disk drives

External Storage:

12 x Sun Storage 2540-M2 array storage, each with
12 x 15K RPM 300 GB drives, 2 controllers, 2 GB cache

Software Configuration:

Oracle Solaris 11 11/11
Oracle Database 11g Release 2 Enterprise Edition

Audited Results:

Database Size: 3000 GB (Scale Factor 3000)
TPC-H Composite: 205,792.0 QphH@3000GB
Price/performance: $4.10/QphH@3000GB
Available: 05/31/2012
Total 3 year Cost: $843,656
TPC-H Power: 190,325.1
TPC-H Throughput: 222,515.9
Database Load Time: 4:08:29

Benchmark Description

The TPC-H benchmark is a performance benchmark established by the Transaction Processing Council (TPC) to demonstrate Data Warehousing/Decision Support Systems (DSS). TPC-H measurements are produced for customers to evaluate the performance of various DSS systems. These queries and updates are executed against a standard database under controlled conditions. Performance projections and comparisons between different TPC-H Database sizes (100GB, 300GB, 1000GB, 3000GB, 10000GB, 30000GB and 100000GB) are not allowed by the TPC.

TPC-H is a data warehousing-oriented, non-industry-specific benchmark that consists of a large number of complex queries typical of decision support applications. It also includes some insert and delete activity that is intended to simulate loading and purging data from a warehouse. TPC-H measures the combined performance of a particular database manager on a specific computer system.

The main performance metric reported by TPC-H is called the TPC-H Composite Query-per-Hour Performance Metric (QphH@SF, where SF is the number of GB of raw data, referred to as the scale factor). QphH@SF is intended to summarize the ability of the system to process queries in both single and multiple user modes. The benchmark requires reporting of price/performance, which is the ratio of the total HW/SW cost plus 3 years maintenance to the QphH. A secondary metric is the storage efficiency, which is the ratio of total configured disk space in GB to the scale factor.

Key Points and Best Practices

  • Twelve Sun Storage 2540-M2 arrays were used for the benchmark. Each Sun Storage 2540-M2 array contains 12 15K RPM drives and is connected to a single dual port 8Gb FC HBA using 2 ports. Each Sun Storage 2540-M2 array showed 1.5 GB/sec for sequential read operations and showed linear scaling, achieving 18 GB/sec with twelve Sun Storage 2540-M2 arrays. These were stand alone IO tests.

  • The peak IO rate measured from the Oracle database was 17 GB/sec.

  • Oracle Solaris 11 11/11 required very little system tuning.

  • Some vendors try to make the point that storage ratios are of customer concern. However, storage ratio size has more to do with disk layout and the increasing capacities of disks – so this is not an important metric in which to compare systems.

  • The SPARC T4-4 server and Oracle Solaris efficiently managed the system load of over one thousand Oracle Database parallel processes.

  • Six Sun Storage 2540-M2 arrays were mirrored to another six Sun Storage 2540-M2 arrays on which all of the Oracle database files were placed. IO performance was high and balanced across all the arrays.

  • The TPC-H Refresh Function (RF) simulates periodical refresh portion of Data Warehouse by adding new sales and deleting old sales data. Parallel DML (parallel insert and delete in this case) and database log performance are a key for this function and the SPARC T4-4 server outperformed both the IBM POWER7 server and HP ProLiant DL980 G7 server. (See the RF columns above.)

See Also

Disclosure Statement

TPC-H, QphH, $/QphH are trademarks of Transaction Processing Performance Council (TPC). For more information, see www.tpc.org. SPARC T4-4 205,792.0 QphH@3000GB, $4.10/QphH@3000GB, available 5/31/12, 4 processors, 32 cores, 256 threads; IBM Power 780 QphH@3000GB, 192,001.1 QphH@3000GB, $6.37/QphH@3000GB, available 11/30/11, 8 processors, 32 cores, 128 threads; HP ProLiant DL980 G7 162,601.7 QphH@3000GB, $2.68/QphH@3000GB available 10/13/10, 8 processors, 64 cores, 128 threads.

Wednesday Nov 09, 2011

SPARC T4-2 Delivers World Record SPECjvm2008 Result with Oracle Solaris 11

Oracle's SPARC T4-2 server equipped with two SPARC T4 processors running at 2.85 GHz delivered a World Record result of 454.25 SPECjvm2008 Peak ops/m on the SPECjvm2008 benchmark.

  • The SPARC T4-2 server demonstrates 41% better performance than the SPARC T3-2 server.

  • The SPARC T4-2 server with hardware cryptography acceleration greatly increases performance with subtests using AES and RSA encryption ciphers.

  • This result was produced using Oracle Solaris 11 and Oracle JDK 7 Update 2.

  • There are no SPECjvm2008 results published by IBM on POWER7 based systems.

  • The SPARC T4-2 server demonstrates Oracle's position of leadership in Java-based computing by publishing world record results for the SPECjvm2008 benchmark.

Performance Landscape

Complete benchmark results are at the SPECjvm2008 website.

SPECjvm2008 Performance Chart
(ordered by performance)
System Processors Performance
base peak
SPARC T4-2 2 x 2.85 GHz SPARC T4 - 454.25
SPARC T3-2 2 x 1.65 GHz SPARC T3 - 320.52
Sun Blade X6270 2 x 2.93 GHz Intel X5570 317.13 -

base: SPECjvm2008 Base ops/m (bigger is better)
peak: SPECjvm2008 Peak ops/m (bigger is better)

SPEC allows base and peak to be submitted separately. The base metric does not allow any optimization of the JVM, the peak metric allows optimization.

Configuration Summary

Hardware Configuration:

SPARC T4-2 server
2 x 2.85 GHz SPARC T4 processors
256 GB memory

Software Configuration:

Oracle Solaris 11 11/11
Java Platform, Standard Edition, JDK 7 Update 2

Benchmark Description

SPECjvm2008 (Java Virtual Machine Benchmark) is a benchmark suite for measuring the performance of a Java Runtime Environment (JRE), containing several real life applications and benchmarks focusing on core Java functionality. The suite focuses on the performance of the JRE executing a single application; it reflects the performance of the hardware processor and memory subsystem, but has low dependence on file I/O and includes no network I/O across machines.

The SPECjvm2008 workload mimics a variety of common general purpose application computations. These characteristics reflect the intent that this benchmark will be applicable to measuring basic Java performance on a wide variety of both client and server systems.

SPECjvm2008 benchmark highlights:

  • Leverages real life applications (like derby, sunflow, and javac) and area-focused benchmarks (like xml, serialization, crypto, and scimark).
  • Also measures the performance of the operating system and hardware in the context of executing the JRE.

The current rules for the benchmark allow either base or peak to be run. The base run is done without any tuning of the JVM to improve the out of the box performance. The peak run allows tuning of the JVM.

Key Points and Best Practices

  • Enhancements to the JVM had a major impact on performance, especially for the security tests.

See Also

Disclosure Statement

SPEC and SPECjvm are registered trademarks of Standard Performance Evaluation Corporation. Results from www.spec.org and this report as of 11/9/2011. SPARC T4-2 454.25 SPECjvm2008 Peak ops/m submitted for review, SPARC T3-2 320.52 SPECjvm2008 Peak ops/m, Sun Blade X6270 317.13 SPECjvm2008 Base ops/m.

Monday Oct 03, 2011

SPARC T4-4 Servers Set World Record on SPECjEnterprise2010, Beats IBM POWER7, Cisco x86

Oracle produced a world record SPECjEnterprise2010 benchmark result of 40,104.86 SPECjEnterprise2010 EjOPS using four of Oracle's SPARC T4-4 servers in the application tier and two more SPARC T4-4 servers for the database server.

  • The four SPARC T4-4 server configuration (sixteen SPARC T4 processors total, 3.0 GHz) demonstrated 2.4x better performance compared to the IBM Power 780 server (eight POWER7 processors, 3.86 THz) result of 16,646.34 SPECjEnterprise2010 EjOPS.

  • In the database tier, two SPARC T4-4 servers with a total of eight SPARC T4 processors at 3.0 GHz, processed 2.4x more transactions compared to the IBM result of 16,646.34 SPECjEnterprise2010 EjOPS which used four POWER7 processors at 3.55 GHz.

  • The four SPARC T4-4 server configuration demonstrated 1.5x better performance compared to the Cisco UCS B440 M2 Blade Server result of 26,118.67 SPECjEnterprise2010 EjOPS.

  • The four SPARC T4-4 server configuration demonstrated 2.3x better performance compared to the Cisco UCS B440 M1 Blade Server result of 17,301.86 SPECjEnterprise2010 EjOPS.

  • This result demonstrated less than 1 second average response times for all SPECjEnterprise2010 transactions and 90% of all transaction times took less than 1 second.

  • This result demonstrated a sustained Java EE 5 transaction load generated by approximately 320,000 users.

  • This result using 16 Oracle WebLogic 10.3.5 server instances demonstrated 4.8x better performance per application server instance when compared to the IBM result which used 32 WebSphere instances.

  • The SPARC T4-4 servers delivered a 6.7x price/performance advantage over the IBM Power 780 for the servers used in the application tier (see disclosure statement below for details). This price/performance advantage in the application tier was accomplished with a SPARC T4-4 server configuration with 2 TB of total memory compared to the IBM solution with 0.5 TB of memory.

  • The SPARC T4-4 servers had a 1.9x advantage over IBM in performance per space for the application tier (see disclosure statement below for details) even though the Oracle solution had four servers.

  • The four SPARC T4-4 servers used for the application tier used Oracle Solaris Containers to consolidate four Oracle WebLogic application server instances on each server to achieve this result.

  • The two SPARC T4-4 servers used for the database tier hosted Oracle Database 11g Release 2 and Oracle RAC cluster software using Oracle Automatic Storage Management (ASM).

  • Oracle Fusion Middleware provides a family of complete, integrated, hot pluggable and best-of-breed products known for enabling enterprise customers to create and run agile and intelligent business applications. Oracle WebLogic Server's on-going, record-setting Java application server performance demonstrates why so many customers rely on Oracle Fusion Middleware as their foundation for innovation.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results.

SPECjEnterprise2010 Performance Chart
as of 10/11/2011
Submitter EjOPS* Java EE Server DB Server
Oracle 40,104.86 4 x SPARC T4-4
4 chips, 32 cores, 3.0 GHz SPARC T4
Oracle WebLogic 11g (10.3.5)
2 x SPARC T4-4
4 chips, 32 cores, 3.0 GHz SPARC T4
Oracle 11g DB 11.2.0.2
Cisco 26,118.67 2 x Cisco UCS B440 M2
4 chips, 40 cores, 2.4 GHz Xeon E7-4870
Oracle WebLogic 11g (10.3.5)
1 x Cisco UCS C460 M2
4 chips, 40 cores, 2.4 GHz Xeon E7-4870
Oracle 11g DB 11.2.0.2
Cisco 17,301.86 2 x Cisco UCS B440 M1
4 chips, 32 cores, 2.26 GHz Xeon X7560
Oracle WebLogic 10.3.4
1 x Cisco UCS C460 M1
4 chips, 32 cores, 2.26 GHz Xeon X7560
Oracle 11g DB 11.2.0.2
IBM 16,646.34 1 x IBM Power 780
8 chips, 64 cores, 3.86 GHz POWER7
WebSphere Application Server V7.0
1 x IBM Power 750 Express
4 chips, 32 cores, 3.55 GHz POWER7
IBM DB2 Universal Database 9.7

* SPECjEnterprise2010 EjOPS (bigger is better)

Configuration Summary

Application Servers:

4 x SPARC T4-4 servers, each with
4 x 3.0 GHz SPARC T4 processors
512 GB memory
2 x 10GbE NIC
Oracle Solaris 10 8/11
Oracle WebLogic Server 11g Release 1 (10.3.5)
Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.6.0_26 (Java SE 6 Update 26)

Database Servers:

2 x SPARC T4-4 servers, each with
4 x 3.0 GHz SPARC T4 processors
1024 GB memory
2 x 10GbE NIC
4 x 8Gb FC HBA
Oracle Solaris 10 8/11
Oracle Database 11g Enterprise Edition Release 11.2.0.2
Oracle Real Application Clusters 11g Release 2

Storage Servers:

8 x Sun Fire X4270 M2 (12-Drive)
1 x 3.0 GHz Intel Xeon
8 GB memory
1 x 8Gb FC HBA
Oracle Solaris 11 Express 2010.11
8 x Sun Storage F5100 Flash Arrays

Switch Hardware:

2 x Sun Network 10GbE 72-port Top of Rack (ToR) Switch
1 x Brocade 5300 80-port Fiber Channel Switch

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The new SPECjEnterprise2010 benchmark has been re-designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems,
  • The web container, servlets, and web services
  • The EJB container
  • JPA 1.0 Persistence Model
  • JMS and Message Driven Beans
  • Transaction management
  • Database connectivity
Moreover, SPECjEnterprise2010 also heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second (SPECjEnterprise2010 EjOPS). The primary metric for the SPECjEnterprise2010 benchmark is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is NO price/performance metric in this benchmark.

Key Points and Best Practices

  • Four Oracle WebLogic server instances on each SPARC T4-4 server were hosted in 4 separate Oracle Solaris Containers to demonstrate consolidation of multiple application servers.
  • Each Oracle Solaris Container was bound to a separate processor set, each contained 7 cores (total 56 threads). This was done to improve performance by reducing memory access latency by using the physical memory closest to the processors. The default set was used for network and disk interrupt handling.
  • The Oracle WebLogic application servers were executed in the FX scheduling class to improve performance by reducing the frequency of context switches.
  • The Oracle database processes were run in 2 processor sets using psrset(1M) and executed in the FX scheduling class. This improved performance by reducing memory access latency and reducing context switches.
  • The Oracle log writer process was run in a separate processor set containing 2 threads and run in the RT scheduling class. This insured that the log writer had the most efficient use of CPU resources.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Results from www.spec.org as of 10/11/2011. SPARC T4-4, 40,104.86 SPECjEnterprise2010 EjOPS; Cisco UCS B440 M2, 26,118.67 SPECjEnterprise2010 EjOPS; Cisco UCS B440 M1, 17,301.86 SPECjEnterprise2010 EjOPS; IBM Power 780, 16,646.34 SPECjEnterprise2010 EjOPS.

SPECjEnterprise2010 models contemporary Java-based applications that run on large Java EE (Java Enterprise Edition) servers, backed by network infrastructure and database servers. Focusing on the critical Java EE server hardware & OS, the IBM result includes a Java EE server with a list price of $1.30 million. The Oracle Java EE servers have a list price of $0.47 million. The Java EE server price versus delivered EjOPS is $77.97/EjOPS for IBM versus $11.67/EjOPS for Oracle. Oracle's $/perf advantage is 6.7x better than IBM ($77.97/$11.67).

Pricing details for IBM, IBM p780 512GB based on public pricing at http://tpc.org/results/FDR/TPCH/TPC-H_1TB_IBM780_Sybase-FDR.pdf. Adjusted hardware costs to license all 64 cores. AIX pricing at: http://www-304.ibm.com/easyaccess3/fileserve?contentid=214347 and AIX Standard Edition V7.1 per processor (5765-G98-0017 64*2,600=$166,400). This gives application tier hardware & OS Price/perf: $77.97/EjOPS (1297956/16646.34)

Pricing details for Oracle, four SPARC T4-4 512 GB, HW acquisition price from Oracle's price list: $467,856 http://www.oracle.com. This gives application tier hardware & OS Price/perf: $11.67/EjOPS (467856/40104.86)

The Oracle application tier servers occupy 20U of space, 40,140.86/20=2005 EjOPS/U. The IBM application tier server occupies 16U of space, 16,646.34/16=1040 EjOPS/U. 2005/1040=1.9x

SPARC T4-4 Produces World Record Oracle OLAP Capacity

Oracle's SPARC T4-4 server delivered world record capacity on the Oracle OLAP Perf workload.

  • The SPARC T4-4 server was able to operate on a cube with a 3 billion row fact table of sales data containing 4 dimensions which represents as many as 70 quintillion aggregate rows (70 followed by 18 zeros).

  • The SPARC T4-4 server supported 3,500 cube-queries/minute against the Oracle OLAP cube with an average response time of 1.5 seconds and the median response time of 0.15 seconds.

Performance Landscape

Oracle OLAP Perf Benchmark
System Fact Table
Num of Rows
Cube-Queries/
minute
Median Response
seconds
Average Response
seconds
SPARC T4-4 3 Billion 3,500 0.15 1.5

Configuration Summary and Results

Hardware Configuration:

SPARC T4-4 server with
4 x SPARC T4 processors, 3.0 GHz
1 TB main memory
2 x Sun Storage F5100 Flash Array

Software Configuration:

Oracle Solaris 10 8/11
Oracle Database 11g Enterprise Edition with Oracle OLAP option

Benchmark Description

OLAP Perf is a workload designed to demonstrate and stress the Oracle OLAP product's core functionalities of fast query, fast update, and rich calculations on a dimensional model to support Enhanced Data Warehousing. The workload uses a set of realistic business intelligence (BI) queries that run against an OLAP cube.

Key Points and Best Practices

  • The SPARC T4-4 server is estimated to support 2,400 interactive users with this fast response time assuming only 5 seconds between query requests.

See Also

Disclosure Statement

Copyright 2011, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 10/3/2011.

Wednesday Sep 28, 2011

SPARC T4 Servers Set World Record on PeopleSoft HRMS 9.1

Oracle's SPARC T4-4 servers running Oracle's PeopleSoft HRMS Self-Service 9.1 benchmark and Oracle Database 11g Release 2 achieved World Record performance on Oracle Solaris 10.

  • Using two SPARC T4-4 servers to run the application and database tiers and one SPARC T4-2 server to run the webserver tier, Oracle demonstrated world record performance of 15,000 concurrent users running the PeopleSoft HRMS Self-Service 9.1 benchmark.

  • The combination of the SPARC T4 servers running the PeopleSoft HRMS 9.1 benchmark supports 3.8x more online users with faster response time compared to the best published result from IBM on the previous PeopleSoft HRMS 8.9 benchmark.

  • The average CPU utilization on the SPARC T4-4 server in the application tier handling 15,000 users was less than 50%, leaving significant room for application growth.

  • The SPARC T4-4 server on the application tier used Oracle Solaris Containers which provide a flexible, scalable and manageable virtualization environment.

Performance Landscape

PeopleSoft HRMS Self-Service 9.1 Benchmark
Systems Processors Users Ave Response -
Search (sec)
Ave Response -
Save (sec)
SPARC T4-2 (web)
SPARC T4-4 (app)
SPARC T4-4 (db)
2 x SPARC T4, 2.85 GHz
4 x SPARC T4, 3.0 GHz
4 x SPARC T4, 3.0 GHz
15,000 1.01 0.63
PeopleSoft HRMS Self-Service 8.9 Benchmark
IBM Power 570 (web/app)
IBM Power 570 (db)
12 x POWER5, 1.9 GHz
4 x POWER5, 1.9 GHz
4,000 1.74 1.25
IBM p690 (web)
IBM p690 (app)
IBM p690 (db)
4 x POWER4, 1.9 GHz
12 x POWER4, 1.9 GHz
6 x 4392 MPIS/Gen1
4,000 1.35 1.01

The main differences between version 9.1 and version 8.9 of the benchmark are:

  • the database expanded from 100K employees and 20K managers to 500K employees and 100K managers,
  • the manager data was expanded,
  • a new transaction, "Employee Add Profile," was added, the percent of users executing it is less then 2%, and the transaction has a heavier footprint,
  • version 9.1 has a different benchmark metric (Average Response search/save time for x number of users) versus single user search/save time,
  • newer versions of the PeopleSoft application and PeopleTools software are used.

Configuration Summary

Application Server:

1 x SPARC T4-4 server
4 x SPARC T4 processors 3.0 GHz
512 GB main memory
5 x 300 GB SAS internal disks,
2 x 100 GB internal SSDs
1 x 300 GB internal SSD
Oracle Solaris 10 8/11
PeopleSoft PeopleTools 8.51.02
PeopleSoft HCM 9.1
Oracle Tuxedo, Version 10.3.0.0, 64-bit, Patch Level 031
Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.6.0_20

Web Server:

1 x SPARC T4-2 server
2 x SPARC T4 processors 2.85 GHz
256 GB main memory
1 x 300 GB SAS internal disks
1 x 300 GB internal SSD
Oracle Solaris 10 8/11
PeopleSoft PeopleTools 8.51.02
Oracle WebLogic Server 11g (10.3.3)
Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.6.0_20

Database Server:

1 x SPARC T4-4 server
4 x SPARC T4 processors 3.0 GHz
256 GB main memory
3 x 300 GB SAS internal disks
1 x Sun Storage F5100 Flash Array (80 flash modules)
Oracle Solaris 10 8/11
Oracle Database 11g Release 2

Benchmark Description

The purpose of the PeopleSoft HRMS Self-Service 9.1 benchmark is to measure comparative online performance of the selected processes in PeopleSoft Enterprise HCM 9.1 with Oracle Database 11g. The benchmark kit is an Oracle standard benchmark kit run by all platform vendors to measure the performance. It's an OLTP benchmark with no dependency on remote COBOL calls, there is no batch workload, and DB SQLs are moderately complex. The results are certified by Oracle and a white paper is published.

PeopleSoft defines a business transaction as a series of HTML pages that guide a user through a particular scenario. Users are defined as corporate Employees, Managers and HR administrators. The benchmark consists of 14 scenarios which emulate users performing typical HCM transactions such as viewing paychecks, promoting and hiring employees, updating employee profiles and other typical HCM application transactions.

All these transactions are well-defined in the PeopleSoft HR Self-Service 9.1 benchmark kit. The benchmark metric is the Average Response Time for search and save for 15,000 users..

Key Points and Best Practices

  • The application tier was configured with two PeopleSoft application server instances on the SPARC T4-4 server hosted in two separate Oracle Solaris Containers to demonstrate consolidation of multiple application, ease of administration, and load balancing.

  • Each PeopleSoft Application Server instance running in an Oracle Solaris Container was configured to run 5 application server Domains with 30 application server instances to be able to effectively handle the 15,000 users workload with zero application server queuing and minimal use of resources.

  • The web tier was configured with 20 WebLogic instances and with 4 GB JVM heap size to load balance transactions across 10 PeopleSoft Domains. That enables equitable distribution of transactions and scaling to high number of users.

  • Internal SSDs were configured in the application tier to host PeopleSoft Application Servers object CACHE file systems and in the web tier for WebLogic servers' logging providing near zero millisecond service time and faster server response time.

See Also

Disclosure Statement

Oracle's PeopleSoft HRMS 9.1 benchmark, www.oracle.com/us/solutions/benchmark/apps-benchmark/peoplesoft-167486.html, results 9/26/2011.

Tuesday Sep 27, 2011

SPARC T4 Servers Set World Record on Siebel Loyalty Batch

Oracle's SPARC T4-2 and SPARC T4-4 servers running Oracle's Siebel Loyalty Batch engine delivered a world record result for batch processing.

  • The SPARC T4-2 and SPARC T4-4 servers running Siebel Loyalty Batch engine, part of Siebel Loyalty Solution, with Oracle Database 11g Release 2 running on Oracle Solaris 10 achieved 7.65M TPH on Accrual (Reward) processing using three Siebel Servers.

  • The world record result was achieved with 24M members and 50M records in the base transaction table.

  • Siebel Loyalty Application was configured with 50 Active Promotions with three Assign Points and four Update Attributes.

  • Oracle's Siebel Server scaled near linearly on SPARC T4 systems achieving 2.72M TPH on a single Siebel Server to 7.65M TPH with three Siebel Servers.

  • The average CPU utilization on the database tier server was 25% and on the application tier server was 65%, leaving significant room for application growth.

Performance Landscape

System Processor TPH Version
3 x SPARC T4-2 (app)
1 x SPARC T4-4 (db)
SPARC T4, 2.85 GHz
SPARC T4, 3.0 GHz
7.65M 8.1.1.1FP
2 x SPARC T3-2 (app)
1 x SPARC T3-1 (app)
1 x SPARC M5000 (db)
SPARC T3, 1.65 GHz
SPARC T3, 1.65 GHz
SPARC64 VII, 2.52 GHz
3.9M 8.1.1.1FP
Customer (app)
Customer (db)
4 x Intel E5540, 2.53 GHz
1 x Itanium, 1.6 GHz
1.5M 8.1.x

Configuration Summary

Hardware Configuration:

3 x SPARC T4-2 servers, each with
2 x SPARC T4 processors, 2.85 GHz
128 GB main memory
1 x SPARC T4-4 server with
4 x SPARC T4 processors, 3.0 GHz
256 GB main memory
1 x Sun Storage 6180 array
16 disk drives
CSM200 with 16 disk drives

Software Configuration:

Oracle Solaris 10
Siebel Server 8.1.1.1FP
Oracle Database 11g Release 2 Enterprise Edition 11.2.0.1

Benchmark Description

Siebel Loyalty enables companies to simulate and process loyalty rewards for their activities across channels and process very high volume accrual and tier assessment transactions via batch process.

The benchmark simulates a workload of Accrual Batch Transactions Processing which imports data through Enterprise Integration Manager (EIM), evaluates eligible promotion and calculates rewards. The key performance metric is transactions per hour (TPH). Key aspects of the workload simulation include:

  • Batch Engine evaluating all accrual promotions and applying all actions in one go,
  • Users do not have control over the sequence in which promotion applied,
  • Promotion actions (assign/redeem points) are rolled back in case of failure.
The number of active promotions and, in particular, the Assign Point action has very significant impact on performance. The load simulated 50 Active promotions with 3 for Assign Points and 7 Update attribute actions configured.

The number of members and the number of queued transactions in the backend database have significant impact on the performance. The benchmark had 24 million members and 52 million records in the base transaction table. The simplified process flow of the benchmark is:

  • calculate accruals base on promotions,
  • credit points to members,
  • initiate any other actions specified in promotions.

See Also

Disclosure Statement

Copyright 2011, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 9/26/2011.

SPARC T4-4 Server Sets World Record on PeopleSoft Payroll (N.A.) 9.1, Outperforms IBM Mainframe, HP Itanium

Oracle's SPARC T4-4 server achieved world record performance on the Unicode version of Oracle's PeopleSoft Enterprise Payroll (N.A) 9.1 extra-large volume model benchmark using Oracle Database 11g Release 2 running on Oracle Solaris 10.

  • The SPARC T4-4 server was able to process 1,460,544 payments/hour using PeopleSoft Payroll N.A 9.1.

  • The SPARC T4-4 server UNICODE result of 30.84 minutes on Payroll 9.1 is 2.8x faster than IBM z10 EC 2097 Payroll 9.0 (UNICODE version) result of 87.4 minutes. The IBM mainframe is rated at 6,512 MIPS.

  • The SPARC T4-4 server UNICODE result of 30.84 minutes on Payroll 9.1 is 3.1x faster than HP rx7640 Itanium2 non-UNICODE result of 96.17 minutes, on Payroll 9.0.

  • The average CPU utilization on the SPARC T4-4 server was only 30%, leaving significant room for business growth.

  • The SPARC T4-4 server processed payroll for 500,000 employees, 750,000 payments, in 30.84 minutes compared to the earlier world record result of 46.76 minutes on Oracle's SPARC Enterprise M5000 server.

  • The SPARC Enterprise M5000 server configured with eight 2.66 GHz SPARC64 VII processors has a result of 46.76 minutes on Payroll 9.1. That is 7% better than the result of 50.11 minutes on the SPARC Enterprise M5000 server configured with eight 2.53 GHz SPARC64 VII processors on Payroll 9.0. The difference in clock speed between the two processors is ~5%. That is close to the difference in the two results, thereby showing that the impact of the Payroll 9.1 benchmark on the overall result is about the same as that of Payroll 9.0.

Performance Landscape

PeopleSoft Payroll (N.A.) 9.1 – 500K Employees (7 Million SQL PayCalc, Unicode)

System OS/Database Payroll Processing
Result (minutes)
Run 1
(minutes)
Num of
Streams
SPARC T4-4, 4 x 3.0 GHz SPARC T4 Solaris/Oracle 11g 30.84 43.76 96
SPARC M5000, 8 x 2.66 GHz SPARC64 VII+ Solaris/Oracle 11g 46.76 66.28 32

PeopleSoft Payroll (N.A.) 9.0 – 500K Employees (3 Million SQL PayCalc, Non-Unicode)

System OS/Database Time in Minutes Num of
Streams
Payroll
Processing
Result
Run 1 Run 2 Run 3
Sun M5000, 8 x 2.53 GHz SPARC64 VII Solaris/Oracle 11g 50.11 73.88 534.20 1267.06 32
IBM z10 EC 2097, 9 x 4.4 GHz Gen1 Z/OS /DB2 58.96 80.5 250.68 462.6 8
IBM z10 EC 2097, 9 x 4.4 GHz Gen1 Z/OS /DB2 87.4 ** 107.6 - - 8
HP rx7640, 8 x 1.6 GHz Itanium2 HP-UX/Oracle 11g 96.17 133.63 712.72 1665.01 32

** This result was run with Unicode. The IBM z10 EC 2097 UNICODE result of 87.4 minutes is 48% slower than IBM z10 EC 2097 non-UNICODE result of 58.96 minutes, both on Payroll 9.0, each configured with nine 4.4GHz Gen1 processors.

Payroll 9.1 Compared to Payroll 9.0

Please note that Payroll 9.1 is Unicode based and Payroll 9.0 had non-Unicode and Unicode versions of the workload. There are 7 million executions of an SQL statement for the PayCalc batch process in Payroll 9.1 and 3 million executions of the same SQL statement for the PayCalc batch process in Payroll 9.0. This gets reflected in the elapsed time (27.33 min for 9.1 and 23.78 min for 9.0). The elapsed times of all other batch processes is lower (better) on 9.1.

Configuration Summary

Hardware Configuration:

SPARC T4-4 server
4 x 3.0 GHz SPARC T4 processors
256 GB memory
Sun Storage F5100 Flash Array
80 x 24 GB FMODs

Software Configuration:

Oracle Solaris 10 8/11
PeopleSoft HRMS and Campus Solutions 9.10.303
PeopleSoft Enterprise (PeopleTools) 8.51.035
Oracle Database 11g Release 2 11.2.0.1 (64-bit)
Micro Focus COBOLServer Express 5.1 (64-bit)

Benchmark Description

The PeopleSoft 9.1 Payroll (North America) benchmark is a performance benchmark established by PeopleSoft to demonstrate system performance for a range of processing volumes in a specific configuration. This information may be used to determine the software, hardware, and network configurations necessary to support processing volumes. This workload represents large batch runs typical of OLTP workloads during a mass update.

To measure five application business process run times for a database representing a large organization. The five processes are:

  • Paysheet Creation: Generates payroll data worksheets consisting of standard payroll information for each employee for a given pay cycle.

  • Payroll Calculation: Looks at paysheets and calculates checks for those employees.

  • Payroll Confirmation: Takes information generated by Payroll Calculation and updates the employees' balances with the calculated amounts.

  • Print Advice forms: The process takes the information generated by Payroll Calculations and Confirmation and produces an Advice for each employee to report Earnings, Taxes, Deduction, etc.

  • Create Direct Deposit File: The process takes information generated by the above processes and produces an electronic transmittal file that is used to transfer payroll funds directly into an employee's bank account.

Key Points and Best Practices

  • The SPARC T4-4 server with the Sun Storage F5100 Flash Array device had an average read throughput of up to 103 MB/sec and an average write throughput of up to 124 MB/sec while consuming 30% CPU on average.

  • The Sun Storage F5100 Flash Array device is a solid-state device that provides a read latency of only 0.5 msec. That is about 10 times faster than the normal disk latencies of 5 msec measured on this benchmark.

See Also

  • Oracle PeopleSoft Benchmark White Papers
    oracle.com
  • PeopleSoft Enterprise Human Capital Management (Payroll)
    oracle.com

  • PeopleSoft Enterprise Payroll 9.1 Using Oracle for Solaris (Unicode) on an Oracle's SPARC T4-4 – White Paper
    oracle.com

  • SPARC T4-4 Server
    oracle.com
  • Oracle Solaris
    oracle.com
  • Oracle Database 11g Release 2 Enterprise Edition
    oracle.com
  • Sun Storage F5100 Flash Array
    oracle.com

Disclosure Statement

Oracle's PeopleSoft Payroll 9.1 benchmark, SPARC T4-4 30.84 min,
http://www.oracle.com/us/solutions/benchmark/apps-benchmark/peoplesoft-167486.html, results 9/26/2011.

Monday Sep 12, 2011

SPARC Enterprise M9000 Produces World Record SAP ATO Benchmark

Oracle delivered an SAP enhancement package 4 for SAP ERP 6.0 Assemble-to-Order (ATO) benchmark world record result using Oracle's SPARC Enterprise M9000 server running Oracle Solaris 10 and Oracle Database 11g along with SAP Enhancement Package 4 for SAP ERP 6.0 (Unicode). The SAP ATO benchmark integrates process chains across SAP Business Suite components, include Financials, Logistics, Human Resources, Basis and Cross Application.

  • The SPARC Enterprise M9000 server containing 64 SPARC64 VII+ 3.0 GHz processors, running Oracle Solaris 10 and Oracle Database 11g along with SAP Enhancement Package 4 for SAP ERP 6.0 (Unicode) delivered a world record 206,000 fully processed assembly orders per hour on the SAP enhancement package 4 for SAP ERP 6.0 ATO benchmark.

  • The SPARC Enterprise M9000 server result shows it can more than consolidate the work of the three-tier HP solution which used 80 different servers.

  • Oracle produced the first SAP ATO benchmark result using Unicode encoding.

  • The SAP ATO benchmark uses multiple components of the SAP Business Suite. See more detail at the SAP ATO benchmark webpage.

Performance Landscape

SAP ATO 2-Tier Performance Table (select results in decreasing performance order)

System OS
Database
Assembly Orders
per hour(*)
SAP
ERP/ECC
Release
Cert Num
SPARC Enterprise M9000
64 x SPARC64 VII+ @3.0 GHz
2048 GB
Oracle Solaris 10
Oracle 11g
206,360 SAP ERP6.0*
(Unicode)
2011033
Fujitsu Siemens Primepower 2000
128 x SPARC64 @560 MHz
128 GB
Solaris 8
Oracle 8.1.7
34,260 4.6B
(non-Unicode)
2001018
HP 9000 Superdome
64 x PA-RISC 8600 @552 MHz
128 GB
HP-UX 11.11
Oracle 8.16
18,870 4.6B
(non-Unicode)
2001014
Fujitsu Siemens Primepower 900
16 x SPARC64 V @1.35 GHz
64 GB
Solaris 8
Oracle 9i
12,170 4.6C
(non-Unicode)
2003012
HP rx5670
4 x Itanium II @1.0 GHz
24 GB
HP-UX 11i
Oracle 9i
3,090 4.6C
(non-Unicode)
2002069

(*) SAP enhancement package 4 for SAP ERP6.0 (Unicode)

SAP ATO 3-Tier Performance Table (top results in decreasing performance order)

System OS
Database
Assembly Orders
per hour(*)
SAP
ERP/ECC
Release
Cert Num
HP 9000 Superdome Enterprise Server
64 x PA-RISC 8700 @ 750MHz
128 GB
HP-UX 11i
Oracle 9i
144,090 4.6 C
(non-Unicode)
2002003
HP 9000 Superdome Enterprise Server
64 x PA-RISC 8700 @750 MHz
128 GB
HP-UX 11i
Oracle 9i
130,570 4.6 C
(non-Unicode)
2001047

(*) Assembly Order: Request to assemble pre-manufactured parts and assemblies to finished products according to an existing sales order.

Complete benchmark results may be found at the SAP benchmark website: http://www.sap.com/benchmark.

Configuration Summary and Results

Hardware Configuration:

SPARC Enterprise M9000
64 SPARC64 VII+ 3.0 GHz processor
2048 GB memory

Software Configuration:

Oracle Solaris 10
SAP enhancement package 4 for SAP ERP 6.0 (Unicode)
Oracle Database 11g

Certified Result:

Fully business processed Assembly Orders/hour:
206,360
SAP Certification Number:
2011033

Benchmark Description

The SAP ATO benchmark integrates process chains across SAP Business Suite components. The ATO scenario is characterized by high volume sales, short production times (from hours to one day), and individual assembly for such products as PCs, pumps, and cars. In general, each benchmark user has its own master data, such as material, vendor, or customer master data to avoid data locking situations. However, the ATO Benchmark has been designed to handle and overcome data locking situations - the ATO benchmark users access common master data, such as material, vendor, or customer master data. (source: http://www12.sap.com/solutions/benchmark/ato.epx).

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

SAP, R/3 are registered trademarks of SAP AG in Germany and other countries. More information may be found at www.sap.com/benchmark

Two-tier SAP ATO standard SAP ERP 6.0 2005/EP4 (Unicode) application benchmarks as of 09/04/11:
Oracle's SPARC Enterprise M9000 (64 processors, 256 cores, 512 threads) 206,360 Assembly Orders/hour, 64 x 3.0 GHz SPARC VIII, 2048 GB memory, Oracle 11g, Oracle Solaris 10, Certification Number 2011033.

Two-tier SAP ATO standard 4.6 C application benchmarks as of 09/04/11:
Fujitsu Siemens Primepower 900 (16-way SMP) 12,170 Assembly Orders/hour, 16 x 1.35 GHz SPARC64 V, 64 GB memory, Oracle 9i, Solaris 8, Certification Number 2003012.
HP rx5670 (4 processors SMP) 3,090 Assembly Orders/hour, 4 x 1.0 GHz Itanium II, 24 GB memory, Oracle 9i, HP-UX 11i, Certification Number 2002069.

Two-tier SAP ATO standard 4.6 B application benchmarks as of 09/04/11:
HP 9000 Superdome (64-way SMP) 18,8770 Assembly Orders/hour, 64 x 552 MHz PA-RISC 8600, 128 GB memory, Oracle 8.1.6, HP-UX 11.11, Certification Number 2001014.
Fujitsu Siemens Primepower 2000 (128 processors SMP) 34,260 Assembly Orders/hour, 128 x 560 MHz SPARC64, 128 GB memory, Oracle 8.1.7, Solaris 8, Certification Number 2001018.

Three-tier SAP ATO standard 4.6 C application benchmarks as of 09/04/11:
HP 9000 Superdome Enterprise Server (64 processors SMP) 144,090 Assembly Orders/hour, 64 x 750 MHz PA-RISC 8700, 128 GB memory, Oracle 9i, HP-UX 11i, Certification Number 2002003
HP 9000 Superdome Enterprise Server (64 processors SMP) 130,570 Assembly Orders/hour, 64 x 750 MHz PA-RISC 8700, 128 GB memory, Oracle 9i, HP-UX 11i, Certification Number 2001047

Friday Aug 12, 2011

Sun Blade X6270 M2 with Oracle WebLogic World Record 2 Processor SPECjEnterprise 2010 Benchmark

Oracle produced a World Record single application server using 2 chips result for the SPECjEnterprise2010 benchmark of 5,427.42 SPECjEnterprise2010 EjOPS using one of Oracle's Sun Blade X6270 M2 server module for the application tier and one Sun Blade X6270 M2 server module for the database.

  • The Sun Blade X6270 M2 server module equipped with two Intel Xeon X5690 processors running at 3.46 GHz, demonstrated 47% better performance compared to the 2-chip IBM System HS22 server result of 3,694.35 SPECjEnterprise2010 EjOPS using the same model of Intel Xeon X5690 processor.

  • The Sun Blade X6270 M2 server module running the application tier demonstrated 33% better performance compared to the 2-chip IBM Power 730 Express server result of 4,062.38 SPECjEnterprise2010 EjOPS.

  • The Sun Blade X6270 M2 server modules used Oracle WebLogic Server 11g Release 1 (10.3.5) application, Java SE 6 Update 26, and Oracle Database 11g Release 2 to produce this result.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results.

SPECjEnterprise2010 Performance Chart
as of 8/11/2011
Submitter EjOPS* Application Server Database Server
Oracle 5,427.42 1x Sun Blade X6270 M2
2x 3.46 GHz Intel Xeon X5690
Oracle WebLogic 11g (10.3.5)
1x Sun Blade X6270 M2
2x 3.46 GHz Intel Xeon X5690
Oracle 11g DB 11.2.0.2
IBM 4,062.38 1x IBM Power 730 Express
2x 3.5 GHz POWER 7
WebSphere Application Server V7
1x IBM BladeCenter PS701
1x 3.0 GHz POWER 7
IBM DB2 9.7 Workgroup Server Edition FP3a
IBM 3,694.35 1x IBM HS22
2x 3.46 GHz Intel Xeon X5690
WebSphere Application Server V8
1x IBM x3850 X5
2x 2.4 GHz Intel Xeon E7-4870
IBM DB2 9.7 FP3a

* SPECjEnterprise2010 EjOPS, bigger is better.

Configuration Summary

Application Server:
    1 x Sun Blade X6270 M2
      2 x 3.46 GHz Intel Xeon X5690
      48 GB memory
      4 x 10 GbE NIC
      Oracle Linux 5 Update 6
      Oracle WebLogic Server 11g Release 1 (10.3.5)
      Java HotSpot(TM) 64-Bit Server VM on Linux, version 1.6.0_26 (Java SE 6 Update 26)

Database Server:

    1 x Sun Blade X6270 M2
      2 x 3.46 GHz Intel Xeon X5690
      144 GB memory
      2 x 10 GbE NIC
      2 x Sun Storage 6180
      Oracle Linux 5 Update 6
      Oracle Database 11g Enterprise Edition Release 11.2.0.2

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The SPECjEnterprise2010 benchmark has been designed and developed to cover the Java EE 5.0 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems.

The workload consists of an end to end web based order processing domain, an RMI and Web Services driven manufacturing domain and a supply chain model utilizing document based Web Services. The application is a collection of Java classes, Java Servlets, Java Server Pages , Enterprise Java Beans, Java Persistence Entities (pojo's) and Message Driven Beans.

The SPECjEnterprise2010 benchmark heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second ("SPECjEnterprise2010 EjOPS"). The primary metric for the SPECjEnterprise2010 benchmark is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is no price/performance metric in this benchmark.

Key Points and Best Practices

  • Two Oracle WebLogic server instances were started using numactl binding 1 instance per chip.
  • Two Oracle database listener processes were started and each was bound to a separate chip.
  • Additional tuning information is in the report at http://spec.org.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Sun Blade X6270 M2, 5,427.42 SPECjEnterprise2010 EjOPS; IBM Power 730 Express, 4,062.38 SPECjEnterprise2010 EjOPS; IBM System HS22, 3,694.35 SPECjEnterprise2010 EjOPS. Results from www.spec.org as of 8/11/2011.

Friday Jun 10, 2011

SPARC Enterprise M5000 Delivers First PeopleSoft Payroll 9.1 Benchmark

Oracle's M-series server sets a world record on Oracle's PeopleSoft Enterprise Payroll (N.A) 9.1 with extra large volume model benchmark (Unicode). Oracle's SPARC Enterprise M5000 server was able to to run faster than the previous generation system result even though the PeopleSoft Payroll 9.1 benchmark is more computationally demanding.

Oracle's SPARC Enterprise M5000 server configured with eight 2.66 GHz SPARC64 VII+ processors together with Oracle's Sun Storage F5100 Flash Array storage achieved world record performance on the Unicode version of Oracle's PeopleSoft Enterprise Payroll (N.A) 9.1 with extra large volume model benchmark using Oracle Database 11g Release 2 running on Oracle Solaris 10.

  • The SPARC Enterprise M5000 server processed payroll payments for the 500K employees PeopleSoft Payroll 9.1 (Unicode) benchmark in 46.76 minutes compared to a previous result of 50.11 minutes for the PeopleSoft Payroll 9.0 (non-Unicode) benchmark configured with 2.53 GHz SPARC64 VII processors resulting in 7% better performance.

  • Note that the IBM z10 Gen1 mainframe running the PeopleSoft Payroll 9.0 (Unicode) benchmark was 48% slower than the 9.0 non-Unicode version. The IBM z10 mainframe with nine 4.4 GHz Gen1 processors has a list price over $6M and is rated at 6,512 MIPS.

  • The SPARC Enterprise M5000 server with the Sun Storage F5100 Flash Array system processed payroll for 500K employees completing the end-to-end run in 66.28 mins, 11% faster than earlier published result of 73.88 mins with Payroll 9.0 configured with 2.53 GHz SPARC64 VII processors.

  • The Sun Storage F5100 Flash Array device is a high performance, high-density solid-state flash array which provides a read latency of only 0.5 msec which is about 10 times faster than the normal disk latencies of 5 msec measured on this benchmark.

Performance Landscape

PeopleSoft Payroll (N.A.) 9.1 – 500K Employees (7 Million SQL PayCalc, Unicode)

System Processor OS/Database Payroll Processing
Result (minutes)
Run 1
(minutes)
Num of
Streams
SPARC M5000 8x 2.66GHz SPARC64 VII+ Solaris/Oracle 11g 46.76 66.28 32

PeopleSoft Payroll (N.A.) 9.0 – 500K Employees (3 Million SQL PayCalc, Non-Unicode)

System Processor OS/Database Time in Minutes Num of
Streams
Payroll
Processing
Result
Run 1 Run 2 Run 3
Sun M5000 8x 2.53GHz SPARC64 VII Solaris/Oracle 11g 50.11 73.88 534.20 1267.06 32
IBM z10 9x 4.4GHz Gen1 Z/OS /DB2 58.96 80.5 250.68 462.6 8
IBM z10 9x 4.4GHz Gen1 Z/OS /DB2 87.4 ** 107.6 - - 8
HP rx7640 8x 1.6GHz Itanium2 HP-UX/Oracle 11g 96.17 133.63 712.72 1665.01 32

** This result was run with Unicode

Payroll 9.1 Compared to Payroll 9.0

Please note that Payroll 9.1 is Unicode based and Payroll 9.0 is non-Unicode. There are 7 million executions of an SQL statement for the PayCalc batch process in Payroll 9.1 and 3 million executions of the same SQL statement for the PayCalc batch process in Payroll 9.0. This gets reflected in the elapsed time (27.33 min for 9.1 and 23.78 min for 9.0). The elapsed times of all other batch processes is lower (better) on 9.1.

Configuration Summary

Hardware Configuration:

SPARC Enterprise M5000 server
8 x 2.66 GHz SPARC64 VII+ processors
128 GB memory
2 x SAS HBA (SG-XPCIE8SAS-E-Z - PCIe HBA for Rack Servers)
Sun Storage F5100 Flash Array
40 x 24 GB FMODs
1 x StorageTek 2501 array with
12 x 146 GB SAS 15K RPM disks
1 x StorageTek 2540 array with
12 x 146 GB SAS 15K RPM disks

Software Configuration:

Oracle Solaris 10 09/10
PeopleSoft HRMS and Campus Solutions 9.10.303
PeopleSoft Enterprise (PeopleTools) 8.51.035
Oracle Database 11g Release 2 11.2.0.1 (64-bit)
Micro Focus COBOLServer Express 5.1 (64-bit)

Benchmark Description

The PeopleSoft 9.1 Payroll (North America) benchmark is a performance benchmark established by PeopleSoft to demonstrate system performance for a range of processing volumes in a specific configuration. This information may be used to determine the software, hardware, and network configurations necessary to support processing volumes. This workload represents large batch runs typical of OLTP workloads during a mass update.

To measure five application business process run times for a database representing a large organization. The five processes are:

  • Paysheet Creation: Generates payroll data worksheets consisting of standard payroll information for each employee for a given pay cycle.

  • Payroll Calculation: Looks at paysheets and calculates checks for those employees.

  • Payroll Confirmation: Takes information generated by Payroll Calculation and updates the employees' balances with the calculated amounts.

  • Print Advice forms: The process takes the information generated by Payroll Calculations and Confirmation and produces an Advice for each employee to report Earnings, Taxes, Deduction, etc.

  • Create Direct Deposit File: The process takes information generated by the above processes and produces an electronic transmittal file that is used to transfer payroll funds directly into an employee's bank account.

For the benchmark, we collected at least three data points with different numbers of job streams (parallel jobs). This batch benchmark allows a maximum of thirty-two job streams to be configured to run in parallel.

See Also

Disclosure Statement

Oracle's PeopleSoft Payroll 9.1 benchmark, SPARC Enterprise M5000 46.76 min, www.oracle.com/apps_benchmark/html/white-papers-peoplesoft.html, results 6/10/2011.

Friday Mar 25, 2011

SPARC Enterprise M9000 with Oracle Database 11g Delivers World Record Single Server TPC-H @3000GB Result

Oracle's SPARC Enterprise M9000 server delivers single-system TPC-H @3000GB world record performance. The SPARC Enterprise M9000 server along with Oracle's Sun Storage 6180 arrays and running Oracle Database 11g Release 2 on the Oracle Solaris operating system proves the power of Oracle's integrated solution.

  • The SPARC Enterprise M9000 server configured with SPARC64 VII+ processors, Sun Storage 6180 arrays and running Oracle Solaris 10 combined with Oracle Database 11g Release 2 achieved World Record TPC-H performance of 386,478.3 QphH@3000GB for non-clustered systems.

  • The SPARC Enterprise M9000 server running the Oracle Database 11g Release 2 software is 2.5 times faster than the IBM p595 (POWER6) server which ran with Sybase IQ v.15.1 database software.

  • The SPARC Enterprise M9000 server is 3.4 times faster than the IBM p595 server for data loading.

  • The SPARC Enterprise M9000 server is 3.5 times faster than the IBM p595 server for Refresh Function.

  • The SPARC Enterprise M9000 server configured with Sun Storage 6180 arrays shows linear scaling up to the maximum delivered IO performance of 48.3 GB/sec as measured by vdbench.

  • The SPARC Enterprise M9000 server running the Oracle Database 11g Release 2 software is 2.4 times faster than the HP ProLiant DL980 server which used Microsoft SQL Server 2008 R2 Enterprise Edition software.

  • The SPARC Enterprise M9000 server is 2.9 times faster than the HP ProLiant DL980 server for data loading.

  • The SPARC Enterprise M9000 server is 4 times faster than the HP ProLiant DL980 server for Refresh Function.

  • A 1.94x improvement was delivered by the SPARC Enterprise M9000 server result using 64 SPARC64 VII+ processors compared to the previous Sun SPARC Enterprise M9000 server result which used 32 SPARC64 VII processes.

  • Oracle's TPC-H result shows that the SPARC Enterprise M9000 server can handle the increasingly large databases required of DSS systems. The IO rate as measured by the Oracle database is over 40 GB/sec.

  • Oracle used Storage Redundancy Level 3 as defined by the TPC-H 2.14.0 specification which is the highest level.

Performance Landscape

TPC-H @3000GB, Non-Clustered Systems

System
CPU type
Memory
Composite
(QphH)
$/perf
($/QphH)
Power
(QppH)
Throughput
(QthH)
Database Available
SPARC Enterprise M9000
3 GHz SPARC64 VII+
1024 GB
386,478.3 $18.19 316,835.8 471,428.6 Oracle 11g 09/22/11
Sun SPARC Enterprise M9000
2.88 GHz SPARC64 VII
512 GB
198,907.5 $15.27 182,350.7 216,967.7 Oracle 11g 12/09/10
HP ProLiant DL980 G7
2.27 GHz Intel Xeon X7560
512 GB
162,601.7 $2.68 185,297.7 142,601.7 SQL Server 10/13/10
IBM Power 595
5.0 GHz POWER6
512 GB
156,537.3 $20.60 142,790.7 171,607.4 Sybase 11/24/09

QphH = the Composite Metric (bigger is better)
$/QphH = the Price/Performance metric (smaller is better)
QppH = the Power Numerical Quantity
QthH = the Throughput Numerical Quantity

Complete benchmark results found at the TPC benchmark website http://www.tpc.org.

Configuration Summary and Results

Server:

SPARC Enterprise M9000
64 x SPARC VII+ 3.0 GHz processors
1024 GB memory
4 x internal SAS (4 x 146 GB)

External Storage:

32 x Sun Storage 6180 arrays (each with 16 x 600 GB)

Software:

Oracle Solaris 10 9/10
Oracle Database 11g Release 2 Enterprise Edition

Audited Results:

Database Size: 3000 GB (Scale Factor 3000)
TPC-H Composite: 386,478.3 QphH@3000GB
Price/performance: $18.19/QphH@3000GB
Available: 09/22/2011
Total 3 year Cost: $7,030,009
TPC-H Power: 316,835.8
TPC-H Throughput: 471,428.6
Database Load Time: 2:59:01

Benchmark Description

The TPC-H benchmark is a performance benchmark established by the Transaction Processing Council (TPC) to demonstrate Data Warehousing/Decision Support Systems (DSS). TPC-H measurements are produced for customers to evaluate the performance of various DSS systems. These queries and updates are executed against a standard database under controlled conditions. Performance projections and comparisons between different TPC-H Database sizes (100GB, 300GB, 1000GB, 3000GB and 10000GB) are not allowed by the TPC.

TPC-H is a data warehousing-oriented, non-industry-specific benchmark that consists of a large number of complex queries typical of decision support applications. It also includes some insert and delete activity that is intended to simulate loading and purging data from a warehouse. TPC-H measures the combined performance of a particular database manager on a specific computer system.

The main performance metric reported by TPC-H is called the TPC-H Composite Query-per-Hour Performance Metric (QphH@SF, where SF is the number of GB of raw data, referred to as the scale factor). QphH@SF is intended to summarize the ability of the system to process queries in both single and multi user modes. The benchmark requires reporting of price/performance, which is the ratio of QphH to total HW/SW cost plus 3 years maintenance.

Key Points and Best Practices

  • The Sun Storage 6180 array showed linear scalability of 48.3 GB/sec Sequential Read with thirty-two Sun Storage 6180 arrays. Scaling could continue if there are more arrays available.
  • Oracle Solaris 10 9/10 required very little system tuning.
  • The optimal Sun Storage 6180 arrays configuration for the benchmark was to set up 1 disk per volume instead of multiple disks per volume and let Oracle Oracle Automatic Storage Management (ASM) mirror. Presenting as many volumes as possible to Oracle database gave the highest scan rate.

  • The storage was managed by ASM with 4 MB stripe size. 1 MB is the default stripe size but 4 MB works better for large databases.

  • All the Oracle database files, except TEMP tablespace, were mirrored under ASM. 16 x Sun Storage 6180 arrays (256 disks) were mirrored to another 16 x Sun Storage 6180 arrays using ASM. IO performance was good and balanced across all the disks. With the ASM mirror the benchmark passed the ACID (Atomicity, Consistency, Isolation and Durablity) test.

  • Oracle database tables were 256-way partitioned. The parallel degree for each table was set to 256 to match the number of available cores. This setting worked the best for performance.

  • Oracle Database 11g Release 2 feature Automatic Parallel Degree Policy was set to AUTO for the benchmark. This enabled automatic degree of parallelism, statement queuing and in-memory parallel execution.

See Also

Disclosure Statement

SPARC Enterprise M9000 386,478.3 QphH@3000GB, $18.19/QphH@3000GB, avail 09/22/11, IBM Power 595 QphH@3000GB, 156,537.3 QphH@3000GB, $20.60/QphH@3000GB, avail 11/24/09, HP ProLiant DL980 G7 162,601.7 QphH@3000GB, $2.68/QphH@3000GB avail 10/13/10, TPC-H, QphH, $/QphH tm of Transaction Processing Performance Council (TPC). More info www.tpc.org.

Thursday Dec 02, 2010

World Record TPC-C Result on Oracle's SPARC Supercluster with T3-4 Servers

Oracle demonstrated the world's fastest database performance using 27 of Oracle's SPARC T3-4 servers, 138 Sun Storage F5100 Flash Array storage systems and Oracle Database 11g Release 2 Enterprise Edition with Real Application Clusters (RAC) and Partitioning delivered a world-record TPC-C benchmark result.

  • The SPARC T3-4 server cluster delivered a world record TPC-C benchmark result of 30,249,688 tpmC and $1.01 $/tpmC (USD) using Oracle Database 11g Release 2 on a configuration available 6/1/2011.

  • The SPARC T3-4 server cluster is 2.9x faster than the performance of the IBM Power 780 (POWER7 3.86 GHz) cluster with IBM DB2 9.7 database and has 27% better price/performance on the TPC-C benchmark. Almost identical price discount levels were applied by Oracle and IBM.

  • The Oracle solution has three times better performance than the IBM configuration and only used twice the power during the run of the TPC-C benchmark.  (Based upon IBM's own claims of energy usage from their August 17, 2010 press release.)

  • The Oracle solution delivered 2.9x the performance in only 71% of the space compared to the IBM TPC-C benchmark result.

  • The SPARC T3-4 server with Sun Storage F5100 Flash Array storage solution demonstrates 3.2x faster response time than IBM Power 780 (POWER7 3.86 GHz) result on the TPC-C benchmark.

  • Oracle used a single-image database, whereas IBM used 96 separate database partitions on their 3-node cluster. It is interesting to note that IBM used 32 database images instead of running each server as a simple SMP.

  • IBM did not use DB2 Enterprise Database, but instead IBM used "DB2 InfoSphere Warehouse 9.7" which is a data warehouse and data management product and not their flagship OLTP product.

  • The multi-node SPARC T3-4 server cluster is 7.4x faster than the HP Superdome (1.6 GHz Itanium2) solution and has 66% better price/performance on the TPC-C benchmark.

  • The Oracle solution utilized Oracle's Sun FlashFire technology to deliver this result. The Sun Storage F5100 Flash Array storage system was used for database storage.

  • Oracle Database 11g Enterprise Edition Release 2 with Real Application Clusters and Partitioning scales and effectively uses all of the nodes in this configuration to produce the world record TPC-C benchmark performance.

  • This result showed Oracle's integrated hardware and software stacks provide industry leading performance.

Performance Landscape

TPC-C results (sorted by tpmC, bigger is better)

System tpmC Price/tpmC Avail Database Cluster Racks
27 x SPARC T3-4 30,249,688 1.01 USD 6/1/2011 Oracle 11g RAC Y 15
3 x IBM Power 780 10,366,254 1.38 USD 10/13/10 DB2 9.7 Y 10
HP Integrity Superdome 4,092,799 2.93 USD 08/06/07 Oracle 10g R2 N 46

Avail - Availability date
Racks - Clients, servers, storage, infrastructure

Oracle and IBM TPC-C Response times

System tpmC Response Time (sec)
New Order 90th%
Response Time (sec)
New Order Average
27 x SPARC T3-4 30,249,688 0.750 0.352
3 x IBM Power 780 10,366,254 2.1 1.137
Response Time Ratio - Oracle Better 2.9x 2.8x 3.2x

Oracle uses Average New Order Response time for comparison between Oracle and IBM.

Graphs of Oracle's and IBM's response times for New-Order can be found in the full disclosure reports on TPC's website TPC-C Official Result Page.

Configuration Summary and Results

Hardware Configuration:

15 racks used to hold

Servers
27 x SPARC T3-4 servers, each with
4 x 1.65 GHz SPARC T3 processors
512 GB memory
3 x 300 GB 10K RPM 2.5" SAS disks

Data Storage
69 x Sun Fire X4270 M2 servers configured as COMSTAR heads, each with
1 x 2.93 GHz Intel Xeon X5670 processor
8 GB memory
9 x 2 TB 7.2K RPM 3.5" SAS disks
2 x Sun Storage F5100 Flash Array storage (1.92 TB each)
1 x Brocade DCX switch

Redo Storage
28 x Sun Fire X4270 M2 servers configured as COMSTAR heads, each with
1 x 2.93 GHz Intel Xeon X5670 processor
8 GB memory
11 x 2 TB 7.2K RPM 3.5" SAS disks
2 x Brocade 5300 switches

Clients
81 x Sun Fire X4170 M2 servers, each with
2 x 2.93 GHz Intel X5670 processors
48 GB memory
2 x 146 GB 10K RMP 2.5" SAS disks

Software Configuration:

Oracle Solaris 10 9/10 (for SPARC T3-4 and Sun Fire X4170 M2)
Oracle Solaris 11 Express (COMSTAR for Sun Fire X4270 M2)
Oracle Database 11g Release 2 Enterprise Edition with Real Application Clusters and Partitioning
Oracle iPlanet Web Server 7.0 U5
Tuxedo CFS-R Tier 1

Results:

System 27 x SPARC T3-4
tpmC 30,249,688
Price/tpmC 1.01 USD
Avail 6/1/2011
Database Oracle Database 11g RAC
Cluster yes
Racks 15
New Order Ave Response 0.352 seconds

Benchmark Description

TPC-C is an OLTP system benchmark. It simulates a complete environment where a population of terminal operators executes transactions against a database. The benchmark is centered around the principal activities (transactions) of an order-entry environment. These transactions include entering and delivering orders, recording payments, checking the status of orders, and monitoring the level of stock at the warehouses.

Key Points and Best Practices

  • Oracle Database 11g Release 2 Enterprise Edition with Real Application Clusters and Partitioning scales easily to this high level of performance.

  • Sun Storage F5100 Flash Array storage provides high performance, very low latency, and very high storage density.

  • COMSTAR (Common Multiprotocol SCSI Target), new in Oracle Solaris 11 Express, is the software framework that enables a Solaris host to serve as a SCSI Target platform. COMSTAR uses a modular approach to break the huge task of handling all the different pieces in a SCSI target subsystem into independent functional modules which are glued together by the SCSI Target Mode Framework (STMF). The modules implementing functionality at SCSI level (disk, tape, medium changer etc.) are not required to know about the underlying transport. And the modules implementing the transport protocol (FC, iSCSI, etc.) are not aware of the SCSI-level functionality of the packets they are transporting. The framework hides the details of allocation providing execution context and cleanup of SCSI commands and associated resources and simplifies the task of writing the SCSI or transport modules.

  • Oracle iPlanet Web Server 7.0 U5 is used in the user tier of the benchmark with each of the web server instance supporting more than a quarter-million users, while satisfying the stringent response time requirement from the TPC-C benchmark.

See Also

Disclosure Statement

TPC Benchmark C, tpmC, and TPC-C are trademarks of the Transaction Processing Performance Council (TPC). 27-node SPARC T3-4 Cluster (4 x 1.65 GHz SPARC T3 processors) with Oracle Database 11g Release 2 Enterprise Edition with Real Application Clusters and Partitioning, 30,249,688 tpmC, $1.01/tpmC, Available 6/1/2011. IBM Power 780 Cluster (3 nodes using 3.86 GHz POWER7 processors) with IBM DB2 InfoSphere Warehouse Ent. Base Ed. 9.7, 10,366,254 tpmC, $1.38 USD/tpmC, available 10/13/2010. HP Integrity Superdome(1.6GHz Itanium2, 64 processors, 128 cores, 256 threads) with Oracle 10g Enterprise Edition, 4,092,799 tpmC, $2.93/tpmC, available 8/06/07. Energy claims based upon IBM calculations and internal measurements. Source: http://www.tpc.org/tpcc, results as of 11/22/2010

World Record SPECweb2005 Result on SPARC T3-2 with Oracle iPlanet Web Server

Oracle's SPARC T3-2 server running Oracle iPlanet Web Server middleware delivered a world record SPECweb2005 benchmark result of 113,857. Oracle's 2-socket SPARC is 9% faster than the fastest 2-socket x86-based competitive server and even 8% faster than the 4-socket HP x86-based server.

  • The SPARC T3-2 server with dual 1.65 GHz SPARC T3 processors using Oracle iPlanet Web Server 7.0.9 middleware delivered a world record result of 113857 on the SPECweb2005 benchmark.

  • This result demonstrates that the SPARC T3-2 running Oracle Solaris and Oracle iPlanet Web Server can support thousands of concurrent web server sessions and is an industry leader in web serving with a high performance and enterprise quality solution.

  • Oracle is the only SPECweb2005 benchmark sponsor who can demonstrate top performance using a commercially viable and production quality web serving solution with the Oracle iPlanet Web Server and the Oracle Solaris 10 operating system.

  • On the SPECweb2005 benchmark, the SPARC T3-2 server with two 1.65 GHz SPARC T3 processors is 8% faster than the latest Hewlett-Packard result that was just published on the HP ProLiant DL585 G7 with four 2.0 GHz AMD 6128HE processors.

  • On the SPECweb2005 benchmark, the SPARC T3-2 server with two 1.65 GHz SPARC T3 processors is 9% faster than the Fujitsu PRIMERGY TX300 S6 with two 3.3 GHz Intel X5680 processors.

  • On the SPECweb2005 benchmark, the SPARC T3-2 server with two 1.65 GHz SPARC T3 processors is 37% faster than the HP ProLiant DL370 G6 with two 3.2 GHz Intel W5580 processors.

  • On the Support workload of SPECweb2005, the SPARC T3-2 server with two 1.65 GHz SPARC T3 processors obtained a 41% higher score than the Fujitsu PRIMERGY TX300 S6 with two 3.3 GHz Intel X5680 processors.

  • The SPARC T3-2 server obtained 14.4 times the result of the 4-core IBM System p5 550 1.9 GHz POWER5+ system on the SPECweb2005 benchmark. There are no IBM POWER7 or POWER6 based system results published on the SPECweb2005 benchmark.

Performance Landscape

SPECweb2005 select results as of 8 December 2010. See the SPEC website for more. Information ordered by Result, bigger is better.

Server Processor OS SPECweb2005 Performance (\*) Web Server
Result Bank Ecom Supp
SPARC T3-2 2 x 1.65 T3 Solaris 113857 165024 160056 123840 iPlanet
HP DL585 G7 4 x 2.0 6128HE RedHat Linux 105586 168192 175104 88576 Rock
Fujitsu TX300 S6 2 x 3.33 X5680 RedHat Linux 104422 162000 177000 88000 Rock
Sun T5440 4 x 1.6 T2 Plus Solaris 100209 176500 133000 95000 Sun
Fujitsu TX300 S5 2 x 2.93 X5570 RedHat Linux 83198 106000 140000 86000 Rock
HP ML370 G6 2 x 3.2 W5580 RedHat Linux 83073 117120 142080 76352 Rock
HP DL370 G6 2 x 3.2 W5580 RedHat Linux 83073 117120 142080 76352 Rock
HP DL585 G5 4 x 3.1 Opt8393 RedHat Linux 71629 117504 123072 56320 Rock
IBM p5 550 2 x 1.9 POWER5+ SuSE Linux 7881 12240 11820 7500 Zeus

(\*) Metrics are
Result - SPECweb2005, overall metric
Bank - SPECweb2005_banking, Banking component metric
Ecom - SPECweb2005_ecommerce, ECommerce component metric
Supp - SPECweb2005_support, Support component metric

Configuration Summary

Hardware Configuration:

1 SPARC T3-2 with
2 x 1.65 GHz SPARC T3 processors
256 GB memory
2 x Sun Storage F5100 Flash Array
4 x Dual 10 GbE SFP+ PCIe LP
4 x 6 GB SAS PCIe HBA

Software Configuration:

Oracle Solaris 10 9/10
Oracle iPlanet Web Server 7.0.9
Java Platform, Standard Edition version 1.6.0_21-b06
Java Hotspot Server VM version 17.0-b16, mixed mode

Benchmark Description

SPECweb2005, successor to SPECweb99 and SPECweb99_SSL, is an industry standard benchmark for evaluating Web Server performance developed by SPEC. The benchmark simulates multiple user sessions accessing a Web Server and generating static and dynamic HTTP requests. The major features of SPECweb2005 are:

  • Measures simultaneous user sessions
  • Dynamic content: currently PHP and JSP implementations
  • Page images requested using 2 parallel HTTP connections
  • Multiple, standardized workloads: Banking (HTTPS), E-commerce (HTTP and HTTPS), and Support (HTTP)
  • Simulates browser caching effects
  • File accesses more accurately simulate today's disk access patterns

SPEC requires the server under test to support SSL Protocol V3 (SSLv3).

Of the various ciphers supported in SSLv3, cipher SSL_RSA_WITH_RC4_128_MD5 is currently required for all workload components that use SSL. It was selected as one of the most commonly used SSLv3 ciphers and allows results to be directly compared to each other. SSL_RSA_WITH_RC4_128_MD5 consists of:

  • RSA public key (asymmetric) encryption with a 1024-bit key
  • RC4 symmetric encryption with a 128-bit key for bulk data encryption
  • MD5 digest algorithm with 128-bit output for the Message Authentication Code (MAC)

A compliant result must use the cipher suite listed above, and must employ the 1024 bit key for RSA public key encryption, 128-bit key for RC4 bulk data encryption, and have a 128-bit output for the Message Authentication code.

All Banking workload request to the server under test use SSL, where the Ecommerce workload requests are a mix of SSL and non-SSL. Non of the Support workload requests to server under test use SSL.

Key Points and Best Practices

  • When multiple 10 GbE Dual Port NICs are used, it is best practice to equally divide these NICs between PCI root nodes that are available.

  • Two web server instances was used. One web server instance was bound to a processor set with CPUs in the first processor chip. The other web server instance was bound to a processor set with CPUs in the second processor chip. The web server instance bound to CPUs in first processor chip was listening on the NIC IP addresses on that processor's chip PCI root node. The same was done with web server instance bound to CPUs in second processor chip. This was done to improve the locality of the processing.

  • Each web server is executed in the FX scheduling class to improve performance by reducing the frequency of context switches.

See Also

Disclosure Statement

SPEC and the benchmark name SPECweb are registered trademarks of Standard Performance Evaluation Corporation. Results are from www.spec.org as of December 8, 2010 and this report. Oracle, SPARC T3-2, 113,857 SPECweb2005. HP ProLiant DL585 G7, 105,586 SPECweb2005. Fujitsu PRIMERGY TX300 S6, 104,422 SPECweb2005. Sun SPARC Enterprise T5440, 100,209 SPECweb2005. Fujitsu PRIMERGY TX300 S5, 83,198 SPECweb2005. HP ProLiant ML370 G6, 83,073 SPECweb2005. HP ProLiant DL370 G6, 83,073 SPECweb2005. HP ProLiant DL585 G5, 71,629 SPECweb2005. IBM System p5 550, 7,881 SPECweb2005.

World Record Performance on PeopleSoft Enterprise Financials Benchmark run on Sun SPARC Enterprise M4000 and M5000

Oracle's Sun SPARC Enterprise M4000 and M5000 servers have combined to produce a world record result on Oracle's PeopleSoft Enterprise Financial Management 9.0 benchmark.

  • The Sun SPARC Enterprise M4000 and M5000 servers configured with SPARC64 VII+ processors along with Oracle's Sun Storage F5100 Flash Array system achieved a world record result using PeopleSoft Enterprise Financial Management and Oracle Database 11g Release 2 software running on the Oracle Solaris 10 operating system.

  • The PeopleSoft Enterprise Financial Management solution processed online business transactions to support 1000 concurrent users using 32 application server threads with compliant response times while simultaneously completing complex batch jobs in record time.

  • The Sun Storage F5100 Flash Array system is a high performance, high-density solid-state flash array which provides a read latency of only 0.5 msec which is about 10 times faster than the normal disk latencies of 5 msec measured on this benchmark.

  • The Sun SPARC Enterprise M4000 and M5000 servers were able to process online users and concurrent batch jobs simultaneously in 34.72 minutes on this benchmark that reflects complex, multi-tier environment and utilizes a large back-end database of nearly 1 TB.

  • The combination of Oracle's PeopleSoft Enterprise Financial Management 9.00.00.331, PeopleSoft PeopleTools 8.49.23 and Oracle WebLogic server was run on the Sun SPARC Enterprise M4000 server and Oracle database 11g Release 2 was run on the Sun SPARC Enterprise M5000 server for this benchmark.

Performance Landscape

The following table discloses the current and the single previously disclosed result for this benchmark. Results are elapsed times therefore the smaller number is better.

Servers CPU Tier Batch (mins) Batch
w/Online (mins)
Sun SPARC Enterprise M4000 2.66 GHz SPARC64 VII+ Web/App
33.09
34.72
Sun SPARC Enterprise M5000 2.66 GHz SPARC64 VII+ DB

SPARC T3-1 1.65 GHz SPARC T3 Web/App 35.82 37.01
Sun SPARC Enterprise M5000 2.5 GHz SPARC64 VII DB

Configuration Summary

Web/Application Tier Configuration:

1 x Sun SPARC Enterprise M4000
4 x 2.66 GHz SPARC64 VII+ processors
128 GB of memory

Database Tier Configuration:

1 x Sun SPARC Enterprise M5000
8 x 2.66 GHz SPARC64 VII+ processors
128 GB of memory
1 x Sun Storage F5100 Flash Array (74 x 24 GB FMODs)
2 x StorageTek 2540 (12 x 146 GB SAS 15K RPM)
1 x StorageTek 2501 (12 x 146 GB SAS 15K RPM)
1 x Dual-Port SAS Fibre Channel Host Bus Adapters (HBA)

Software Configurations:

Oracle Solaris 10 10/09
PeopleSoft Enterprise Financial Management/SCM 9.00.00.311 64-bit
PeopleSoft Enterprise (PeopleTools) 8.49.23 64-bit
Oracle Database 11g Release 2 11.1.0.6 64-bit
Oracle Tuxedo 9.1 RP36 with Jolt 9.1
Micro Focus COBOL Server Express 4.0 SP4 64-bit

Benchmark Description

This Day-in-the-Life benchmark measured the concurrent batch and online performance for a large database model. This scenario more accurately represents a production environment where users and scheduled batch jobs must run concurrently. This benchmark measured performance results during a Close-the-Books process.

The PeopleSoft Enterprise Financials 9 batch processes included in this benchmark are as follows:

  • Journal Generator: (AE) This process creates journals from accounting entries (AE) generated from various data sources, including non-PeopleSoft systems as well as PeopleSoft applications. In the benchmark, the Journal Generator (FS_JGEN) process is set up to create accounting entries from Oracle's PeopleSoft applications in the same database, such as PeopleSoft Enterprise Payables, Receivables, Asset Management, Expenses, Cash Management. The process is run with the option of Edit and Post turned on to edit and post the journals created by Journal generator. Journal Edit is an AE program and Post is a COBOL program.

  • Allocation: (AE) This process allocates balances held or accumulated in one or more entities to more than one business unit, department or other entities based on user-defined rules.

  • Journal Edit & Post: (AE & COBOL) Journal Edit validates journal transactions before posting them to the ledger. This validation ensures that journals are valid, for example: valid ChartFields values and combinations, debits and credits equal, and inter/intra-unit balanced, Journal Post process posts only valid, edited journals, ensures each journal line posts to the appropriate target detail ledgers, and then changes the journal's status to posted. In this benchmark, the Journal Edit & Post is also set up to edit and post Oracle's PeopleSoft applications from another database, such as PeopleSoft Enterprise Payroll data.

  • Summary Ledger: (AE) Summary Ledger processing summarizes detail ledger data across selected GL BUs. Summary Ledgers can be generated for reporting purposes or used in consolidations.

  • Consolidations: (COBOL) Consolidation processing summarizes ledger balances and generates elimination journal entries across business units based on user-defined rules.

  • SQR & nVision Reporting: Reporting will consist of nVision and SQR reports. A balance sheet, an income statement, and a trial balance will be generated for each GL BU by SQR processes GLS7002 and GLS7012. The consolidated results of the nVision reports are run by 10 nVision users using 4 standard delivered report request definitions such as BALANCE, INCOME, CONSBAL, and DEPTINC. Each of the nVision users will have ownership over 10 Business Units and each of the nVision users will submit multiple runs that are being executed in parallel to generate a total of 40 nVision reports.

Batch processes are run concurrently with more than 1000 emulated users executing 30 pre-defined online applications. Response times for the online applications are collected and must conform to a maximum time.

Key Points and Best Practices

The Sun SPARC Enterprise M4000 and M5000 servers were able process online users and concurrent batch jobs simultaneously in 34.72 minutes.

The Sun Storage F5100 Flash Array system, which is highly tuned for IOPS, contributed to the result through reduced IO latency.

The family of Sun SPARC Enterprise M-series servers, with Sun Storage F5100 Flash Array systems, form an ideal environment for hosting complex multi-tier applications. This is the second public disclosure of any system running this benchmark.

The Sun SPARC Enterprise M4000 server hosted the web and application server tiers providing good response time to emulated user requests. The benchmark specification allows 1000 users, but there is headroom for increased load.

The Sun SPARC Enterprise M5000 server was used for the database server along with a Sun Storage F5100 Flash Array system. The speed of the M-series server with the low latency of the Flash Array provided the overall low latency for user requests, even while completing complex batch jobs.

Despite the systems being lightly loaded, the increased frequency of the SPARC64 VII+ processors yielded lower latencies and faster elapsed times than previously disclosed results.

The low latency of the Sun Storage F5100 Flash Array storage contributed to the excellent response times of emulated users by making data quickly available to the database back-end. The array was configured as several RAID 0 volumes and data was distributed across the volumes, maximizing storage bandwidth.

The transaction processing capacity of the Sun SPARC Enterprise M5000 server enabled very fast batch processing times while supporting over 1000 online users.

While running the maximum workload specified by the benchmark, the systems were lightly loaded, providing headroom to grow.

Please see the white paper for information on PeopleSoft payroll best practices using flash.

See Also

Disclosure Statement

Oracle's PeopleSoft Financials 9.0 benchmark, Oracle's Sun SPARC Enterprise M4000 (4 2.66 SPARC64 VII+), Oracle's Sun SPARC Enterprise M5000 (8 2.66 SPARC64 VII+), 34.72 min. Results as of 12/02/2010, see www.oracle.com/apps_benchmark/html/white-papers-peoplesoft.html for more about PeopleSoft.

Monday Oct 11, 2010

Sun SPARC Enterprise M9000 Server Delivers World Record Non-Clustered TPC-H @3000GB Performance

Oracle's Sun SPARC Enterprise M9000 server delivered a single-system TPC-H 3000GB world record performance. The Sun SPARC Enterprise M9000 server, running Oracle Database 11g Release 2 on the Oracle Solaris operating system proves the power of Oracle's integrated solution.

  • Oracle beats IBM Power with better performance and price/performance (3 Year TCO). This shows that Oracle's focus on integrated system design provides more customer value than IBM's focus on "per core performance"!

  • The Sun SPARC Enterprise M9000 server is 27% faster than the IBM Power 595.

  • The Sun SPARC Enterprise M9000 server is 22% faster than the HP ProLiant DL980 G7.

  • The Sun SPARC Enterprise M9000 server is 26% lower than the IBM Power 595 for price/performance.

  • The Sun SPARC Enterprise M9000 server is 2.7 times faster than the IBM Power 595 for data loading.

  • The Sun SPARC Enterprise M9000 server is 2.3 times faster than the HP ProLiant DL980 for data loading.

  • The Sun SPARC Enterprise M9000 server is 2.6 times faster than the IBM p595 for Refresh Function.

  • The Sun SPARC Enterprise M9000 server is 3 times faster than the HP ProLiant DL980 for Refresh Function.

  • Oracle used Storage Redundancy Level 3 as defined by the TPC-H 2.12.0 specification, which is the highest level. IBM is the only other vendor to secure the storage to this level.

  • One should focus on the performance of the complete hardware and software stack since server implementation details such as the number of cores or the number of threads will obscure the important metrics of delivered system performance and system price/performance.

  • The Sun SPARC Enterprise M9000 server configured with SPARC VII processors, Sun Storage 6180 arrays, and running Oracle Solaris 10 operating system combined with Oracle Database 11g Release 2 achieved World Record TPC-H performance of 198,907.5 QphH@3000GB for non-clustered systems.

  • The Sun SPARC Enterprise M9000 server is over three times faster than the HP Itanium2 Superdome.

  • The Sun Storage 6180 array configuration (a total of 16 6180 arrays) in this benchmark delivered IO performance of over 21 GB/sec Sequential Read performance as measured by the vdbench tool.

  • This TPC-H result demonstrates that the Sun SPARC Enterprise M9000 server can handle the increasingly large databases required of DSS systems. The server delivered more than 18 GB/sec of real IO throughput as measured by the Oracle Database 11g Release 2 software.

  • Both Oracle and IBM had the same level of hardware discounting as allowed by TPC rules to provide a effective comparison of price/performance.

  • IBM has not shown any delivered I/O performance results for the high-end IBM POWER7 systems. In addition, they have not delivered any commercial benchmarks (TPC-C, TPC-H, etc.) which have heavy I/O demands.

Performance Landscape

TPC-H @3000GB, Non-Clustered Systems

System
CPU type
Memory
Composite
(QphH)
$/perf
($/QphH)
Power
(QppH)
Throughput
(QthH)
Database Available
Sun SPARC Enterprise M9000
2.88GHz SPARC64 VII
512GB
198,907.5 $15.27 182,350.7 216,967.7 Oracle 12/09/10
HP ProLiant DL980 G7
2.27GHz Intel Xeon X7560
512GB
162,601.7 $2.68 185,297.7 142,601.7 SQL Server 10/13/10
IBM Power 595
5.0GHz POWER6
512GB
156,537.3 $20.60 142,790.7 171,607.4 Sybase 11/24/09
Unisys ES7000 7600R
2.6GHz Intel Xeon
1024GB
102,778.2 $21.05 120,254.8 87,841.4 SQL Server 05/06/10
HP Integrity Superdome
1.6GHz Intel Itanium
256GB
60,359.3 $32.60 80,838.3 45,068.3 SQL Server 05/21/07

QphH = the Composite Metric (bigger is better)
$/QphH = the Price/Performance metric (smaller is better)
QppH = the Power Numerical Quantity
QthH = the Throughput Numerical Quantity

Complete benchmark results found at the TPC benchmark website http://www.tpc.org.

Configuration Summary and Results

Server:

Sun SPARC Enterprise M9000
32 x SPARC VII 2.88 GHz processors
512 GB memory
4 x internal SAS (4 x 300 GB)

External Storage:

16 x Sun Storage 6180 arrays (16x 16 x 300 GB)

Software:

Operating System: Oracle Solaris 10 10/09
Database: Oracle Database 11g Release 2 Enterprise Edition

Audited Results:

Database Size: 3000 GB (Scale Factor 3000)
TPC-H Composite: 198,907.5 QphH@3000GB
Price/performance: $15.27/QphH@3000GB
Available: 12/09/2010
Total 3 year Cost: $3,037,900
TPC-H Power: 182,350.7
TPC-H Throughput: 216,967.7
Database Load Time: 3:40:11

Benchmark Description

The TPC-H benchmark is a performance benchmark established by the Transaction Processing Council (TPC) to demonstrate Data Warehousing/Decision Support Systems (DSS). TPC-H measurements are produced for customers to evaluate the performance of various DSS systems. These queries and updates are executed against a standard database under controlled conditions. Performance projections and comparisons between different TPC-H Database sizes (100GB, 300GB, 1000GB, 3000GB and 10000GB) are not allowed by the TPC.

TPC-H is a data warehousing-oriented, non-industry-specific benchmark that consists of a large number of complex queries typical of decision support applications. It also includes some insert and delete activity that is intended to simulate loading and purging data from a warehouse. TPC-H measures the combined performance of a particular database manager on a specific computer system.

The main performance metric reported by TPC-H is called the TPC-H Composite Query-per-Hour Performance Metric (QphH@SF, where SF is the number of GB of raw data, referred to as the scale factor). QphH@SF is intended to summarize the ability of the system to process queries in both single and multi user modes. The benchmark requires reporting of price/performance, which is the ratio of QphH to total HW/SW cost plus 3 years maintenance.

Key Points and Best Practices

  • The Sun Storage 6180 array showed good scalability and these sixteen 6180 arrays showed over 21 GB/sec Sequential Read performance as measured by the vdbench tool.
  • Oracle Solaris 10 10/09 required little system tuning.
  • The optimal 6180 configuration for the benchmark was to set up 1 disk per volume instead of multiple disks per volume and let Oracle Solaris Volume Manager (SVM) mirror. Presenting as many volumes as possible to Oracle database gave the highest scan rate.

  • The storage was managed by SVM with 1MB stripe size to match with Oracle's database IO size. The default 16K stripe size is just too small for this DSS benchmark.

  • All the Oracle files, except TEMP tablespace, were mirrored under SVM. Eight 6180 arrays (128 disks) were mirrored to another 8 6180 arrays using 128-way stripe. IO performance was good and balanced across all the disks with a round robin order. Read performance was the same with mirror or without mirror. With the SVM mirror the benchmark passed the ACID (Atomicity, Consistency, Isolation and Durablity) test.

  • Oracle tables were 128-way partitioned and parallel degree for each table was set to 128 because the system had 128 cores. This setting worked the best for performance.

  • CPU usage during the Power run was not so high. This is because parallel degree was set to 128 for the tables and indexes so it utilized 128 vcpus for the most of the queries but the system had 256 vcpus.

See Also

Disclosure Statement

Sun SPARC Enterprise M9000 198,907.5 QphH@3000GB, $15.27/QphH@3000GB, avail 12/09/10, IBM Power 595 QphH@3000GB, 156,537.3 QphH@3000GB, $20.60/QphH@3000GB, avail 11/24/09, HP Integrity Superdome 60,359.3 QphH@3000GB, $32.60/QphH@3000GB avail 06/18/07, TPC-H, QphH, $/QphH tm of Transaction Processing Performance Council (TPC). More info www.tpc.org.

Friday Sep 24, 2010

SPARC T3-2 sets World Record on SPECjvm2008 Benchmark

World Record SPECjvm2008 Result

Oracle's SPARC T3-2 server equipped with two SPARC T3 processors running at 1.65 GHz delivered a World Record result of 320.52 SPECjvm2008 Peak ops/m on the SPECjvm2008 benchmark.
  • The SPARC T3-2 server demonstrates better performance than servers equipped with 2 Intel Xeon processors.

  • This result was produced using Oracle Solaris 10 and Oracle JDK 6 Update 21 Performance Release.

  • There are no SPECjvm2008 results published by IBM on POWER7 based systems.

  • The SPARC T3-2 server demonstrates Oracle's position of leadership in Java-based computing by publishing world record results for the SPECjvm2008 benchmark.

Performance Landscape

Complete benchmark results are at the SPECjvm2008 website.

SPECjvm2008 Performance Chart
(ordered by performance)
System Processors Performance
base peak
SPARC T3-2 2 x 1.65 GHz SPARC T3 - 320.52
Sun Blade X6270 2 x 2.93 GHz Intel X5570 317.13 -
Sun Fire X4450 4 x 2.66 GHz Intel X7450 283.79 -
Sun Fire X4450 4 x 2.93 GHz Intel X7350 260.08 -

base: SPECjvm2008 Base ops/m (bigger is better)
peak: SPECjvm2008 Peak ops/m (bigger is better)

Results and Configuration Summary

Hardware Configuration:

SPARC T3-2 server
2 x 1.65 GHz SPARC T3 processors
256 GB memory

Software Configuration:

Oracle Solaris 10 9/10
Java Platform, Standard Edition, JDK 6 Update 21 Performance Release

Benchmark Description

SPECjvm2008 (Java Virtual Machine Benchmark) is a benchmark suite for measuring the performance of a Java Runtime Environment (JRE), containing several real life applications and benchmarks focusing on core java functionality. The suite focuses on the performance of the JRE executing a single application; it reflects the performance of the hardware processor and memory subsystem, but has low dependence on file I/O and includes no network I/O across machines.

The SPECjvm2008 workload mimics a variety of common general purpose application computations. These characteristics reflect the intent that this benchmark will be applicable to measuring basic Java performance on a wide variety of both client and server systems.

SPEC also finds user experience of Java important, and the suite, therefore, includes startup benchmarks and has a required run category called base, which must be run without any tuning of the JVM to improve the out of the box performance.

SPECjvm2008 benchmark highlights:

  • Leverages real life applications (like derby, sunflow, and javac) and area-focused benchmarks (like xml, serialization, crypto, and scimark).
  • Also measures the performance of the operating system and hardware in the context of executing the JRE.

Key Points and Best Practices

  • Enhancements to the JVM had a major impact on performance

See Also

Disclosure Statement

SPEC and SPECjvm are registered trademarks of Standard Performance Evaluation Corporation. Results from www.spec.org and this report as of 9/16/2010. SPARC T3-2 320.52 SPECjvm2008 Peak ops/m, Sun Blade X6270 317.13 SPECjvm2008 Base ops/m.

Thursday Sep 23, 2010

SPARC T3-1 Performance on PeopleSoft Enterprise Financials 9.0 Benchmark

Oracle's SPARC T3-1 and Sun SPARC Enterprise M5000 servers combined with Oracle's Sun Storage F5100 Flash Array storage has produced the first world-wide disclosure and World Record performance on the PeopleSoft Enterprise Financials 9.0 benchmark.

  • Using SPARC T3-1 and Sun SPARC Enterprise M5000 servers along with a Sun Storage F5100 Flash Array system, the Oracle solution processed online business transactions to support 1000 concurrent users using 32 application server threads with compliant response times while simultaneously completing complex batch jobs. This is the first publication of this benchmark by any vendor world-wide.

  • The Sun Storage F5100 Flash Array system is a high performance, high-density solid-state flash array which provides a read latency of only 0.5 msec which is about 10 times faster than the normal disk latencies of 5 msec measured on this benchmark.

  • The SPARC T3-1 and Sun SPARC Enterprise M5000 servers were able process online users and concurrent batch jobs simultaneously in 38.66 minutes on this benchmark that reflects complex, multi-tier environment and utilizes a large back-end database of nearly 1 TB.

  • Both the SPARC T3-1 and Sun SPARC Enterprise M5000 servers used the Oracle Solaris 10 operating system.

  • The combination of Oracle's PeopleSoft Enterprise Financials/SCM 9.00.00.331, PeopleSoft Enterprise (PeopleTools) 8.49.23 and Oracle WebLogic server running on the SPARC T3-1 server and the Oracle database 11g Release 1 was run on the Sun SPARC Enterprise M5000 server for this benchmark.

Performance Landscape

As the first world-wide disclosure of this benchmark, no competitive results exist with which the current result may be compared.

Batch Processing Times
Batch Process Elapsed Time in Minutes
Batch Alone\* Batch with
1000 Online Users\*
JGEN Subsystem 7.30 7.78
JEDIT1 2.52 3.77
ALLOCATION 6.05 10.15
ALLOC EDIT/POST 2.32 2.23
SUM LEDGER 1.00 1.18
CONSOLIDATION 1.50 1.55
Total Main Batch Stream 20.69 26.66
SQR/GL_LEDGER 8.92 9.12
SQR/GL_TBAL 3.33 3.35
SQR 11.83 12.00
nVisions 8.78 8.83
nVision 11.83 12.00
Max SQR and nVision Stream 11.83 12.00
Total Batch (sum of Main Batch and Max SQR) 32.52 38.66

\* PeopleSoft Enterprise Financials batch processing and post-processing elapsed times.

Results and Configuration Summary

Hardware Configuration:

1 x SPARC T3-1 (1 x T3 at 1.65 GHz with 128 GB of memory)
1 x Sun SPARC Enterprise M5000 (8 x SPARC64 at 2.53 GHz with 64 GB of memory)
1 x Sun Storage F5100 Flash Array (74 x 24 GB FMODs)
2 x StorageTek 2540 (12 x 146 GB SAS 15K RPM)
1 x StorageTek 2501 (12 x 146 GB SAS 15K RPM)
1 x Dual-Port SAS Fibre Channel Host Bus Adapters (HBA)

Software Configuration:

Oracle Solaris 10 10/09
Oracle's PeopleSoft Enterprise Financials/SCM 9.00.00.311 64-bit
Oracle's PeopleSoft Enterprise (PeopleTools) 8.49.23 64-bit
Oracle 11g R2 11.1.0.6 64-bit
Oracle Tuxedo 9.1 RP36 with Jolt 9.1
Micro Focus COBOL Server Express 4.0 SP4 64-bit

Benchmark Description

The PeopleSoft Enterprise Financials batch processes included in this benchmark are as follows:

  • Journal Generator: (AE) This process creates journals from accounting entries (AE) generated from various data sources, including non-PeopleSoft systems as well as PeopleSoft applications. In the benchmark, the Journal Generator (FS_JGEN) process is set up to create accounting entries from Oracle's PeopleSoft applications in the same database, such as PeopleSoft Enterprise Payables, Receivables, Asset Management, Expenses, Cash Management. The process is run with the option of Edit and Post turned on to edit and post the journals created by Journal generator. Journal Edit is an AE program and Post is a COBOL program.

  • Allocation: (AE) This process allocates balances held or accumulated in one or more entities to more than one business unit, department or other entities based on user-defined rules.

  • Journal Edit & Post: (AE & COBOL) Journal Edit validates journal transactions before posting them to the ledger. This validation ensures that journals are valid, for example: valid ChartFields values and combinations, debits and credits equal, and inter/intra-unit balanced, Journal Post process posts only valid, edited journals, ensures each journal line posts to the appropriate target detail ledgers, and then changes the journal's status to posted. In this benchmark, the Journal Edit & Post is also set up to edit and post Oracle's PeopleSoft applications from another database, such as PeopleSoft Enterprise Payroll data.

  • Summary Ledger: (AE) Summary Ledger processing summarizes detail ledger data across selected GL BUs. Summary Ledgers can be generated for reporting purposes or used in consolidations.

  • Consolidations: (COBOL) Consolidation processing summarizes ledger balances and generates elimination journal entries across business units based on user-defined rules.

  • SQR & nVision Reporting: Reporting will consist of nVision and SQR reports. A balance sheet, and income statement, and a trial balance will be generated for each GL BU by SQR processes GLS7002 and GLS7012. The consolidated results of the nVision reports are run by 10 nVision users using 4 standard delivered report request definitions such as BALANCE, INCOME, CONSBAL, and DEPTINC. Each of the nVision users will have ownership over 10 Business Units and each of the nVision users will submit multiple runs that are being executed in parallel to generate a total of 40 nVision reports.

Batch processes are run concurrently with more than 1000 emulated users executing 30 pre-defined online applications. Response times for the online applications are collected and must conform to a maximum time.

Key Points and Best Practices

Oracle's SPARC T3-1 and Oracle's Sun SPARC Enterprise M5000 servers published the first result for Oracle's PeopleSoft Enterprise Financials 9.0 benchmark for concurrent batch and 1000 online users using the large database model on Oracle 11g running Oracle Solaris 10.

The SPARC T3-1 and Sun SPARC Enterprise M5000 servers were able process online users and concurrent batch jobs simultaneously in 38.66 minutes.

The Sun Storage F5100 Flash Array system, which is highly tuned for IOPS, contributed to the result through reduced IO latency.

The combination of the SPARC T3-1 and Sun SPARC Enterprise M5000 servers, with a Sun Storage F5100 Flash Array system, form an ideal environment for hosting complex multi-tier applications. This is the first public disclosure of any system running this benchmark.

The SPARC T3-1 server hosted the web and application server tiers, providing good response time to emulated user requests. The benchmark specification allows 1000 users, but there is headroom for increased load.

The Sun SPARC Enterprise M5000 server was used for the database server along with a Sun Storage F5100 Flash Array system. The speed of the M-series server with the low latency of the Flash Array provided the overall low latency for user requests, even while completing complex batch jobs.

The parallelism of the SPARC T3-1 server, when used as an application and web server tier, is best taken advantage of by configuring sufficient server processes. With sufficient server processes distributed across the hardware cores, acceptable user response times are achieved.

The low latency of the Sun Storage F5100 Flash Array storage contributed to the excellent response times of emulated users by making data quickly available to the database back-end. The array was configured as several RAID 0 volumes and data was distributed across the volumes, maximizing storage bandwidth.

The transaction processing capacity of the Sun SPARC Enterprise M5000 server enabled very fast batch processing times while supporting over 1000 online users.

While running the maximum workload specified by the benchmark, the systems were lightly loaded, providing headroom to grow.

Please see the white paper for information on PeopleSoft payroll best practices using flash.

See Also

Disclosure Statement

Oracle's PeopleSoft Financials 9.0 benchmark, Oracle's SPARC T3-1 (1 1.65GHz SPARC-T3), Oracle's SPARC Enterprise M5000 (8 2.53GHz SPARC64), 38.66 min. www.oracle.com/apps_benchmark/html/white-papers-peoplesoft.html Results 09/20/2010.

Tuesday Jun 29, 2010

Sun Fire X2270 M2 Sets World Record on SPEC OMP2001 Benchmark

Oracle's Sun Fire X2270 M2 server running the Oracle Solaris 10 10/09 with the Oracle Solaris Studio 12 Update 1 compiler, produced the top x86 SPECompM2001 result for all 2-socket servers.

  • The Sun Fire X2270 M2 server with two Intel Xeon X5670 processors running 24 OpenMP threads achieved a SPEC OMP2001 result of 55,178 SPECompM2001.

  • The Sun Fire X2270 M2 server beat the Cisco B200 M2 system even thought the Cisco system used the faster Intel Xeon X5680 (3.33GHz) chips.

Performance Landscape

SPEC OMP2001 Performance Charts - bigger is better, selected results, please see www.spec.org for complete results. All results as of 06/28/10.

In the tables below
"Base" = SPECompMbase2001 and "Peak" = SPECompMpeak2001

SPEC OMPM2001 results

System Processors Base
Threads
Performance Results
Cores/
Chips
Type GHz Peak Base
Sun Fire X2270 M2 12/2 Xeon X5670 2.93 24 55178 49548
Cisco B200 M2 12/2 Xeon X5680 3.33 24 55072 52314
Intel SR1600UR 12/2 Xeon X5680 3.33 24 54249 51510
Intel SR1600UR 12/2 Xeon X5670 2.93 24 53313 50283

Results and Configuration Summary

Hardware Configuration:

Sun Fire X2270 M2
2 x 2.93 GHz Intel Xeon X5670
24 GB

Software Configuration:

Oracle Solaris 10 10/09
Oracle Solaris Studio 12 Update 1
SPEC OMP2001 suite v3.2

Benchmark Description

The SPEC OMPM2001 Benchmark Suite was released in June 2001 and tests HPC performance using OpenMP for parallelism.

  • 11 programs (3 in C and 8 in Fortran) parallelized using OpenMP API
Goals of the suite:
  • Targeted to mid-range (4-32 processor) parallel systems
  • Run rules, tools and reporting similar to SPEC CPU2006
  • Programs representative of HPC and Scientific Applications

The SPEC OMPL2001 Benchmark Suite was released in June 2001 and tests HPC performance using OpenMP for parallelism.

  • 9 programs (2 in C and 7 in Fortran) parallelized using OpenMP API
Goals of the suite:
  • Targeted to larger parallel systems
  • Run rules, tools and reporting similar to SPEC CPU2006
  • Programs representative of HPC and Scientific Applications

There are "base" variants of both the above metrics that require more conservative compilation, such as using the same flags for all benchmarks.

See Also

Disclosure Statement

SPEC, SPEComp reg tm of Standard Performance Evaluation Corporation. Results from www.spec.org as of 28 June 2010 and this report. Sun Fire X2270 M2 (2 chips, 12 cores, 24 OpenMP threads) 55,178 SPECompM2001;

Sun Fire X4170 M2 Sets World Record on SPEC CPU2006 Benchmark

Oracle's Sun Fire X4170 M2 server equipped with two Intel Xeon X5670 2.93 GHz processors and running the Oracle Solaris 10 operating system delivered the a world record score of 53.5 SPECfp_base2006.

  • The Sun Fire X4170 M2 server using the Oracle Solaris Studio Express 06/10 compiler delivered a world record result of 53.5 SPECfp_base2006.

  • The Sun Fire X4170 M2 server delivered 20% better performance on the SPECfp_base2006 benchmark compared to the IBM 780 POWER7 based system.

  • The Sun Fire X4170 M2 server beat systems from Supermicro (X8DTU-LN4F+), Dell (R710), IBM (x3650 M3) and Bull (R460 F2) on SPECfp_base2006.

Performance Landscape

SPEC CPU2006 Performance Charts - bigger is better, selected results, please see www.spec.org for complete results. All results as of 06/28/10.

In the tables below
"Base" = SPECint_base2006, SPECfp_base2006, SPECint_rate_base2006 or SPECfp_rate_base2006
"Peak" = SPECint2006, SPECfp2006, SPECint_rate2006 or SPECfp_rate2006

SPECfp2006 results

System Processors Performance Results
Cores/
Chips
Type GHz Peak Base
Sun Fire X4170 M2 12/2 Xeon X5670 2.93 57.6 53.5
Sun Fire X2270 M2 12/2 Xeon X5670 2.93 58.6 49.9
Supermicro X8DTU-LN4F+ 8/2 Xeon X5677 3.46 48.8 45.9
IBM x3650 M3 8/2 Xeon X5677 3.46 48.9 45.8
Bull R460 F2 8/2 Xeon X5677 3.46 49.3 45.8
Dell R710 8/2 Xeon X5677 3.46 49.3 45.8
Dell R710 12/2 Xeon X5680 3.33 48.5 45.0
IBM 780 16/2 POWER7 3.94 71.5 44.5
Dell R710 12/2 Xeon X5670 2.93 45.8 42.5

SPECint_rate2006 results

System Processors Base
Copies
Performance Results
Cores/
Chips
Type GHz Peak Base
Dell R815 24/2 Opteron 6176 2.3 24 401 314
Fijitsu BX922 S2 12/2 Xeon X5680 3.33 24 381 354
Dell R710 12/2 Xeon X5680 3.33 24 379 355
Sun Blade X6270 M2 12/2 Xeon X5680 3.33 24 369 337
Sun Fire X4170 M2 12/2 Xeon X5670 2.93 24 353 316
Sun Fire X2270 M2 (S10) 12/2 Xeon X5670 2.93 24 346 311
Sun Fire X2270 M2 (OEL) 12/2 Xeon X5670 2.93 24 342 320

SPECfp_rate2006 results

System Processors Base
Copies
Performance Results
Cores/
Chips
Type GHz Peak Base
Dell R815 24/2 Opteron 6176 2.3 24 323 295
Dell R710 12/2 Xeon X5680 3.33 24 256 248
Fijitsu BX922 S2 12/2 Xeon X5680 3.33 24 256 248
Sun Blade X6270 M2 12/2 Xeon X5680 3.33 24 255 247
Sun Fire X4170 M2 12/2 Xeon X5670 2.93 24 245 234
Sun Fire X2270 M2 (S10) 12/2 Xeon X5670 2.93 24 240 231
Sun Fire X2270 M2 (OEL) 12/2 Xeon X5670 2.93 24 235 226

Results and Configuration Summary

Hardware Configuration:

Sun Fire X4170
2 x 2.93 GHz Intel Xeon X5670
48 GB
Sun Fire X2270
2 x 2.93 GHz Intel Xeon X5670
48 GB
Sun Blade X6270
2 x 3.33 GHz Intel Xeon X5680
48 GB

Software Configuration:

Oracle Solaris 10 10/09
Oracle Solaris Studio Express 6/10
SPEC CPU2006 suite v1.1
MicroQuill SmartHeap Library v8.1

Benchmark Description

SPEC CPU2006 is SPEC's most popular benchmark, with over 8000 results published in the three years since it was introduced. It measures:

  • "Speed" - single copy performance of chip, memory, compiler
  • "Rate" - multiple copy (throughput)

The rate metrics are used for the throughput-oriented systems described on this page. These metrics include:

  • SPECint_rate2006: throughput for 12 integer benchmarks derived from real applications such as perl, gcc, XML processing, and pathfinding
  • SPECfp_rate2006: throughput for 17 floating point benchmarks derived from real applications, including chemistry, physics, genetics, and weather.

There are base variants of both the above metrics that require more conservative compilation. In particular, all benchmarks of a particular programming language must use the same compilation flags.

See Also

Disclosure Statement

SPEC, SPECint, SPECfp reg tm of Standard Performance Evaluation Corporation. Results from www.spec.org as of 24 June 2010 and this report. Sun Fire X4170 M2 53.5 SPECfp_base2006.

Monday Jun 28, 2010

Sun Fire X4470 Sets World Records on SPEC OMP2001 Benchmarks

Oracle's Sun Fire X4470 server, with four Intel Xeon X7560 processors capable of running OpenMP applications with 64 compute threads, delivered outstanding performance on the both medium and large suites of the industry-standard SPEC OMP2001 benchmark.

  • The Sun Fire X4470 server running the Oracle Solaris 10 10/09 operating system with Oracle Solaris Studio 12 Update 1 compiler software, produced the top x86 result on SPECompM2001.

  • The Sun Fire X4470 server running the Oracle Solaris 10 10/09 operating system with Oracle Solaris Studio 12 Update 1 compiler software, produced the top x86 result on SPECompL2001.

  • The Sun Fire X4470 server beats IBM Power 750 Express POWER7 3.55 GHz SPECompM2001 score by 14%, while the Sun Fire X4470 server uses half the number of OpenMP threads compared to the IBM Power 750.
  • The Sun Fire X4470 server with four Intel Xeon 7560 processors, running 64 OpenMP threads, achieved SPEC OMP2001 results of 118,264 SPECompM2001 and 642,479 SPECompL2001.

  • The Sun Fire X4470 server produced better SPECompL2001 results than Cisco (UCS C460 M1) and Intel (QSSC-S4R) even though they all used the same number of Intel Xeon X7560 processors.

  • The Sun Fire X4470 server produced better SPECompM2001 results than Cisco (UCS C460 M1), SGI (Altix UV 10) and Intel (QSSC-S4R) even though they all used the same number of Intel Xeon X7560 processors.

Performance Landscape

SPEC OMP2001 Performance Charts - bigger is better, selected results, please see www.spec.org for complete results. All results as of 06/28/10.

In the tables below
"Base" = SPECompLbase2001 or SPECompMbase2001
"Peak" = SPECompLpeak2001 or SPECompMpeak2001

SPEC OMPL2001 results

System Processors Base
Threads
Performance Results
Cores/
Chips
Type GHz Peak Base
Sun Fire X4470 32/4 Xeon 7560 2.26 64 642479 615790
Cisco UCS C460 M1 32/4 Xeon 7560 2.26 64 628126 607818
Intel QSSC-S4R X7560 32/4 Xeon 7560 2.26 64 610386 591375
Sun/Fujitsu SPARC M8000 64/16 SPARC64 VII 2.52 64 581807 532576

SPEC OMPM2001 results

System Processors Base
Threads
Performance Results
Cores/
Chips
Type GHz Peak Base
Sun Fire X4470 32/4 Xeon 7560 2.26 64 118264 95650
Cisco UCS C460 M1 32/4 Xeon 7560 2.26 64 109077 100258
SGI Altix UV 10 32/4 Xeon 7560 2.26 64 107248 96797
Intel QSSC-S4R X7560 32/4 Xeon 7560 2.26 64 106369 98288
Sun/Fujitsu SPARC M8000 64/16 SPARC64 VII 2.52 64 104714 75418
IBM Power 750 Express 32/4 POWER7 3.55 128 104175 92957

Results and Configuration Summary

Hardware Configuration:

Sun Fire X4470
4 x 2.26 GHz Intel Xeon X7560
256 GB

Software Configuration:

Oracle Solaris 10 10/09
Oracle Solaris Studio Update 1
SPEC OMP2001 suite v3.2

Benchmark Description

The SPEC OMPM2001 Benchmark Suite was released in June 2001 and tests HPC performance using OpenMP for parallelism.

  • 11 programs (3 in C and 8 in Fortran) parallelized using OpenMP API
Goals of the suite:
  • Targeted to mid-range (4-32 processor) parallel systems
  • Run rules, tools and reporting similar to SPEC CPU2006
  • Programs representative of HPC and Scientific Applications

The SPEC OMPL2001 Benchmark Suite was released in June 2001 and tests HPC performance using OpenMP for parallelism.

  • 9 programs (2 in C and 7 in Fortran) parallelized using OpenMP API
Goals of the suite:
  • Targeted to larger parallel systems
  • Run rules, tools and reporting similar to SPEC CPU2006
  • Programs representative of HPC and Scientific Applications

There are "base" variants of both the above metrics that require more conservative compilation, such as using the same flags for all benchmarks.

See Also

Disclosure Statement

SPEC, SPEComp reg tm of Standard Performance Evaluation Corporation. Results from www.spec.org as of 28 June 2010 and this report. Sun Fire X4470 (4 chips, 32 cores, 64 OpenMP threads) 642,479 SPECompL2001, 118264 SPECompM2001.

Sun Fire X4470 Sets World Record on SPEC CPU2006 Rate Benchmark

Oracle's Sun Fire X4470 server delivered a world record SPECint_rate2006 result for all x86 systems with 4 chips.

  • The Sun Fire X4470 server with four Intel Xeon X7560 processors achieved a SPECint_rate2006 score of 788 and a SPECfp_rate2006 score of 573

  • The Sun Fire X4470 server delivered better 4 socket x86 system performance on the SPECint_rate2006 benchmark compared to HP (DL585 G7), Cicso (UCS C460 M1), Dell (R815) and IBM (x3850 X5).

  • The Sun Fire X4470 server delivered better performance on the SPECfp_rate2006 benchmark compared to similar Intel Xeon X7560 processor based systems from Cisco (UCS C460 M1), IBM (x3850 X5), and Fujitsu (RX600 S5).

Performance Landscape

SPEC CPU2006 Performance Charts - bigger is better, selected results, please see www.spec.org for complete results. All results as of 06/28/10.

In the tables below
"Base" = SPECint_rate_base2006 or SPECfp_rate_base2006
"Peak" = SPECint_rate2006 or SPECfp_rate2006

SPECint_rate2006 results

System Processors Base
Copies
Performance Results
Cores/
Chips
Type GHz Peak Base
Sun Fire X4470 32/4 Xeon X7560 2.26 64 788 724
HP DL585 G7 48/4 Opteron 6176 2.3 48 782 610
Cisco UCS C460 M1 32/4 Xeon X7560 2.26 64 772 723
Dell R815 48/4 Opteron 6174 2.20 48 771 602
IBM x3850 X5 32/4 Xeon X7560 2.26 64 770 720
Sun Fire X4640 48/8 Opteron 8435 2.6 48 730 574

SPECfp_rate2006 results

System Processors Base
Copies
Performance Results
Cores/
Chips
Type GHz Peak Base
Dell R815 48/4 Opteron 6174 2.20 48 626 574
HP DL585 G7 48/4 Opteron 6176 2.3 48 619 572
Sun Fire X4470 32/4 Xeon X7560 2.26 64 573 547
Cicso UCS C460 M1 32/4 Xeon X7560 2.26 64 568 549
IBM x3850 X5 32/4 Xeon X7560 2.26 64 560 543
Fujitsu RX600 S5 32/4 Xeon X7560 2.26 64 559 538
Sun Fire X4640 48/8 Opteron 8435 2.6 48 470 434

Results and Configuration Summary

Hardware Configuration:

Sun Fire X4470
4 x 2.26 GHz Intel Xeon X7560
256 GB

Software Configuration:

Oracle Solaris 10 10/09
Oracle Solaris Studio Express 6/10
SPEC CPU2006 suite v1.1
MicroQuill SmartHeap Library v8.1

Benchmark Description

SPEC CPU2006 is SPEC's most popular benchmark, with over 8000 results published in the three years since it was introduced. It measures:

  • "Speed" - single copy performance of chip, memory, compiler
  • "Rate" - multiple copy (throughput)

The rate metrics are used for the throughput-oriented systems described on this page. These metrics include:

  • SPECint_rate2006: throughput for 12 integer benchmarks derived from real applications such as perl, gcc, XML processing, and pathfinding
  • SPECfp_rate2006: throughput for 17 floating point benchmarks derived from real applications, including chemistry, physics, genetics, and weather.

There are base variants of both the above metrics that require more conservative compilation. In particular, all benchmarks of a particular programming language must use the same compilation flags.

See Also

Disclosure Statement

SPEC, SPECint, SPECfp reg tm of Standard Performance Evaluation Corporation. Results from www.spec.org as of 24 June 2010 and this report. Sun Fire X4470 788 SPECint_rate2006.

Sun Fire X4800 Sets World Records on SPEC CPU2006 Rate Benchmarks

Oracle's Sun Fire X4800 server combines the compute capacity of up to eight Intel Xeon X7560 processors with Oracle Solaris 10 and Oracle Solaris Studio Express software in a compact 5U form factor to provide better performance than any other x86 8-chip server, setting world records for both integer and floating point throughput on the SPEC CPU2006 benchmark for 8-chip x86 systems.
  • The Sun Fire X4800 server, in combination with Oracle Solaris 10 and Oracle Solaris Studio Express 6/10 compiler, delivered results of 1380 SPECint_rate2006 and 936 SPECfp_rate2006.

  • The combination of the Sun Fire X4800, Oracle Solaris and Oracle Solaris Studio compiler software resulted in a better score than Fujitsu's PRIMEQUEST 1800E server, running Red Hat Linux and using Intel compiler software.

Performance Landscape

SPEC CPU2006 Performance Charts - bigger is better, selected results, please see www.spec.org for complete results. All results as of 06/28/10.

In the tables below
"Base" = SPECint_rate_base2006 or SPECfp_rate_base2006
"Peak" = SPECint_rate2006 or SPECfp_rate2006

SPECint_rate2006 results

System Processors Base
Copies
Performance Results Comments
Cores/
Chips
Type GHz Peak Base
Sun Fire X4800 64/8 Xeon X7560 2.26 128 1380 1260 Solaris 10
Fujitsu 1800E 64/8 Xeon X7560 2.26 128 1339 1254 RHEL5.4 Linux
HP DL785 G6 48/8 Opteron 8439 2.8 48 800 629
HP DL585 G7 48/4 Opteron 6176 2.3 48 782 610
Sun Fire X4640 48/8 Opteron 8435 2.6 48 730 574
HP BL890c i2 32/8 Itanium 9350 1.73 32 531 507

SPECfp_rate2006 results

System Processors Base
Copies
Performance Results Comments
Cores/
Chips
Type GHz Peak Base
Sun Fire X4800 64/8 Xeon X7560 2.26 128 936 888 Solaris 10
Fujitsu 1800E 64/8 Xeon X7560 2.26 128 904 887 RHEL5.4 Linux
HP DL585 G7 48/4 Opteron 6176 2.3 48 619 572
HP BL890c i2 32/8 Itanium 9350 1.73 32 520 508
HP DL785 G6 48/8 Opteron 8439 2.8 48 513 473
Sun Fire X4640 48/8 Opteron 8435 2.6 48 470 434

Results and Configuration Summary

Hardware Configuration:

Sun Fire X4800
8 x 2.26 GHz Intel Xeon X7560
256 GB

Software Configuration:

Oracle Solaris 10 10/09
Oracle Solaris Studio Express 6/10
SPEC CPU2006 suite v1.1
MicroQuill SmartHeap Library v8.1

Benchmark Description

SPEC CPU2006 is SPEC's most popular benchmark, with over 8000 results published in the three years since it was introduced. It measures:

  • "Speed" - single copy performance of chip, memory, compiler
  • "Rate" - multiple copy (throughput)

The rate metrics are used for the throughput-oriented systems described on this page. These metrics include:

  • SPECint_rate2006: throughput for 12 integer benchmarks derived from real applications such as perl, gcc, XML processing, and pathfinding
  • SPECfp_rate2006: throughput for 17 floating point benchmarks derived from real applications, including chemistry, physics, genetics, and weather.

There are base variants of both the above metrics that require more conservative compilation. In particular, all benchmarks of a particular programming language must use the same compilation flags.

See Also

Disclosure Statement

SPEC, SPECint, SPECfp reg tm of Standard Performance Evaluation Corporation. Results from www.spec.org as of 24 June 2010 and this report. Sun Fire X4800 1380 SPECint_rate2006, 936 SPECfp_rate2006.

Wednesday Jun 09, 2010

PeopleSoft Payroll 500K Employees on Sun SPARC Enterprise M5000 World Record

Oracle's Sun SPARC Enterprise M5000 server combined with Oracle's Sun Storage F5100 Flash Array system has produced World Record Performance on PeopleSoft Payroll 9.0 (North American) 500K employees benchmark.
  • The Sun SPARC Enterprise M5000 server and the Sun Storage F5100 Flash Array system processed payroll for 500K employees using 32 payroll threads 18% faster than the IBM z10 EC 2097-709 mainframe as measured for payroll processing tasks in the Peoplesoft Payroll 9.0 (North American) benchmark. This IBM mainframe is rated at 6,512 MIPS.

  • The IBM z10 mainframe with nine 4.4 GHz Gen1 processors has a list price over $6M.

  • The Sun SPARC Enterprise M5000 server together with the Sun Storage F5100 Flash Array system processed payroll for 500K employees using 32 payroll threads 92% faster than an HP rx7640 as measured for payroll processing tasks in the Peoplesoft Payroll 9.0 (North American) benchmark.

  • The Sun Storage F5100 Flash Array system is a high performance, high density solid state flash array which provides a read latency of only 0.5 msec which is about 10 times faster than the normal disk latencies 5 msec measured on this benchmark.

  • The Sun SPARC Enterprise M5000 server used the Oracle Solaris 10 operating system and ran with the Oracle 11gR1 database for this benchmark.

Performance Landscape

500K Employees

System Processor OS/Database Time in Minutes Num of
Streams
Payroll
Processing
Result
Run 1 Run 2 Run 3
Sun M5000 8x 2.53GHz SPARC64 VII Solaris/Oracle 11g 50.11 73.88 534.20 1267.06 32
IBM z10 9x 4.4GHz Gen1, 6,512 MIPS Z/OS /DB2 58.96 80.5 250.68 462.6 8
HP rx7640 8x 1.6GHz Itanium2 HP-UX/Oracle 11g 96.17 133.63 712.72 1665.01 32

Times under all Run columns above represent Payroll processing and Post-processing elapsed times and furthermore:

  • Run 1 = 32 parallel job streams & Single Check option = "No"
  • Run 2 = 32 sequential jobs for Pay Calculation process & 32 parallel job streams for the rest. Single Check option = "Yes"
  • Run 3 = One job stream & Single Check option = "Yes"

Times under Result column represents Payroll processing only.

Results and Configuration Summary

Hardware Configuration:

    1 x Sun SPARC Enterprise M5000 (8 x 2.53 GHz/64 GB)
    1 x Sun Storage F5100 Flash Array (40 x 24 GB FMODs)
    1 x StorageTek 2510 (4 x 136 GB SAS 15K RPM)
    4 x Dual-Port SAS Fibre Channel Host Bus Adapters (HBA)

Software Configuration:

    Oracle Solaris 10 10/09
    Oracle PeopleSoft HCM and Campus Solutions 9.00.00.311 64-bit
    Oracle PeopleSoft Enterprise (PeopleTools) 8.49.25 64-bit
    Oracle 11g R1 11.1.0.7 64-bit
    Micro Focus COBOL Server Express 4.0 SP4 64-bit

Benchmark Description

The PeopleSoft 9.0 Payroll (North America) benchmark is a performance benchmark established by PeopleSoft to demonstrate system performance for a range of processing volumes in a specific configuration. This information may be used to determine the software, hardware, and network configurations necessary to support processing volumes. This workload represents large batch runs typical of OLTP workloads during a mass update.

To measure five application business process run times for a database representing large organization. The five processes are:

  • Paysheet Creation: generates payroll data worksheet for employees, consisting of std payroll information for each employee for given pay cycle.

  • Payroll Calculation: Looks at Paysheets and calculates checks for those employees.

  • Payroll Confirmation: Takes information generated by Payroll Calculation and updates the employees' balances with the calculated amounts.

  • Print Advice forms: The process takes the information generated by payroll Calculations and Confirmation and produces an Advice for each employee to report Earnings, Taxes, Deduction, etc.

  • Create Direct Deposit File: The process takes information generated by above processes and produces an electronic transmittal file use to transfer payroll funds directly into an employee bank a/c.

For the benchmark, we collect at least three data points with different number of job streams (parallel jobs). This batch benchmark allows a maximum of thirty-two job streams to be configured to run in parallel.

Key Points and Best Practices

Please see the white paper for information on PeopleSoft payroll best practices using flash.

See Also

Disclosure Statement

Oracle PeopleSoft Payroll 9.0 benchmark, Sun SPARC Enterprise M5000 (8 2.53GHz SPARC64 VII) 50.11 min, IBM z10 (9 gen1) 58.96 min, HP rx7640 (8 1.6GHz Itanium2) 96.17 min, www.oracle.com/apps_benchmark/html/white-papers-peoplesoft.html, results 6/3/2010.

Thursday Jun 03, 2010

Sun SPARC Enterprise T5440 World Record SPECjAppServer2004

Using five of Oracle's Sun SPARC Enterprise T5440 systems for application serving along with one of Oracle's Sun SPARC Enterprise M9000 servers for the database server allowed Oracle to deliver a world record result of 28,648.74 SPECjAppServer2004 JOPS@Standard on the SPECjAppServer2004 benchmark.

This result was run using the Oracle WebLogic 10.3.3 Application Server, a component of Oracle Fusion Middleware, and Oracle Database 11g Enterprise Edition with the Oracle Solaris 10 operating system to obtain this world record result.

Oracle Performance Advantages
  • This Oracle result is 26% faster than the IBM result of 22,634.13 SPECjAppServer2004 JOPS@Standard. For the application tier of the benchmark, Oracle used five Sun SPARC Enterprise T5440 servers compared to the sixteen IBM BladeCenter HS blades used by IBM. For the database tier, Oracle used a Sun SPARC Enterprise M9000 server compared to a IBM system p5 595 used by IBM. 

  • The Oracle result is faster than the HP result of 28,463.03 SPECjAppServer2004 JOPS@Standard. For the application tier of the benchmark, Oracle used five Sun SPARC Enterprise T5440 servers compared to the seventeen HP BL870c blade servers used by HP. For the database tier, Oracle used a Sun SPARC Enterprise M9000 server compared to a HP Superdome used by HP. 

Oracle's Advantages in Reduced Space and Reduced Number of Servers
  • The five Sun SPARC Enterprise T5440 servers used a total of 20 RU of space to obtain this result which is 26% less than the 27 RU space used by the three blade chassis containing sixteen IBM BladeCenter HS blades.

  • IBM uses more than 3.4 times the number of application servers than Oracle.

  • The five Sun SPARC Enterprise T5440 servers occupied 40% of the 50 RU space used by the five blade chassis containing 17 HP BL870c blade servers to obtain this leading result. 

  • HP uses more than 3.2 times the number of application servers than Oracle.

Oracle's Storage Advantages:
  • The six Sun Storage F5100 Flash Array storage used in this result occupied 6U of rack space which is 13% of the 44U space used by the database storage in the IBM result. 

  • The database storage in the HP result used 4x EVA81000 Storage arrays consuming112U of space is more than 18 times the 6U space used for database storage in the Oracle result.

  • The application server storage in the HP result used an EVA6100 storage space which consumed 16U of space for JMS logs. The 5x T5440 each used internal SSDs for the same function - no additional external storage was used.

Oracle Technologies Utilized:
  • Six of Oracle's Sun Storage F5100 Flash Array storage were used with Oracle 11g Enterprise Edition on the Sun SPARC Enterprise M9000 server to show outstanding database performance in this benchmark. 

  • These results were obtained using Java Platform, Standard Edition JDK 6 Update 20 on the Sun SPARC Enterprise T5440 servers and running the Oracle Solaris 10 10/09 operating system.

  • The five Sun SPARC Enterprise T5440 servers used Oracle Solaris Containers to consolidate eight Oracle Weblogic application server instances on each server to achieve this result. 

  • Oracle Fusion Middleware provides a family of complete, integrated, hot pluggable and best-of-breed products known for enabling enterprise customers to create and run agile and intelligent business applications. Oracle WebLogic Servers on-going, record-setting Java application server performance demonstrates why so many customers rely on Oracle Fusion Middleware as their foundation for innovation.

Oracle has other benchmarks that show that Oracle's "Optimized System Performance" is more important than IBM's "Per-core Performance Focus".

Performance Landscape

SPECjAppServer2004 Performance Chart as of 6/2/2010. Complete benchmark results may be found at the SPEC benchmark website http://www.spec.org. SPECjAppServer2004 JOPS@Standard (bigger is better)

Submitter SPECjAppServer2004
JOPS@Standard
J2EE Server DB Server
Oracle 28,648.74 5x Sun SPARC Enterprise T5440
1.6 GHz US-T2 Plus
Oracle WebLogic 10.3.3
1x Sun SPARC Enterprise M9000
2.88 GHz SPARC64-VII
Oracle 11g DB 11.1.0.7
HP 28,463.03 17x HP BL870c Server Blade
1.6 Ghz Itanium
Oracle WebLogic 10.3
1x HP Superdome
1.6 GHz Itanium
Oracle 11g DB 11.1.0.7
IBM 22,634.13 16x IBM BladeCenter HS21
3.32 GHz Intel X5470
WebSphere Application Server V7.0.0.1
1x IBM System p5 595
2.1 GHz POWER5+
IBM DB2 Universal Database 9.5 FP3

Results and Configuration Summary

Application Server:
    5x Sun SPARC Enterprise T5440
      4 x 1.6 GHz UltraSPARC T2 Plus
      256 GB memory
      2 x 10GbE NIC
      2 x 32GB SATA SSD
      Oracle Solaris 10 10/09
      Oracle Solaris Containers
      Oracle WebLogic 10.3.3 Application Server - Standard Edition
      Oracle Fusion Middleware
      Java Platform, Standard Edition JDK 6 Update 20

Database Server:

    Sun SPARC Enterprise M9000
      64x 2.88 GHz SPARC64-VII
      2048 GB memory
      6 x Sun Storage F5100 Flash Array
      Oracle Solaris 10 10/09
      Oracle Database Enterprise Edition Release 11.1.0.7

Benchmark Description

SPECjAppServer2004 (Java Application Server) is a multi-tier benchmark for measuring the performance of Java 2 Enterprise Edition (J2EE) technology-based application servers. SPECjAppServer2004 is an end-to-end application which exercises all major J2EE technologies implemented by compliant application servers as follows:
  • The web container, including servlets and JSPs
  • The EJB container
  • EJB2.0 Container Managed Persistence
  • JMS and Message Driven Beans
  • Transaction management
  • Database connectivity
Moreover, SPECjAppServer2004 also heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network. The primary metric of the SPECjAppServer2004 benchmark is jAppServer Operations Per Second (JOPS) which is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is NO price/performance metric in this benchmark.

Key Points and Best Practices

  • 8x Oracle WebLogic server instances on each Sun SPARC Enterprise T5440 server were hosted in 4x separate Solaris Containers to demonstrate consolidation of multiple application servers.
  • The Oracle WebLogic application servers were executed in the FX scheduling class to improve performance by reducing the frequency of context switches.
  • Enhancements in Java to the JVM had a major impact on performance.
  • Each Sun SPARC Enterprise T5440 used 2x 10GbE NICs for network traffic from the driver systems.

See Also

Disclosure Statement

SPECjAppServer2004, 5x Sun SPARC Enterprise T5440 (4 chips, 32 cores) 28648.74 SPECjAppServer2004 JOPS@Standard; 17x HP BL870c (4 chips, 8 cores) 28463.03 SPECjAppServer2004 JOPS@Standard; 16x IBM HS21 (2 chips, 8 cores) 22634.13 SPECjAppServer2004 JOPS@Standard; SPEC, SPECjAppServer reg tm of Standard Performance Evaluation Corporation. Results from www.spec.org as of 6/2/2010.

HP C7000 Blade Chassis (10 RU each). 5x Blade Chassis total 50 RU.
HP EVA8100 2C6D Storage Array(112 disks): 2x HSV210-B controllers (2U each) and 8x M5314C Disk Enclosures (3U each) total 28 RU. 4x EVA8100 2C6D total 112 RU.
HP EVA6100 2C4D Storage Array: 2x HSV200-B controllers (2U each) and 4x M5314C Disk Enclosures (3U each) total 16 RU.
http://h20000.www2.hp.com/bc/docs/support/SupportManual/c00816246/c00816246.pdf
http://h18004.www1.hp.com/products/quickspecs/12745_div/12745_div.pdf

IBM BladeCenter H Chassis (9 RU each). 3x Chassis Total 27 RU.
IBM DS4800 Disk System Model 82 (4U each). 6x IBM DS4000 EXP810 (3U each) total 22 RU. 2x Total STorage DS4800 total 44 RU.
http://www-03.ibm.com/systems/xbc/cog/bc_h_8852/bc_h_8852aag.html
ftp://ftp.software.ibm.com/systems/support/system_x_pdf/59y7294.pdf
ftp://ftp.software.ibm.com/systems/support/bladecenter/gc26779809.pdf

About

BestPerf is the source of Oracle performance expertise. In this blog, Oracle's Strategic Applications Engineering group explores Oracle's performance results and shares best practices learned from working on Enterprise-wide Applications.

Index Pages
Search

Archives
« May 2016
SunMonTueWedThuFriSat
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
    
       
Today