Thursday Mar 31, 2016

SPECjEnterprise2010: Oracle Server X6-2 World Record 2-Chip Single Application Server

Two Oracle Server X6-2 systems, using the Intel Xeon E5-2699 v4 processor, produced a world record x86 two-chip single application server SPECjEnterprise2010 benchmark result of 27,509.59 SPECjEnterprise2010 EjOPS. One Oracle Server X6-2 system ran the application tier and the second Oracle Server X6-2 system ran the database tier.

  • The Oracle Server X6-2 system demonstrated 22% better performance when compared to the IBM Power System S824 server result of 22,543.34 SPECjEnterprise2010 EjOPS.

  • The Oracle Server X6-2 system demonstrated 43% better performance when compared to the IBM X3650 M5 server result of 19,282.14 SPECjEnterprise2010 EjOPS.

  • The Oracle Server X6-2 system demonstrated 28% better performance when compared to the previous generation Oracle Server X5-2 system result of 21,504.30 SPECjEnterprise2010 EjOPS.

  • This result used Oracle WebLogic Server 12c, Java HotSpot(TM) 64-Bit Server 1.8.0_74, Oracle Database 12c, and Oracle Linux.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results. The table below shows the top two- and four-chip single application server results.

SPECjEnterprise2010 Performance Chart
as of 3/31/2016
Submitter EjOPS* Application Server Database Server
Oracle 27,509.59 1 x Oracle Server X6-2
2 x 2.2 GHz Intel Xeon E5-2699 v4
Oracle WebLogic 12c (12.2.1)
1 x Oracle Server X6-2
2 x 2.2 GHz Intel Xeon E5-2699 v4
Oracle Database 12c (12.1.0.2)
IBM 22,543.34 1 x IBM Power System S824
4 x 3.5 GHz POWER8
WebSphere Application Server V8.5
1 x IBM Power System S824
4 x 3.5 GHz POWER8
IBM DB2 10.5 FP3
Oracle 21,504.30 1 x Oracle Server X5-2
2 x 2.3 GHz Intel Xeon E5-2699 v3
Oracle WebLogic 12c (12.1.3)
1 x Oracle Server X5-2
2 x 2.3 GHz Intel Xeon E5-2699 v3
Oracle Database 12c (12.1.0.2)
IBM 19,282.14 1 x IBM X3650 M5
2 x 2.6 GHz Intel Xeon E5-2697 v3
WebSphere Application Server V8.5
1 x IBM X3850 X6
4 x 2.8 GHz Intel Xeon E7-4890 v2
IBM DB2 10.5

* EjOPS – SPECjEnterprise2010 EjOPS, bigger is better.

Configuration Summary

Application Server:

1 x Oracle Server X6-2
2 x Intel Xeon Processor E5-2699 v4 (2.2 GHz)
256 GB memory
3 x 10 GbE NIC
Oracle Linux 6 Update 6 (kernel-2.6.39-400.243.1.el6uek.x86_64)
Oracle WebLogic Server 12c (12.2.1)
Java HotSpot(TM) 64-Bit Server VM on Linux, version 1.8.0_74 (Java SE 8 Update 74)

Database Server:

1 x Oracle Server X6-2
2 x Intel Xeon Processor E5-2699 v4 (2.2 GHz)
512 GB memory
2 x 10 GbE NIC
1 x 16 Gb FC HBA
2 x Oracle Server X5-2L Storage
Oracle Linux 6 Update 6 (kernel-2.6.39-400.243.1.el6uek.x86_64)
Oracle Database 12c Enterprise Edition Release 12.1.0.2

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The SPECjEnterprise2010 benchmark has been designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems.

The workload consists of an end to end web based order processing domain, an RMI and Web Services driven manufacturing domain and a supply chain model utilizing document based Web Services. The application is a collection of Java classes, Java Servlets, Java Server Pages, Enterprise Java Beans, Java Persistence Entities (pojo's) and Message Driven Beans.

The SPECjEnterprise2010 benchmark heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second ("SPECjEnterprise2010 EjOPS"). This metric is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is no price/performance metric in this benchmark.

Key Points and Best Practices

  • Four Oracle WebLogic server instances were started using numactl binding 2 instances per chip.
  • Four Oracle database listener processes were started, 2 processes bound per processor.
  • Additional tuning information is in the report at spec.org.
  • COD (Cluster on Die) is enabled in the BIOS on the application server.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Oracle Server X6-2, 27,509.59 SPECjEnterprise2010 EjOPS; Oracle Server X5-2, 21,504.30 SPECjEnterprise2010 EjOPS; IBM Power System S824, 22,543.34 SPECjEnterprise2010 EjOPS; IBM System x3650 M5, 19,282.14 SPECjEnterprise2010 EjOPS. Results from www.spec.org as of 3/31/2016.

Thursday Mar 17, 2016

OLTPbenchmark Workload, Open-Source Benchmark: SPARC T7-1 Performance Beats IBM S824, Beats x86 E5-2699 v3

OLTPbenchmark is an open-source database benchmarking tool that includes an On-Line Transaction Processing (OLTP) transactional workload derived from the industry standard TPC-C workload.

Oracle's SPARC T7-1 server demonstrated OLTP performance that is 2.76 times faster per chip than Intel Xeon Processor E5-2699 v3 and 5.47 times faster per chip than an IBM POWER8 (3.5 GHz) processor. This means that a SPARC T7-1 is 1.38 times faster than a 2-chip x86 E5 v3 based server. The SPARC T7-1 server is also 1.37 times faster than an IBM Power System S824 (POWER8) server. On per core performance, the SPARC M7 processor used in the SPARC T7-1 server out performed the IBM POWER8 processor. All of these tests used Oracle Database 12c Release 1 (12.1.0.2) Enterprise Edition for the database.

Comparing the SPARC T7-1 server to the 2-chip x86 E5 v3 server equipped with two 2.3 GHz Intel Xeon Processor E5-2699 v3, we see the following advantages for the the SPARC T7-1 server.

  • On a per chip basis, the SPARC T7-1 server demonstrated 2.76 times better performance compared to the 2-chip x86 E5 v3 server.

  • At the system level, the SPARC T7-1 server demonstrated 1.38 times better performance compared to the 2-chip x86 E5 v3 server.

Comparing the SPARC T7-1 server to an IBM Power System S824 server equipped with four 3.5 GHz POWER8 processors (6 cores), we see the following advantages for the the SPARC T7-1 server.

  • On a per chip basis, the SPARC T7-1 server demonstrated nearly 5.47 times better performance compared to an IBM Power System S824 server.

  • On a per core basis the SPARC T7-1 server demonstrated nearly 3% better performance per core compared to an IBM Power System S824 server.

  • At the system level, the SPARC T7-1 server demonstrated nearly 1.37 times better performance compared to the IBM Power System S824 server.

The OLTPbenchmark transactional workload is based upon the TPC-C benchmark specification. Details of the configuration and parameters used are available in the reports referenced in the See Also section.

Performance Landscape

All OLTPbenchmark server results were run as part of this benchmark effort (except as noted). All results are run with Oracle Database 12c Release 1 Enterprise Edition. Results are ordered by TPM/core, highest to lowest.

OLTPbenchmark Transactional Workload
Relative Performance to x86 System
System TPM TPM/chip TPM/core
SPARC T7-1
1 x SPARC M7 (32 cores/chip, 32 total)
1.38x 2.76x 1.55x
IBM Power System S824
4 x POWER8 (6 cores/chip, 24 total)
1.01x 0.50x 1.51x
Oracle Server X5-2
2 x Intel E5-2699 v3 (18 cores/chip, 36 total)
1.00x 1.00x 1.00x

TPM – OLTPbenchmark transactions per minute

Results on the IBM Power System S824 were run by Oracle engineers using Oracle Database 12c.

Configuration Summary

Systems Under Test:

SPARC T7-1 server with
1 x SPARC M7 processor (4.13 GHz)
512 GB memory
2 x 600 GB 10K RPM SAS2 HDD
1 x Sun Dual Port 10 GbE PCIe 2.0 Networking card with Intel 82599 10 GbE Controller
1 x Sun Storage 16 Gb Fibre Channel Universal HBA
Oracle Solaris 11.3
Oracle Database 12c Release 1 (12.1.0.2) Enterprise Edition
Oracle Grid Infrastructure 12c Release 1 (12.1.0.2)

Oracle Server X5-2 with
2 x Intel Xeon processor E5-2699 v3 (2.3 GHz)
512 GB memory
2 x 600 GB 10K RPM SAS2 HDD
1 x Sun Dual Port 10 GbE PCIe 2.0 Networking card with Intel 82599 10 GbE Controller
1 x Sun Storage 16 Gb Fibre Channel Universal HBA
Oracle Linux 6.5
Oracle Database 12c Release 1 (12.1.0.2) Enterprise Edition
Oracle Grid Infrastructure 12c Release 1 (12.1.0.2)

IBM Power System S824 with
4 x POWER8 (3.5 GHz)
512 GB memory
4 x 300 GB 15K RPM SAS HDD
1 x 10 GbE Network Interface
1 x 16 Gb Fibre Channel HBA
AIX 7.1 TL3 SP3
Oracle Database 12c Release 1 (12.1.0.2) Enterprise Edition
Oracle Grid Infrastructure 12c Release 1 (12.1.0.2)

Storage Servers:

1 x Oracle Server X5-2L with
2 x Intel Xeon Processor E5-2630 v3 (2.4 GHz)
32 GB memory
1 x Sun Storage 16 Gb Fibre Channel Universal HBA
4 x 1.6 TB NVMe SSD
2 x 600 GB SAS HDD
Oracle Solaris 11.3

1 x Oracle Server X5-2L with
2 x Intel Xeon Processor E5-2630 v3 (2.4 GHz)
32 GB memory
1 x Sun Storage 16 Gb Fibre Channel Universal HBA
14 x 600 GB SAS HDD
Oracle Solaris 11.3

Benchmark Description

The OLTPbenchmark workload as described from the OLTPbenchmark website:

This is a database performance testing tool that allows you to conduct database workload replay, industry-standard benchmark testing, and scalability testing under various loads, such as scaling a population of users who executes order-entry transactions against a wholesale supplier database.
OLTPbenchmark supports many databases including Oracle, SQL Server, DB2, TimesTen, MySQL, MariaDB, PostgreSQL, Greenplum, Postgres Plus Advanced Server, Redis and Trafodion SQL on Hadoop.

Key Points and Best Practices

  • For these tests, an 800 warehouse database was created to compare directly with results posted by Intel.

  • To improve the scalability, the OrderLine table was partitioned and loaded into a separate tablespace using the OLTPbenchmark GUI. The default blocksize was 8K and the OrderLine tablespace blocksize was 16K.

  • To reduce latency of Oracle "cache chains buffers" wait events, the OLTPbenchmark kit was modified by adding partitioning to the NEW_ORDER table as well as the ORDERS_I1 and ORDERS_I2 indexes.

  • To reduce latency of Oracle "library cache: mutex X" wait events, added recommended workarounds from the following Intel blog

  • Refer to the detailed configuration documents in the See Also section below for the list of Oracle parameters.

See Also

Disclosure Statement

Copyright 2016, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of March 17, 2016.

Monday Oct 26, 2015

Memory and Bisection Bandwidth: SPARC T7 and M7 Servers Faster Than x86 and POWER8

The STREAM benchmark measures delivered memory bandwidth on a variety of memory intensive tasks. Delivered memory bandwidth is key to a server delivering high performance on a wide variety of workloads. The STREAM benchmark is typically run where each chip in the system gets its memory requests satisfied from local memory. This report presents performance of Oracle's SPARC M7 processor based servers and compares their performance to x86 and IBM POWER8 servers.

Bisection bandwidth on a server is a measure of the cross-chip data bandwidth between the processors of a system where no memory access is local to the processor. Systems with large cross-chip penalties show dramatically lower bisection bandwidth. Real-world ad hoc workloads tend to perform better on systems with better bisection bandwidth because their memory usage characteristics tend to be chaotic.

IBM says the sustained or delivered bandwidth of the IBM POWER8 12-core chip is 230 GB/s. This number is a peak bandwidth calculation: 230.4 GB/sec = 9.6 GHz * 3 (r+w) * 8 byte. A similar calculation is used by IBM for the POWER8 dual-chip-module (two 6-core chips) to show a sustained or delivered bandwidth of 192 GB/sec (192.0 GB/sec = 8.0 GHz * 3 (r+w) * 8 byte). Peaks are the theoretical limits used for marketing hype, but true measured delivered bandwidth is the only useful comparison to help one understand delivered performance of real applications.

The STREAM benchmark is easy to run and anyone can measure memory bandwidth on a target system (see Key Points and Best Practices section).

  • The SPARC M7-8 server delivers over 1 TB/sec on the STREAM benchmark. This is over 2.4 times the triad bandwidth of an eight-chip x86 E7 v3 server.

  • The SPARC T7-4 delivered 2.2 times the STREAM triad bandwidth of a four-chip x86 E7 v3 server and 1.7 times the triad bandwidth of a four-chip IBM Power System S824 server.

  • The SPARC T7-2 delivered 2.5 times the STREAM triad bandwidth of a two-chip x86 E5 v3 server.

  • The SPARC M7-8 server delivered over 8.5 times the triad bisection bandwidth of an eight-chip x86 E7 v3 server.

  • The SPARC T7-4 server delivered over 2.7 times the triad bisection bandwidth of a four-chip x86 E7 v3 server and 2.3 times the triad bisection bandwidth of a four-chip IBM Power System S824 server.

  • The SPARC T7-2 server delivered over 2.7 times the triad bisection bandwidth of a two-chip x86 E5 v3 server.

Performance Landscape

The following SPARC, x86, and IBM S824 STREAM results were run as part of this benchmark effort. The IBM S822L result is from the referenced web location. The following SPARC results were all run using 32 GB dimms.

Maximum STREAM Benchmark Performance
System Chips Bandwidth (MB/sec - 10^6)
Copy Scale Add Triad
SPARC M7-8 8 995,402 995,727 1,092,742 1,086,305
x86 E7 v3 8 346,771 354,679 445,550 442,184
SPARC T7-4 4 512,080 510,387 556,184 555,374
IBM S824 4 251,533 253,216 322,399 319,561
IBM S822L 4 252,743 247,314 295,556 305,955
x86 E7 v3 4 230,027 232,092 248,761 251,161
SPARC T7-2 2 259,198 259,380 285,835 285,905
x86 E5 v3 2 105,622 105,808 113,116 112,521
SPARC T7-1 1 131,323 131,308 144,956 144,706

All of the following bisection bandwidth results were run as part of this benchmark effort.

Bisection Bandwidth Benchmark Performance (Nonlocal STREAM)
System Chips Bandwidth (MB/sec - 10^6)
Copy Scale Add Triad
SPARC M7-8 8 383,479 381,219 375,371 375,851
SPARC T5-8 8 172,195 172,354 250,620 250,858
x86 E7 v3 8 42,636 42,839 43,753 43,744
SPARC T7-4 4 142,549 142,548 142,645 142,729
SPARC T5-4 4 75,926 75,947 76,975 77,061
IBM S824 4 53,940 54,107 60,746 60,939
x86 E7 v3 4 41,636 47,740 51,206 51,333
SPARC T7-2 2 127,372 127,097 129,833 129,592
SPARC T5-2 2 91,530 91,597 91,761 91,984
x86 E5 v3 2 45,211 45,331 47,414 47,251

The following SPARC results were all run using 16 GB dimms.

SPARC T7 Servers – 16 GB DIMMS
Maximum STREAM Benchmark Performance
System Chips Bandwidth (MB/sec - 10^6)
Copy Scale Add Triad
SPARC T7-4 4 520,779 521,113 602,137 600,330
SPARC T7-2 2 262,586 262,760 302,758 302,085
SPARC T7-1 1 132,154 132,132 168,677 168,654

Configuration Summary

SPARC Configurations:

SPARC M7-8
8 x SPARC M7 processors (4.13 GHz)
4 TB memory (128 x 32 GB dimms)

SPARC T7-4
4 x SPARC M7 processors (4.13 GHz)
2 TB memory (64 x 32 GB dimms)
1 TB memory (64 x 16 GB dimms)

SPARC T7-2
2 x SPARC M7 processors (4.13 GHz)
1 TB memory (32 x 32 GB dimms)
512 GB memory (32 x 16 GB dimms)

SPARC T7-1
1 x SPARC M7 processor (4.13 GHz)
512 GB memory (16 x 32 GB dimms)
256 GB memory (16 x 16 GB dimms)

Oracle Solaris 11.3
Oracle Solaris Studio 12.4

x86 Configurations:

Oracle Server X5-8
8 x Intel Xeon Processor E7-8995 v3
2 TB memory (128 x 16 GB dimms)

Oracle Server X5-4
4 x Intel Xeon Processor E7-8995 v3
1 TB memory (64 x 16 GB dimms)

Oracle Server X5-2
2 x Intel Xeon Processor E5-2699 v3
256 GB memory (16 x 16 GB dimms)

Oracle Linux 7.1
Intel Parallel Studio XE Composer Version 2016 compilers

Benchmark Description

STREAM

The STREAM benchmark measures sustainable memory bandwidth (in MB/s) for simple vector compute kernels. All memory accesses are sequential, so a picture of how fast regular data may be moved through the system is portrayed. Properly run, the benchmark displays the characteristics of the memory system of the machine and not the advantages of running from the system's memory caches.

STREAM counts the bytes read plus the bytes written to memory. For the simple Copy kernel, this is exactly twice the number obtained from the bcopy convention. STREAM does this because three of the four kernels (Scale, Add and Triad) do arithmetic, so it makes sense to count both the data read into the CPU and the data written back from the CPU. The Copy kernel does no arithmetic, but, for consistency, counts bytes the same way as the other three.

The sequential nature of the memory references is the benchmark's biggest weakness. The benchmark does not expose limitations in a system's interconnect to move data from anywhere in the system to anywhere.

Bisection Bandwidth – Easy Modification of STREAM Benchmark

To test for bisection bandwidth, processes are bound to processors in sequential order. The memory is allocated in reverse order, so that the memory is placed non-local to the process. The benchmark is then run. If the system is capable of page migration, this feature must be turned off.

Key Points and Best Practices

The stream benchmark code was compiled for the SPARC M7 processor based systems with the following flags (using cc):

    -fast -m64 -W2,-Avector:aggressive -xautopar -xreduction -xpagesize=4m

The benchmark code was compiled for the x86 based systems with the following flags (Intel icc compiler):

    -O3 -m64 -xCORE-AVX2 -ipo -openmp -mcmodel=medium -fno-alias -nolib-inline

On Oracle Solaris, binding is accomplished with either setting the environment variable SUNW_MP_PROCBIND or the OpenMP variables OMP_PROC_BIND and OMP_PLACES.

    export OMP_NUM_THREADS=512
    export SUNW_MP_PROCBIND=0-511

On Oracle Linux systems using Intel compiler, binding is accomplished by setting the environment variable KMP_AFFINITY.

    export OMP_NUM_THREADS=72
    export KMP_AFFINITY='verbose,granularity=fine,proclist=[0-71],explicit'

The source code change in the file stream.c to do the reverse allocation

    <     for (j=STREAM_ARRAY_SIZE-1; j>=0; j--) { 
                a[j] = 1.0; 
                b[j] = 2.0; 
                c[j] = 0.0; 
            }
    ---
    >     for (j=0; j<STREAM_ARRAY_SIZE; j++) {
                a[j] = 1.0; 
                b[j] = 2.0; 
                c[j] = 0.0; 
            }
    

See Also

Disclosure Statement

Copyright 2015, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 10/25/2015.

Hadoop TeraSort: SPARC T7-4 Top Per-Chip Performance

Oracle's SPARC T7-4 server using virtualization delivered an outstanding single server result running the Hadoop TeraSort benchmark. The SPARC T7-4 server was run with and without security. Even the secure runs on the SPARC M7 processor based server performed much faster per chip compared to competitive unsecure results.

  • The SPARC T7-4 server on a per chip basis is 4.7x faster than an IBM POWER8 based cluster on the 10 TB Hadoop TeraSort benchmark.

  • The SPARC T7-4 server running with ZFS encryption enabled on the 10 TB Hadoop TeraSort benchmark is 4.6x faster than an unsecure x86 v2 cluster on a per chip basis.

  • The SPARC T7-4 server running with ZFS encryption (AES-256-GCM) enabled on the 10 TB Hadoop TeraSort benchmark is 4.3x faster than an unsecure (plain-text) IBM POWER8 cluster on a per chip basis.

  • The SPARC T7-4 server ran the 10 TB Hadoop TeraSort benchmark in 4,259 seconds.

Performance Landscape

The following table presents results for the 10 TB Hadoop TeraSort benchmark. The rate results are determined by taking the dataset size (10**13) and dividing by the time (in minutes). These rates are further normalized by the number of systems or chips used in obtaining the results.

10 TB Hadoop TeraSort Performance Landscape
System Security Nodes Total
Chips
Time
(sec)
Sort Rate (GB/min)
Per Node Per Chip
SPARC T7-4
SPARC M7 (4.13 GHz)
unsecure 1 4 4,259 140.9 35.2
SPARC T7-4
SPARC M7 (4.13 GHz)
AES-256-GCM 1 4 4,657 128.8 32.2
IBM Power System S822L
POWER8 (3.0 GHz)
unsecure 8 32 2,490 30.1 7.5
Dell R720xd/VMware
Intel Xeon E5-2680 v2 (2.8 GHz)
unsecure 32 64 1,054 17.8 8.9
Cisco UCS CPA C240 M3
Intel Xeon E5-2665 (2.4 GHz)
unsecure 16 32 3,112 12.0 6.0

Configuration Summary

Server:

SPARC T7-4
4 x SPARC M7 processors (4.13 GHz)
2 TB memory (64 x 32 GB)
6 x 600 GB 10K RPM SAS-2 HDD
10 GbE
Oracle Solaris 11.3 (11.3.0.29)
Oracle Solaris Studio 12.4
Java SE Runtime Environment (build 1.7.0_85-b33)
Hadoop 1.2.1

External Storage (Common Multiprotocol SCSI TARget, or COMSTAR enables system to be seen as a SCSI target device):

16 x Sun Server X3-2L
2 x Intel Xeon E5-2609 (2.4 GHz)
16 GB memory (2 x 8 GB)
2 x 600 GB SAS-2 HDD
12 x 3 TB SAS-1 HDD
4 x Sun Flash Accelerator F40 PCIe Card
Oracle Solaris 11.1 (11.1.16.5.0)
Please note: These devices are only used as storage. No Hadoop is run on these COMSTAR storage nodes. There was no compression or encryption done on these COMSTAR storage nodes.

Benchmark Description

The Hadoop TeraSort benchmark sorts 100-byte records by a contained 10-byte random key. Hadoop TeraSort is characterized by high I/O bandwidth between each compute/data node of a Hadoop cluster and the disk drives that are attached to that node.

Note: benchmark size is measured by power-of-ten not power-of-two bytes; 1 TB sort is sorting 10^12 Bytes = 10 billion 100-byte rows using an embedded 10-Byte key field of random characters, 100 GB sort is sorting 10^11 Bytes = 1 billion 100-byte rows, etc.

Key Points and Best Practices

  • The SPARC T7-4 server was configured with 15 Oracle Solaris Zones. Each Zone was running one Hadoop data-node with HDFS layered on an Oracle Solaris ZFS volume.

  • Hadoop uses a distributed, shared nothing, batch processing framework employing divide-conquer serial Map and Reduce JVM tasks with performance coming from scale-out concurrency (e.g. more tasks) rather than parallelism. Only one job scheduler and task manager can be configured per data/compute-node and both (job scheduler and task manager) have inherent scaling limitations (the hadoop design target being small compute-nodes and hundreds or even thousands of them).

  • Multiple data-nodes significantly help improve overall system utilization – HDFS becomes more distributed with more processes servicing file system operations, and more task-trackers are managing all the MapReduce work.

  • On large node systems virtualization is required to improve utilization by increasing the number of independent data/compute nodes each running their own hadoop processes.

  • I/O bandwidth to the local disk drives and network communication bandwidth are the primary determinants of Hadoop performance. Typically, Hadoop reads input data files from HDFS during the Map phase of computation, and stores intermediate file back to HDFS. Then during the subsequent Reduce phase of computation, Hadoop reads the intermediate files, and outputs the final result. The Map and Reduce phases are executed concurrently by multiple Map tasks and Reduce tasks. Tasks are purpose-built stand-alone serial applications often written in Java (but can be written in any programming language or script).

See Also

Disclosure Statement

Copyright 2015, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 25 October 2015.

Competitive results found at: Dell R720xd/VMware, IBM S822L, Cisco C240 M3

SPECjEnterprise2010: SPARC T7-1 World Record with Single Application Server Using 1 to 4 Chips

Oracle's SPARC T7-1 servers have set a world record for the SPECjEnterprise2010 benchmark for solutions using a single application server with one to four chips. The result of 25,818.85 SPECjEnterprise2010 EjOPS used two SPARC T7-1 servers, one server for the application tier and the other server for the database tier.

  • The SPARC T7-1 servers obtained a result of 25,093.06 SPECjEnterprise2010 EjOPS using encrypted data. This secured result used Oracle Advanced Security Transparent Data Encryption (TDE) for the application database tablespaces with the AES-256-CFB cipher. The network connection between the application server and the database server was also encrypted using the secure JDBC.

  • The SPARC T7-1 server solution delivered 34% more performance compared to the two-chip IBM x3650 M5 server result of 19,282.14 SPECjEnterprise2010 EjOPS.

  • The SPARC T7-1 server solution delivered 14% more performance compared to the four-chip IBM Power System S824 server result of 22,543.34 SPECjEnterprise2010 EjOPS.

  • The SPARC T7-1 server based results demonstrated 20% more performance compared to the Oracle Server X5-2 system result of 21,504.30 SPECjEnterprise2010 EjOPS. Oracle holds the top x86 two-chip application server SPECjEnterprise2010 result.

  • The application server used Oracle Fusion Middleware components including the Oracle WebLogic 12.1 application server and Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.8.0_60. The database server was configured with Oracle Database 12c Release 1.

  • For the secure result, the application data was encrypted in the Oracle database using the Oracle Advanced Security Transparent Data Encryption (TDE) feature. Hardware accelerated cryptography support in the SPARC M7 processor for the AES-256-CFB cipher was used to provide data security.

  • The benchmark performance using the secure SPARC T7-1 server configuration with encryption was less than 3% when compared to the peak result.

  • This result demonstrated less than 1 second average response times for all SPECjEnterprise2010 transactions and represents Jave EE 5.0 transactions generated by over 210,000 users.

Performance Landscape

Select single application server results. Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results.

SPECjEnterprise2010 Performance Chart
10/25/2015
Submitter EjOPS* Java EE Server DB Server Notes
Oracle 25,818.85 1 x SPARC T7-1
1 x 4.13 GHz SPARC M7
Oracle WebLogic 12c (12.1.3)
1 x SPARC T7-1
1 x 4.13 GHz SPARC M7
Oracle Database 12c (12.1.0.2)
-
Oracle 25,093.06 1 x SPARC T7-1
1 x 4.13 GHz SPARC M7
Oracle WebLogic 12c (12.1.3)
Network Data Encryption for JDBC
1 x SPARC T7-1
1 x 4.13 GHz SPARC M7
Oracle Database 12c (12.1.0.2)
Transparent Data Encryption
Secure
IBM 22,543.34 1 x IBM Power S824
4 x 3.5 GHz POWER 8
WebSphere Application Server V8.5
1 x IBM Power S824
4 x 3.5 GHz POWER 8
IBM DB2 10.5 FP3
-
Oracle 21,504.30 1 x Oracle Server X5-2
2 x 2.3 GHz Intel Xeon E5-2699 v3
Oracle WebLogic 12c (12.1.3)
1 x Oracle Server X5-2
2 x 2.3 GHz Intel Xeon E5-2699 v3
Oracle Database 12c (12.1.0.2)
COD
IBM 19,282.14 1 x System x3650 M5
2 x 2.6 GHz Intel Xeon E5-2697 v3
WebSphere Application Server V8.5
1 x System x3850 X6
4 x 2.8 GHz Intel Xeon E7-4890 v2
IBM DB2 10.5 FP5
-

* SPECjEnterprise2010 EjOPS (bigger is better)

The Cluster on Die (COD) mode is a BIOS setting that effectively splits the chip in half, making the operating system think it has twice as many chips as it does (in this case, four, 9 core chips). Intel has stated that COD is appropriate only for highly NUMA optimized workloads. Dell has shown that there is a 3.7x slower bandwidth to the other half of the chip split by COD.

Configuration Summary

Application Server:

1 x SPARC T7-1 server, with
1 x SPARC M7 processor (4.13 GHz)
256 GB memory (16 x 16 GB)
2 x 600 GB SAS HDD
2 x 400 GB SAS SSD
3 x Sun Dual Port 10 GbE PCIe 2.0 Networking card with Intel 82599 10 GbE Controller
Oracle Solaris 11.3 (11.3.0.0.30)
Oracle WebLogic Server 12c (12.1.3)
Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.8.0_60

Database Server:

1 x SPARC T7-1 server, with
1 x SPARC M7 processor (4.13 GHz)
512 GB memory (16 x 32 GB)
2 x 600 GB SAS HDD
1 x Sun Dual Port 10 GbE PCIe 2.0 Networking card with Intel 82599 10 GbE Controller
1 x Sun Storage 16 Gb Fibre Channel Universal HBA
Oracle Solaris 11.3 (11.3.0.0.30)
Oracle Database 12c (12.1.0.2)

Storage Servers:

1 x Oracle Server X5-2L (8-Drive), with
2 x Intel Xeon Processor E5-2699 v3 (2.3 GHz)
32 GB memory
1 x Sun Storage 16 Gb Fibre Channel Universal HBA
4 x 1.6 TB NVMe SSD
2 x 600 GB SAS HDD
Oracle Solaris 11.3 (11.3.0.0.30)
1 x Oracle Server X5-2L (24-Drive), with
2 x Intel Xeon Processor E5-2699 v3 (2.3 GHz)
32 GB memory
1 x Sun Storage 16 Gb Fibre Channel Universal HBA
14 x 600 GB SAS HDD
Oracle Solaris 11.3 (11.3.0.0.30)

1 x Brocade 6510 16 Gb FC switch

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The SPECjEnterprise2010 benchmark has been designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems,

  • The web zone, servlets, and web services
  • The EJB zone
  • JPA 1.0 Persistence Model
  • JMS and Message Driven Beans
  • Transaction management
  • Database connectivity
Moreover, SPECjEnterprise2010 also heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second (SPECjEnterprise2010 EjOPS). The primary metric for the SPECjEnterprise2010 benchmark is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is NO price/performance metric in this benchmark.

Key Points and Best Practices

  • Four Oracle WebLogic server instances on the SPARC T7-1 server were hosted in 4 separate Oracle Solaris Zones.
  • The Oracle WebLogic application servers were executed in the FX scheduling class to improve performance by reducing the frequency of context switches.
  • The Oracle log writer process was run in the RT scheduling class.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Results from www.spec.org as of 10/25/2015. SPARC T7-1, 25,818.85 SPECjEnterprise2010 EjOPS (unsecure); SPARC T7-1, 25,093.06 SPECjEnterprise2010 EjOPS (secure); Oracle Server X5-2, 21,504.30 SPECjEnterprise2010 EjOPS (unsecure); IBM Power S824, 22,543.34 SPECjEnterprise2010 EjOPS (unsecure); IBM x3650 M5, 19,282.14 SPECjEnterprise2010 EjOPS (unsecure);

Oracle E-Business Suite Applications R12.1.3 (OLTP X-Large): SPARC M7-8 World Record

Oracle's SPARC M7-8 server, using a four-chip Oracle VM Server for SPARC (LDom) virtualized server, produced a world record 20,000 users running the Oracle E-Business OLTP X-Large benchmark. The benchmark runs five Oracle E-Business online workloads concurrently: Customer Service, iProcurement, Order Management, Human Resources Self-Service, and Financials.

  • The virtualized four-chip LDom on the SPARC M7-8 was able to handle more users than the previous best result which used eight processors of Oracle's SPARC M6-32 server.

  • The SPARC M7-8 server using Oracle VM Server for SPARC provides enterprise applications high availability, where each application is executed on its own environment, insulated and independent of the others.

Performance Landscape

Oracle E-Business (3-tier) OLTP X-Large Benchmark
System Chips Total Online Users Weighted Average
Response Time (sec)
90th Percentile
Response Time (s)
SPARC M7-8 4 20,000 0.70 1.13
SPARC M6-32 8 18,500 0.61 1.16

Break down of the total number of users by component.

Users per Component
Component SPARC M7-8 SPARC M6-32
Total Online Users 20,000 users 18,500 users
HR Self-Service
Order-to-Cash
iProcurement
Customer Service
Financial
5000 users
2500 users
2700 users
7000 users
2800 users
4000 users
2300 users
2400 users
7000 users
2800 users

Configuration Summary

System Under Test:

SPARC M7-8 server
8 x SPARC M7 processors (4.13 GHz)
4 TB memory
2 x 600 GB SAS-2 HDD
using a Logical Domain with
4 x SPARC M7 processors (4.13 GHz)
2 TB memory
2 x Sun Storage Dual 16Gb Fibre Channel PCIe Universal HBA
2 x Sun Dual Port 10GBase-T Adapter
Oracle Solaris 11.3
Oracle E-Business Suite 12.1.3
Oracle Database 11g Release 2

Storage Configuration:

4 x Oracle ZFS Storage ZS3-2 appliances each with
2 x Read Flash Accelerator SSD
1 x Storage Drive Enclosure DE2-24P containing:
20 x 900 GB 10K RPM SAS-2 HDD
4 x Write Flash Accelerator SSD
1 x Sun Storage Dual 8Gb FC PCIe HBA
Used for Database files, Zones OS, EBS Mid-Tier Apps software stack
and db-tier Oracle Server
2 x Sun Server X4-2L server with
2 x Intel Xeon Processor E5-2650 v2
128 GB memory
1 x Sun Storage 6Gb SAS PCIe RAID HBA
4 x 400 GB SSD
14 x 600 GB HDD
Used for Redo log files, db backup storage.

Benchmark Description

The Oracle E-Business OLTP X-Large benchmark simulates thousands of online users executing transactions typical of an internal Enterprise Resource Processing, simultaneously executing five application modules: Customer Service, Human Resources Self Service, iProcurement, Order Management and Financial.

Each database tier uses a database instance of about 600 GB in size, supporting thousands of application users, accessing hundreds of objects (tables, indexes, SQL stored procedures, etc.).

Key Points and Best Practices

This test demonstrates virtualization technologies running concurrently various Oracle multi-tier business critical applications and databases on four SPARC M7 processors contained in a single SPARC M7-8 server supporting thousands of users executing a high volume of complex transactions with constrained (<1 sec) weighted average response time.

The Oracle E-Business LDom is further configured using Oracle Solaris Zones.

This result of 20,000 users was achieved by load balancing the Oracle E-Business Suite Applications 12.1.3 five online workloads across two Oracle Solaris processor sets and redirecting all network interrupts to a dedicated third processor set.

Each applications processor set (set-1 and set-2) was running concurrently two Oracle E-Business Suite Application servers and two database servers instances, each within its own Oracle Solaris Zone (4 x Zones per set).

Each application server network interface (to a client) was configured to map with the locality group associated to the CPUs processing the related workload, to guarantee memory locality of network structures and application servers hardware resources.

All external storage was connected with at least two paths to the host multipath-capable fibre channel controller ports and Oracle Solaris I/O multipathing feature was enabled.

See Also

Disclosure Statement

Oracle E-Business Suite R12 extra-large multiple-online module benchmark, SPARC M7-8, SPARC M7, 4.13 GHz, 4 chips, 128 cores, 1024 threads, 2 TB memory, 20,000 online users, average response time 0.70 sec, 90th percentile response time 1.13 sec, Oracle Solaris 11.3, Oracle Solaris Zones, Oracle VM Server for SPARC, Oracle E-Business Suite 12.1.3, Oracle Database 11g Release 2, Results as of 10/25/2015.

SAP Two-Tier Standard Sales and Distribution SD Benchmark: SPARC T7-2 World Record 2 Processors

Oracle's SPARC T7-2 server produces a world record result for 2-processors on the SAP two-tier Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement Package 5 for SAP ERP 6.0 (2 chips / 64 cores / 512 threads).

  • The SPARC T7-2 server achieved 30,800 SAP SD benchmark users running the two-tier SAP Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement Package 5 for SAP ERP 6.0.

  • The SPARC T7-2 server achieved 1.9 times more users than the Dell PowerEdge R730 server result.

  • The SPARC T7-2 server achieved 1.5 times more users than the IBM Power System S824 server result.

  • The SPARC T7-2 server achieved 1.9 times more users than the HP ProLiant DL380 Gen9 server result.

  • The SPARC T7-2 server result was run with Oracle Solaris 11 and used Oracle Database 12c.

Performance Landscape

SAP-SD 2-tier performance table in decreasing performance order for leading two-processor systems and four-processor IBM Power System S824 server, with SAP ERP 6.0 Enhancement Package 5 for SAP ERP 6.0 results (current version of the benchmark as of May, 2012).

SAP SD Two-Tier Benchmark
System
Processor
OS
Database
Users Resp Time
(sec)
Version Cert#
SPARC T7-2
2 x SPARC M7 (2x 32core)
Oracle Solaris 11
Oracle Database 12c
30,800 0.96 EHP5 2015050
IBM Power S824
4 x POWER8 (4x 6core)
AIX 7
DB2 10.5
21,212 0.98 EHP5 2014016
Dell PowerEdge R730
2 x Intel E5-2699 v3 (2x 18core)
Red Hat Enterprise Linux 7
SAP ASE 16
16,500 0.99 EHP5 2014033
HP ProLiant DL380 Gen9
2 x Intel E5-2699 v3 (2x 18core)
Red Hat Enterprise Linux 6.5
SAP ASE 16
16,101 0.99 EHP5 2014032

Version – Version of SAP, EHP5 refers to SAP ERP 6.0 Enhancement Package 5 for SAP ERP 6.0

Number of cores presented are per chip, to get system totals, multiple by the number of chips.

Complete benchmark results may be found at the SAP benchmark website http://www.sap.com/benchmark.

Configuration Summary and Results

Database/Application Server:

1 x SPARC T7-2 server with
2 x SPARC M7 processors (4.13 GHz, total of 2 processors / 64 cores / 512 threads)
1 TB memory
Oracle Solaris 11.3
Oracle Database 12c

Database Storage:
3 x Sun Server X3-2L each with
2 x Intel Xeon Processors E5-2609 (2.4 GHz)
16 GB memory
4 x Sun Flash Accelerator F40 PCIe Card
12 x 3 TB SAS disks
Oracle Solaris 11

REDO log Storage:
1 x Pillar FS-1 Flash Storage System, with
2 x FS1-2 Controller (Netra X3-2)
2 x FS1-2 Pilot (X4-2)
4 x DE2-24P Disk enclosure
96 x 300 GB 10000 RPM SAS Disk Drive Assembly

Certified Results (published by SAP)

Number of SAP SD benchmark users: 30,800
Average dialog response time: 0.96 seconds
Throughput:
  Fully processed order line items per hour: 3,372,000
  Dialog steps per hour: 10,116,000
  SAPS: 168,600
Average database request time (dialog/update): 0.022 sec / 0.047 sec
SAP Certification: 2015050

Benchmark Description

The SAP Standard Application SD (Sales and Distribution) Benchmark is an ERP business test that is indicative of full business workloads of complete order processing and invoice processing, and demonstrates the ability to run both the application and database software on a single system. The SAP Standard Application SD Benchmark represents the critical tasks performed in real-world ERP business environments.

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

Two-tier SAP Sales and Distribution (SD) standard application benchmarks, SAP Enhancement Package 5 for SAP ERP 6.0 as of 10/23/15:

SPARC T7-2 (2 processors, 64 cores, 512 threads) 30,800 SAP SD users, 2 x 4.13 GHz SPARC M7, 1 TB memory, Oracle Database 12c, Oracle Solaris 11, Cert# 2015050.
IBM Power System S824 (4 processors, 24 cores, 192 threads) 21,212 SAP SD users, 4 x 3.52 GHz POWER8, 512 GB memory, DB2 10.5, AIX 7, Cert#2014016
Dell PowerEdge R730 (2 processors, 36 cores, 72 threads) 16,500 SAP SD users, 2 x 2.3 GHz Intel Xeon Processor E5-2699 v3 256 GB memory, SAP ASE 16, RHEL 7, Cert#2014033
HP ProLiant DL380 Gen9 (2 processors, 36 cores, 72 threads) 16,101 SAP SD users, 2 x 2.3 GHz Intel Xeon Processor E5-2699 v3 256 GB memory, SAP ASE 16, RHEL 6.5, Cert#2014032

SAP, R/3, reg TM of SAP AG in Germany and other countries. More info www.sap.com/benchmark

SPARC T7-1 Delivers 1-Chip World Records for SPEC CPU2006 Rate Benchmarks

This page has been updated on November 19, 2015. The SPARC T7-1 server results have been published at www.spec.org.

Oracle's SPARC T7-1 server delivered world record SPEC CPU2006 rate benchmark results for systems with one chip. This was accomplished with Oracle Solaris 11.3 and Oracle Solaris Studio 12.4 software.

  • The SPARC T7-1 server achieved world record scores of 1200 SPECint_rate2006, 1120 SPECint_rate_base2006, 832 SPECfp_rate2006, and 801 SPECfp_rate_base2006.

  • The SPARC T7-1 server beat the 1 chip Fujitsu CELSIUS C740 with an Intel Xeon Processor E5-2699 v3 by 1.7x on the SPECint_rate2006 benchmark. The SPARC T7-1 server beat the 1 chip NEC Express5800/R120f-1M with an Intel Xeon Processor E5-2699 v3 by 1.8x on the SPECfp_rate2006 benchmark.

  • The SPARC T7-1 server beat the 1 chip IBM Power S812LC server with a POWER8 processor by 1.9 times on the SPECint_rate2006 benchmark and by 1.8 times on the SPECfp_rate2006 benchmark.

  • The SPARC T7-1 server beat the 1 chip Fujitsu SPARC M10-4S with a SPARC64 X+ processor by 2.2x on the SPECint_rate2006 benchmark and by 1.6x on the SPECfp_rate2006 benchmark.

  • The SPARC T7-1 server improved upon the previous generation SPARC platform which used the SPARC T5 processor by 2.5 on the SPECint_rate2006 benchmark and by 2.3 on the SPECfp_rate2006 benchmark.

The SPEC CPU2006 benchmarks are derived from the compute-intensive portions of real applications, stressing chip, memory hierarchy, and compilers. The benchmarks are not intended to stress other computer components such as networking, the operating system, or the I/O system. Note that there are many other SPEC benchmarks, including benchmarks that specifically focus on Java computing, enterprise computing, and network file systems.

Performance Landscape

Complete benchmark results are at the SPEC website. The tables below provide the new Oracle results, as well as select results from other vendors.

Presented are single chip SPEC CPU2006 rate results. Only the best results published at www.spec.org per chip type are presented (best Intel, IBM, Fujitsu, Oracle chips).

SPEC CPU2006 Rate Results – One Chip
System Chip Peak Base
  SPECint_rate2006
SPARC T7-1 SPARC M7 (4.13 GHz, 32 cores) 1200 1120
Fujitsu CELSIUS C740 Intel E5-2699 v3 (2.3 GHz, 18 cores) 715 693
IBM Power S812LC POWER8 (2.92 GHz, 10 cores) 642 482
Fujitsu SPARC M10-4S SPARC64 X+ (3.7 GHz, 16 cores) 546 479
SPARC T5-1B SPARC T5 (3.6 GHz, 16 cores) 489 441
IBM Power 710 Express POWER7 (3.55 GHz, 8 cores) 289 255
  SPECfp_rate2006
SPARC T7-1 SPARC M7 (4.13 GHz, 32 cores) 832 801
NEC Express5800/R120f-1M Intel E5-2699 v3 (2.3 GHz, 18 cores) 474 460
IBM Power S812LC POWER8 (2.92 GHz, 10 cores) 468 394
Fujitsu SPARC M10-4S SPARC64 X+ (3.7 GHz, 16 cores) 462 418
SPARC T5-1B SPARC T5 (3.6 GHz, 16 cores) 369 350
IBM Power 710 Express POWER7 (3.55 GHz, 8 cores) 248 229

The following table highlights the performance of the single-chip SPARC M7 processor based server to the best published two-chip POWER8 processor based server.

SPEC CPU2006 Rate Results
Comparing One SPARC M7 Chip to Two POWER8 Chips
System Chip Peak Base
  SPECint_rate2006
SPARC T7-1 1 x SPARC M7 (4.13 GHz, 32core) 1200 1120
IBM Power S822LC 2 x POWER8 (2.92 GHz, 2x 10core) 1100 853
  SPECfp_rate2006
SPARC T7-1 1 x SPARC M7 (4.13 GHz, 32 cores) 832 801
IBM Power S822LC 2 x POWER8 (2.92 GHz, 2x 10core) 888 745

Configuration Summary

System Under Test:

SPARC T7-1
1 x SPARC M7 processor (4.13 GHz)
512 GB memory (16 x 32 GB dimms)
800 GB on 4 x 400 GB SAS SSD (mirrored)
Oracle Solaris 11.3
Oracle Solaris Studio 12.4 with 4/15 Patch Set

Benchmark Description

SPEC CPU2006 is SPEC's most popular benchmark. It measures:

  • Speed — single copy performance of chip, memory, compiler
  • Rate — multiple copy (throughput)

The benchmark is also divided into integer intensive applications and floating point intensive applications:

  • integer: 12 benchmarks derived from applications such as artificial intelligence chess playing, artificial intelligence go playing, quantum computer simulation, perl, gcc, XML processing, and pathfinding
  • floating point: 17 benchmarks derived from applications, including chemistry, physics, genetics, and weather.

It is also divided depending upon the amount of optimization allowed:

  • base: optimization is consistent per compiled language, all benchmarks must be compiled with the same flags per language.
  • peak: specific compiler optimization is allowed per application.

The overall metrics for the benchmark which are commonly used are:

  • SPECint_rate2006, SPECint_rate_base2006: integer, rate
  • SPECfp_rate2006, SPECfp_rate_base2006: floating point, rate
  • SPECint2006, SPECint_base2006: integer, speed
  • SPECfp2006, SPECfp_base2006: floating point, speed

Key Points and Best Practices

  • Jobs were bound using pbind.

See Also

Disclosure Statement

SPEC and the benchmark names SPECfp and SPECint are registered trademarks of the Standard Performance Evaluation Corporation. Results as of November 19, 2015 from www.spec.org.
SPARC T7-1: 1200 SPECint_rate2006, 1120 SPECint_rate_base2006, 832 SPECfp_rate2006, 801 SPECfp_rate_base2006; SPARC T5-1B: 489 SPECint_rate2006, 440 SPECint_rate_base2006, 369 SPECfp_rate2006, 350 SPECfp_rate_base2006; Fujitsu SPARC M10-4S: 546 SPECint_rate2006, 479 SPECint_rate_base2006, 462 SPECfp_rate2006, 418 SPECfp_rate_base2006. IBM Power 710 Express: 289 SPECint_rate2006, 255 SPECint_rate_base2006, 248 SPECfp_rate2006, 229 SPECfp_rate_base2006; Fujitsu CELSIUS C740: 715 SPECint_rate2006, 693 SPECint_rate_base2006; NEC Express5800/R120f-1M: 474 SPECfp_rate2006, 460 SPECfp_rate_base2006; IBM Power S822LC: 1100 SPECint_rate2006, 853 SPECint_rate_base2006, 888 SPECfp_rate2006, 745 SPECfp_rate_base2006; IBM Power S812LC: 642 SPECint_rate2006, 482 SPECint_rate_base2006, 468 SPECfp_rate2006, 394 SPECfp_rate_base2006.

Friday Apr 03, 2015

Oracle Server X5-2 Produces World Record 2-Chip Single Application Server SPECjEnterprise2010 Result

Two Oracle Server X5-2 systems, using the Intel Xeon E5-2699 v3 processor, produced a World Record x86 two-chip single application server SPECjEnterprise2010 benchmark result of 21,504.30 SPECjEnterprise2010 EjOPS. One Oracle Server X5-2 ran the application tier and the second Oracle Server X5-2 was used for the database tier.

  • The Oracle Server X5-2 system demonstrated 11% better performance when compared to the IBM X3650 M5 server result of 19,282.14 SPECjEnterprise2010 EjOPS.

  • The Oracle Server X5-2 system demonstrated 1.9x better performance when compared to the previous generation Sun Server X4-2 server result of 11,259.88 SPECjEnterprise2010 EjOPS.

  • This result used Oracle WebLogic Server 12c, Java HotSpot(TM) 64-Bit Server 1.8.0_40 Oracle Database 12c, and Oracle Linux.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results. The table below shows the top single application server, two-chip x86 results.

SPECjEnterprise2010 Performance Chart
as of 4/1/2015
Submitter EjOPS* Application Server Database Server
Oracle 21,504.30 1x Oracle Server X5-2
2x 2.3 GHz Intel Xeon E5-2699 v3
Oracle WebLogic 12c (12.1.3)
1x Oracle Server X5-2
2x 2.3 GHz Intel Xeon E5-2699 v3
Oracle Database 12c (12.1.0.2)
IBM 19,282.14 1x IBM X3650 M5
2x 2.6 GHz Intel Xeon E5-2697 v3
WebSphere Application Server V8.5
1x IBM X3850 X6
4x 2.8 GHz Intel Xeon E7-4890 v2
IBM DB2 10.5
Oracle 11,259.88 1x Sun Server X4-2
2x 2.7 GHz Intel Xeon E5-2697 v2
Oracle WebLogic 12c (12.1.2)
1x Sun Server X4-2L
2x 2.7 GHz Intel Xeon E5-2697 v2
Oracle Database 12c (12.1.0.1)

* SPECjEnterprise2010 EjOPS, bigger is better.

Configuration Summary

Application Server:

1 x Oracle Server X5-2
2 x 2.3 GHz Intel Xeon E5-2699 v3 processors
256 GB memory
3 x 10 GbE NIC
Oracle Linux 6 Update 5 (kernel-2.6.39-400.243.1.el6uek.x86_64)
Oracle WebLogic Server 12c (12.1.3)
Java HotSpot(TM) 64-Bit Server VM on Linux, version 1.8.0_40 (Java SE 8 Update 40)
BIOS SW 1.2

Database Server:

1 x Oracle Server X5-2
2 x 2.3 GHz Intel Xeon E5-2699 v3 processors
512 GB memory
2 x 10 GbE NIC
1 x 16 Gb FC HBA
2 x Oracle Server X5-2L Storage
Oracle Linux 6 Update 5 (kernel-3.8.13-16.2.1.el6uek.x86_64)
Oracle Database 12c Enterprise Edition Release 12.1.0.2

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The SPECjEnterprise2010 benchmark has been designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems.

The workload consists of an end to end web based order processing domain, an RMI and Web Services driven manufacturing domain and a supply chain model utilizing document based Web Services. The application is a collection of Java classes, Java Servlets, Java Server Pages, Enterprise Java Beans, Java Persistence Entities (pojo's) and Message Driven Beans.

The SPECjEnterprise2010 benchmark heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second ("SPECjEnterprise2010 EjOPS"). This metric is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is no price/performance metric in this benchmark.

Key Points and Best Practices

  • Four Oracle WebLogic server instances were started using numactl binding 2 instances per chip.
  • Four Oracle database listener processes were started, 2 processes bound per processor.
  • Additional tuning information is in the report at http://spec.org.
  • COD (Cluster on Die) is enabled in the BIOS on the application server.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Oracle Server X5-2, 21,504.30 SPECjEnterprise2010 EjOPS; IBM System X3650 M5, 19,282.14 SPECjEnterprise2010 EjOPS. Sun Server X4-2, 11,259.88 SPECjEnterprise2010 EjOPS; Results from www.spec.org as of 4/1/2015.

Friday Mar 20, 2015

Oracle ZFS Storage ZS4-4 Shows 1.8x Generational Performance Improvement on SPC-2 Benchmark

The Oracle ZFS Storage ZS4-4 appliance delivered 1.8x improved performance and 1.3x improved price performance over the previous generation Oracle ZFS Storage ZS3-4 appliance as shown by the SPC-2 benchmark.

  • Running the SPC-2 benchmark, the Oracle ZFS Storage ZS4-4 appliance delivered SPC-2 Price-Performance of $17.09 and an overall score of 31,486.23 SPC-2 MBPS.

  • The Oracle ZFS Storage continues its strong price performance by occupying the three of the top five SPC-2 price performance.

  • Oracle holds the three of the top four performance results on the SPC-2 benchmark for HDD based systems.

  • The Oracle ZFS Storage ZS4-4 appliance has a 7.6x price-performance advantage over the IBM DS8870 and 2x performance advantage as measured by the SPC-2 benchmark.

  • The Oracle ZFS Storage ZS4-4 appliance has a 5.0x performance advantage over the new Fujitsu DX200 S3 as measured by the SPC-2 benchmark.

  • The Oracle ZFS Storage ZS4-4 appliance has a 4.6x price-performance advantage over the Fujitsu ET8700 S2 and 1.9x performance advantage as shown by the SPC-2 benchmark.

  • The Oracle ZFS Storage ZS4-4 appliance has a 4.6x price-performance advantage over the Hitachi Virtual Storage Platform (VSP) and 1.96x performance advantage as measured by the SPC-2 benchmark.

  • The Oracle ZFS Storage ZS4-4 appliance has a 1.6x price-performance advantage over the HP XP7 disk array as shown by the SPC-2 benchmark (HP even discounted their hardware 63%).

Performance Landscape

SPC-2 Price-Performance

Below is a table of the top SPC-2 Price-Performance results for HDD storage based systems, presented in increasing price-performance order (as of 03/17/2015). The complete set of results may be found at SPC2 top 10 Price-Performance list.

System SPC-2
MBPS
$/SPC-2
MBPS
Results
Identifier
Oracle ZFS Storage ZS3-2 16,212.66 $12.08 BE00002
Fujitsu Eternus DX200 S3 6,266.50 $15.42 B00071
SGI InfiniteStorage 5600 8,855.70 $15.97 B00065
Oracle ZFS Storage ZS4-4 31,486.23 $17.09 B00072
Oracle ZFS Storage ZS3-4 17,244.22 $22.53 B00067
NEC Storage M700 14,408.89 $25.10 B00066
Sun StorageTek 2530 663.51 $26.48 B00026
HP XP7 storage 43,012.53 $28.30 B00070
Fujitsu ETERNUS DX80 S2 2,685.50 $28.48 B00055
SGI InfiniteStorage 5500-SP 4,064.49 $28.57 B00059
Hitachi Unified Storage VM 11,274.83 $32.64 B00069

SPC-2 MBPS = the Performance Metric
$/SPC-2 MBPS = the Price-Performance Metric
Results Identifier = A unique identification of the result

SPC-2 Performance

The following table list the top SPC-2 -Performance results for HDD storage based systems, presented in decreasing performance order (as of 03/17/2015). The complete set of results may be found at the SPC2 top 10 Performance list.

HDD Based Systems SPC-2
MBPS
$/SPC-2
MBPS
TSC Price Results
Identifier
HP XP7 storage 43,012.52 $28.30 $1,217,462 B00070
Oracle ZFS Storage ZS4-4 31,486.23 $17.09 $538,050 B00072
Oracle ZFS Storage ZS3-4 17,244.22 $22.53 $388,472 B00067
Oracle ZFS Storage ZS3-2 16,212.66 $12.08 $195,915 BE00002
Fujitsu ETERNUS DX8870 S2 16,038.74 $79.51 $1,275,163 B00063
IBM System Storage DS8870 15,423.66 $131.21 $2,023,742 B00062
IBM SAN VC v6.4 14,581.03 $129.14 $1,883,037 B00061
Hitachi Virtual Storage Platform (VSP) 13,147.87 $95.38 $1,254,093 B00060
HP StorageWorks P9500 XP Storage Array 13,147.87 $88.34 $1,161,504 B00056

SPC-2 MBPS = the Performance Metric
$/SPC-2 MBPS = the Price-Performance Metric
TSC Price = Total Cost of Ownership Metric
Results Identifier = A unique identification of the result Metric

Complete SPC-2 benchmark results may be found at
http://www.storageperformance.org/results/benchmark_results_spc2.

Configuration Summary

Storage Configuration:

Oracle ZFS Storage ZS4-4 storage system in clustered configuration
2 x Oracle ZFS Storage ZS4-4 controllers with
8 x Intel Xeon processors
3 TB memory
24 x Oracle Storage Drive Enclosure DE2-24P, each with
24 x 300 GB 10K RPM SAS-2 drives

Benchmark Description

SPC Benchmark 2 (SPC-2): Consists of three distinct workloads designed to demonstrate the performance of a storage subsystem during the execution of business critical applications that require the large-scale, sequential movement of data. Those applications are characterized predominately by large I/Os organized into one or more concurrent sequential patterns. A description of each of the three SPC-2 workloads is listed below as well as examples of applications characterized by each workload.

  • Large File Processing: Applications in a wide range of fields, which require simple sequential process of one or more large files such as scientific computing and large-scale financial processing.
  • Large Database Queries: Applications that involve scans or joins of large relational tables, such as those performed for data mining or business intelligence.
  • Video on Demand: Applications that provide individualized video entertainment to a community of subscribers by drawing from a digital film library.

SPC-2 is built to:

  • Provide a level playing field for test sponsors.
  • Produce results that are powerful and yet simple to use.
  • Provide value for engineers as well as IT consumers and solution integrators.
  • Is easy to run, easy to audit/verify, and easy to use to report official results.

See Also

Disclosure Statement

SPC-2 and SPC-2 MBPS are registered trademarks of Storage Performance Council (SPC). Results as of March 17, 2015, for more information see www.storageperformance.org.

Oracle ZFS Storage ZS4-4 - B00072, Oracle ZFS Storage ZS3-2 - BE00002, Oracle ZFS Storage ZS3-4 - B00067, Fujitsu ETERNUS DX80 S2, B00055, Fujitsu ETERNUS DX8870 S2 - B00063, Fujitsu ETERNUS DX200 S3 - B00071, HP StorageWorks P9500 XP Storage Array - B00056, HP XP7 Storage Array - B00070, Hitachi Unified Storage VM - B00069, Hitachi Virtual Storage Platform (VSP) - B00060, IBM SAN VC v6.4 - B00061, IBM System Storage DS8870 - B00062, IBM XIV Storage System Gen3 - BE00001, NEC Storage M700 - B00066, SGI InfiniteStorage 5500-SP - B00059, SGI InfiniteStorage 5600 - B00065, Sun StorageTek 2530 - B00026.

Wednesday Jun 25, 2014

Oracle ZFS Storage ZS3-2 Delivers World Record Price-Performance on SPC-2/E

The Oracle ZFS Storage ZS3-2 appliance delivered a world record Price-Performance result, world record energy result and excellent overall performance for the SPC-2/E benchmark.

  • The Oracle ZFS Storage ZS3-2 appliance delivered the top SPC-2 Price-Performance of $12.08 and it delivered an overall score of 16,212.66 SPC-2 MBPS for the SPC-2/E benchmark.

  • The Oracle ZFS Storage ZS3-2 appliance produced the top Performance-Energy SPC-2/E benchmark result of 3.67 SPC2 MBPS / watt.

  • Oracle holds the top two performance results on the SPC-2 benchmark for HDD based systems.

  • The Oracle ZFS Storage ZS3-2 appliance has an 11x price-performance advantage over the IBM DS8870.

  • The Oracle ZFS Storage ZS3-2 appliance has an 8x price-performance advantage over the Hitachi Virtual Storage Platform (VSP).

  • The Oracle ZFS Storage ZS3-2 appliance has an 7.3x price-performance advantage over the HP P9500 XP disk array.

Performance Landscape

SPC-2 Price-Performance

Below is a table of the top SPC-2 Price-Performance results for HDD storage based systems, presented in increasing price-performance order (as of 06/25/2014). The complete set of results may be found at SPC2 top 10 Price-Performance list.

System SPC-2
MBPS
$/SPC-2
MBPS
Results
Identifier
Oracle ZFS Storage ZS3-2 16,212.66 $12.08 BE00002
SGI InfiniteStorage 5600 8,855.70 $15.97 B00065
Oracle ZFS Storage ZS3-4 17,244.22 $22.53 B00067
NEC Storage M700 14,408.89 $25.10 B00066
Sun StorageTek 2530 663.51 $26.48 B00026
Fujitsu ETERNUS DX80 S2 2,685.50 $28.48 B00055
SGI InfiniteStorage 5500-SP 4,064.49 $28.57 B00059
Hitachi Unified Storage VM 11,274.83 $32.64 B00069

SPC-2 MBPS = the Performance Metric
$/SPC-2 MBPS = the Price-Performance Metric
Results Identifier = A unique identification of the result

SPC-2/E Results

The table below list all SPC-2/E results. The SPC-2/E benchmark extends the SPC-2 benchmark by additionally measuring power consumption during the SPC-2 benchmark run.

System SPC-2
MBPS
$/SPC-2
MBPS
TSC Price SPC2 MBPS /
watt
Results
Identifier
Oracle ZFS Storage ZS3-2 16,212.66 $12.08 $195,915 3.67 BE00002
IBM XIV Storage System Gen3 7,467.99 $152.34 $1,137,641 0.81 BE00001

SPC-2 MBPS = the Performance Metric
$/SPC-2 MBPS = the Price-Performance Metric
TSC Price = Total Cost of Ownership Metric
SPC2 MBPS / watt = Number of SPC2 MB/second produced per watt consumed. Higher is Better.
Results Identifier = A unique identification of the result

SPC-2 Performance

The following table list the top SPC-2 -Performance results for HDD storage based systems, presented in decreasing performance order (as of 06/25/2014). The complete set of results may be found at the SPC2 top 10 Performance list.

System SPC-2
MBPS
$/SPC-2
MBPS
TSC Price Results
Identifier
Oracle ZFS Storage ZS3-4 17,244.22 $22.53 $388,472 B00067
Oracle ZFS Storage ZS3-2 16,212.66 $12.08 $195,915 BE00002
Fujitsu ETERNUS DX8870 S2 16,038.74 $79.51 $1,275,163 B00063
IBM System Storage DS8870 15,423.66 $131.21 $2,023,742 B00062
IBM SAN VC v6.4 14,581.03 $129.14 $1,883,037 B00061
Hitachi Virtual Storage Platform (VSP) 13,147.87 $95.38 $1,254,093 B00060
HP StorageWorks P9500 XP Storage Array 13,147.87 $88.34 $1,161,504 B00056

SPC-2 MBPS = the Performance Metric
$/SPC-2 MBPS = the Price-Performance Metric
TSC Price = Total Cost of Ownership Metric
Results Identifier = A unique identification of the result Metric

Complete SPC-2 benchmark results may be found at
http://www.storageperformance.org/results/benchmark_results_spc2.

Configuration Summary

Storage Configuration:

Oracle ZFS Storage ZS3-2 storage system in clustered configuration
2 x Oracle ZFS Storage ZS3-2 controllers, each with
4 x 2.1 GHz 8-core Intel Xeon processors
512 GB memory
12 x Sun Disk shelves, each with
24 x 300 GB 10K RPM SAS-2 drives

Benchmark Description

SPC Benchmark 2 (SPC-2): Consists of three distinct workloads designed to demonstrate the performance of a storage subsystem during the execution of business critical applications that require the large-scale, sequential movement of data. Those applications are characterized predominately by large I/Os organized into one or more concurrent sequential patterns. A description of each of the three SPC-2 workloads is listed below as well as examples of applications characterized by each workload.

  • Large File Processing: Applications in a wide range of fields, which require simple sequential process of one or more large files such as scientific computing and large-scale financial processing.
  • Large Database Queries: Applications that involve scans or joins of large relational tables, such as those performed for data mining or business intelligence.
  • Video on Demand: Applications that provide individualized video entertainment to a community of subscribers by drawing from a digital film library.

SPC-2 is built to:

  • Provide a level playing field for test sponsors.
  • Produce results that are powerful and yet simple to use.
  • Provide value for engineers as well as IT consumers and solution integrators.
  • Is easy to run, easy to audit/verify, and easy to use to report official results.

SPC Benchmark 2/Energy (SPC-2/E): consists of the complete set of SPC-2 performance measurement and reporting plus the measurement and reporting of energy use. This benchmark extension provides measurement and reporting to complete storage configurations, complementing SPC-2C/E, which focuses on storage component configurations.

See Also

Disclosure Statement

SPC-2 and SPC-2 MBPS are registered trademarks of Storage Performance Council (SPC). Results as of June 25, 2014, for more information see www.storageperformance.org.

Fujitsu ETERNUS DX80 S2, B00055, Fujitsu ETERNUS DX8870 S2 - B00063, HP StorageWorks P9500 XP Storage Array - B00056, Hitachi Unified Storage VM - B00069, Hitachi Virtual Storage Platform (VSP) - B00060, IBM SAN VC v6.4 - B00061, IBM System Storage DS8870 - B00062, IBM XIV Storage System Gen3 - BE00001, NEC Storage M700 - B00066, Oracle ZFS Storage ZS3-2 - BE00002, Oracle ZFS Storage ZS3-4 - B00067, SGI InfiniteStorage 5500-SP - B00059, SGI InfiniteStorage 5600 - B00065, Sun StorageTek 2530 - B00026.

Thursday Mar 27, 2014

SPARC M6-32 Produces SAP SD Two-Tier Benchmark World Record for 32-Processor Systems

Oracle's SPARC M6-32 server produced a world record result for 32-processors on the SAP two-tier Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement Package 5 for SAP ERP 6.0 (32 chips / 384 cores / 3072 threads).

  • SPARC M6-32 server achieved 140,000 SAP SD benchmark users with a low average dialog response time of 0.58 seconds running the SAP two-tier Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement package 5 for SAP ERP 6.0.

  • The SPARC M6-32 delivered 2.5 times more users than the IBM Power 780 result using SAP Enhancement Package 5 for SAP ERP 6.0. The IBM result also had 1.7 times worse average dialog response time compared to the SPARC M6-32 server result.

  • The SPARC M6-32 delivered 3.0 times more users than the Fujitsu PRIMEQUEST 2800E (with Intel Xeon E7-8890 v2 processors) result. The Fujitsu result also had 1.7 times worse average dialog response time compared to the SPARC M6-32 server result.

  • The SPARC M6-32 server solution was run with Oracle Solaris 11 and used Oracle Database 11g.

Performance Landscape

SAP-SD 2-Tier Performance Table (in decreasing performance order). With SAP ERP 6.0 Enhancement Package 4 for SAP ERP 6.0 (Old version of the benchmark, obsolete at the end of April, 2012), and SAP ERP 6.0 Enhancement Package 5 for SAP ERP 6.0 results (current version of the benchmark as of May, 2012).

System
Processor
Ch / Co / Th — Memory
OS
Database
Users Resp Time
(sec)
Version Cert#
Fujitsu SPARC M10-4S
SPARC64 X @3.0 GHz
40 / 640 / 1280 — 10 TB
Solaris 11
Oracle 11g
153,000 0.87 EHP5 2013014
SPARC M6-32 Server
SPARC M6 @3.6 GHz
32 / 384 / 3072 — 16 TB
Solaris 11
Oracle 11g
140,000 0.58 EHP5 2014008
IBM Power 795
POWER7 @4 GHz
32 / 256 / 1024 — 4 TB
AIX 7.1
DB2 9.7
126,063 0.98 EHP4 2010046
IBM Power 780
POWER7+ @3.72 GHz
12 / 96 / 834 — 1536 GB
AIX 7.1
DB2 10
57,024 0.98 EHP5 2012033
Fujitsu PRIMEQUEST 2800E
Intel Xeon E7-8890 v2 @2.8 GHz
8 / 120 / 240 — 1024 GB
Windows Server 2012 SE
SQL Server 2012
47,500 0.97 EHP5 2014003
IBM Power 760
POWER7+ @3.41 GHz
8 / 48 / 192 — 1024 GB
AIX 7.1
DB2 10
25,488 0.99 EHP5 2013004

Version – Version of SAP, EHP5 refers to SAP ERP 6.0 Enhancement Package 5 for SAP ERP 6.0 and EHP4 refers to SAP ERP 6.0 Enhancement Package 4 for SAP ERP 6.0

Ch / Co / Th – Total chips, coreas and threads

Complete benchmark results may be found at the SAP benchmark website http://www.sap.com/benchmark.

Configuration Summary and Results

Hardware Configuration:

1 x SPARC M6-32 server with
32 x 3.6 GHz SPARC M6 processors (total of 32 processors / 384 cores / 3072 threads)
16 TB memory
6 x Sun Server X3-2L each with
2 x Intel Xeon E5-2609 2.4 GHz Processors
16 GB Memory
4 x Flash Accelerator F40
12 x 3 TB SAS disks
2 x Sun Server X3-2L each with
2 x Intel Xeon E5-2609 2.4 GHz Processors
16 GB Memory
1 x 8-Port 6Gbps SAS-2 RAID PCI Express HBA
12 x 3 TB SAS disks

Software Configuration:

Oracle Solaris 11
SAP Enhancement Package 5 for SAP ERP 6.0
Oracle Database 11g Release 2

Certified Results (published by SAP)

Number of SAP SD benchmark users:
140,000
Average dialog response time:
0.58 seconds
Throughput:

  Fully processed order line items per hour:
15,878,670
  Dialog steps per hour:
47,636,000
  SAPS:
793,930
Average database request time (dialog/update):
0.020 sec / 0.041 sec
SAP Certification:
2014008

Benchmark Description

The SAP Standard Application SD (Sales and Distribution) Benchmark is an ERP business test that is indicative of full business workloads of complete order processing and invoice processing, and demonstrates the ability to run both the application and database software on a single system. The SAP Standard Application SD Benchmark represents the critical tasks performed in real-world ERP business environments.

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

Two-tier SAP Sales and Distribution (SD) standard application benchmarks, SAP Enhancement Package 5 for SAP ERP 6.0 as of 3/26/14:

SPARC M6-32 (32 processors, 384 cores, 3072 threads) 140,000 SAP SD users, 32 x 3.6 GHz SPARC M6, 16 TB memory, Oracle Database 11g, Oracle Solaris 11, Cert# 2014008. Fujitsu SPARC M10-4S (40 processors, 640 cores, 1280 threads) 153,000 SAP SD users, 40 x 3.0 GHz SPARC65 X, 10 TB memory, Oracle Database 11g, Oracle Solaris 11, Cert# 2013014. IBM Power 780 (12 processors, 96 cores, 384 threads) 57,024 SAP SD users, 12 x 3.72 GHz IBM POWER7+, 1536 GB memory, DB210, AIX7.1, Cert#2012033. Fujitsu PRIMEQUEST 2800E (8 processors, 120 cores, 240 threads) 47,500 SAP SD users, 8 x 2.8 GHz Intel Xeon Processor E7-8890 v2, 1024 GB memory, SQL Server 2012, Windows Server 2012 Standard Edition, Cert# 2014003. IBM Power 760 (8 processors, 48 cores, 192 threads) 25,488 SAP SD users, 8 x 3.41 GHz IBM POWER7+, 1024 GB memory, DB2 10, AIX 7.1, Cert#2013004.

Two-tier SAP Sales and Distribution (SD) standard application benchmarks, SAP Enhancement Package 4 for SAP ERP 6.0 as of 3/26/14:

IBM Power 795 (32 processors, 256 cores, 1024 threads) 126,063 SAP SD users, 32 x 4 GHz IBM POWER7, 4 TB memory, DB2 9.7, AIX7.1, Cert#2010046.

SAP, R/3, reg TM of SAP AG in Germany and other countries. More info www.sap.com/benchmark

Thursday Jan 23, 2014

SPARC T5-2 Delivers World Record 2-Socket Application Server for SPECjEnterprise2010 Benchmark

Oracle's SPARC T5-2 servers have set the world record for the SPECjEnterprise2010 benchmark using two-socket application servers with a result of 17,033.54 SPECjEnterprise2010 EjOPS. The result used two SPARC T5-2 servers, one server for the application tier and the other server for the database tier.

  • The SPARC T5-2 server delivered 29% more performance compared to the 2-socket IBM PowerLinux server result of 13,161.07 SPECjEnterprise2010 EjOPS.

  • The two SPARC T5-2 servers have 1.2x better price performance than the two IBM PowerLinux 7R2 POWER7+ processor-based servers (based on hardware plus software configuration costs for both tiers). The price performance of the SPARC T5-2 server is $35.99 compared to the IBM PowerLinux 7R2 at $44.75.

  • The SPARC T5-2 server demonstrated 1.5x more performance compared to Oracle's x86-based 2-socket Sun Server X4-2 system (Ivy Bridge) result of 11,259.88 SPECjEnterprise2010 EjOPS. Oracle holds the top x86 2-socket application server SPECjEnterprise2010 result.

  • This SPARC T5-2 server result represents the best performance per socket for a single system in the application tier of 8,516.77 SPECjEnterprise2010 EjOPS per socket.

  • The application server used Oracle Fusion Middleware components including the Oracle WebLogic 12.1 application server and Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.7.0_45. The database server was configured with Oracle Database 12c Release 1.

  • This result demonstrated less than 1 second average response times for all SPECjEnterprise2010 transactions and represents Jave EE 5.0 transactions generated by 139,000 users.

Performance Landscape

Select 2-socket single application server results. Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results.

SPECjEnterprise2010 Performance Chart
1/22/2014
Submitter EjOPS* Java EE Server DB Server
Oracle 17,033.54 1 x SPARC T5-2
2 x 3.6 GHz SPARC T5
Oracle WebLogic 12c (12.1.2)
1 x SPARC T5-2
2 x 3.6 GHz SPARC T5
Oracle Database 12c (12.1.0.1)
IBM 13,161.07 1x IBM PowerLinux 7R2
2 x 4.2 GHz POWER 7+
WebSphere Application Server V8.5
1x IBM PowerLinux 7R2
2 x 4.2 GHz POWER 7+
IBM DB2 10.1 FP2
Oracle 11,259.88 1x Sun Server X4-2
2 x 2.7 GHz Intel Xeon E5-2697 v2
Oracle WebLogic 12c (12.1.2)
1x Sun Server X4-2L
2 x 2.7 GHz Intel Xeon E5-2697 v2
Oracle Database 12c (12.1.0.1)

* SPECjEnterprise2010 EjOPS (bigger is better)

Configuration Summary

Application Server:

1 x SPARC T5-2 server, with
2 x 3.6 GHz SPARC T5 processors
512 GB memory
2 x 10 GbE dual-port NIC
Oracle Solaris 11.1 (11.1.13.6.0)
Oracle WebLogic Server 12c (12.1.2)
Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.7.0_45

Database Server:

1 x SPARC T5-2 server, with
2 x 3.6 GHz SPARC T5 processors
512 GB memory
1 x 10 GbE dual-port NIC
2 x 8 Gb FC HBA
Oracle Solaris 11.1 (11.1.13.6.0)
Oracle Database 12c (12.1.0.1)

Storage Servers:

2 x Sun Server X4-2L (24-Drive), with
2 x 2.6 GHz Intel Xeon
64 GB memory
1 x 8 Gb FC HBA
4 x Sun Flash Accelerator F80 PCI-E Cards
Oracle Solaris 11.1

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The new SPECjEnterprise2010 benchmark has been re-designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems,

  • The web zone, servlets, and web services
  • The EJB zone
  • JPA 1.0 Persistence Model
  • JMS and Message Driven Beans
  • Transaction management
  • Database connectivity
Moreover, SPECjEnterprise2010 also heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second (SPECjEnterprise2010 EjOPS). The primary metric for the SPECjEnterprise2010 benchmark is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is NO price/performance metric in this benchmark.

Key Points and Best Practices

  • Two Oracle WebLogic server instances on the SPARC T5-2 server were hosted in 2 separate Oracle Solaris Zones.
  • The Oracle WebLogic application servers were executed in the FX scheduling class to improve performance by reducing the frequency of context switches.
  • The Oracle log writer process was run in the RT scheduling class.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Results from www.spec.org as of 1/22/2014. SPARC T5-2, 17,033.54 SPECjEnterprise2010 EjOPS; IBM PowerLinux 7R2, 13,161.07 SPECjEnterprise2010 EjOPS; Sun Server X4-2, 11,259.88 SPECjEnterprise2010 EjOPS.

The SPARC T5-2 configuration cost is the total application and database server hardware plus software. List price is $613,052 from http://www.oracle.com as of 1/22/2014. The IBM PowerLinux 7R2 configuration total hardware plus software list price is $588,970 based on public pricing from http://www.ibm.com as of 1/22/2014. Pricing does not include database storage hardware for IBM or Oracle.

Thursday Sep 26, 2013

SPARC T5-8 Delivers World Record Single Server SPECjEnterprise2010 Benchmark, Utilizes Virtualized Environment

Oracle produced a world record single-server SPECjEnterprise2010 benchmark result of 36,571.36 SPECjEnterprise2010 EjOPS using one of Oracle's SPARC T5-8 servers for both the application and the database tier. Oracle VM Server for SPARC was used to virtualize the system to achieve this result.

  • The 8-chip SPARC T5 processor based server is 3.3x faster than the 8-chip IBM Power 780 server (POWER7+ processor based).

  • The SPARC T5-8 has 4.4x better price performance than the IBM Power 780, a POWER7+ processor based server (based on hardware plus software configuration costs). The price performance of the SPARC T5-8 server is $40.68 compared to the IBM Power 780 at $177.41. The IBM Power 780, POWER7+ based system has 1.2x better performance per core, but this did not reduce the total software and hardware cost to the customer. As shown by this comparison, performance-per-core is a poor predictor of characteristics relevant to customers. The SPARC T5-8 virtualized price performance was also less than the low-end IBM PowerLinux 7R2 at $62.26.

  • The SPARC T5-8 server ran the Oracle Solaris 11.1 operating system and used Oracle VM Server for SPARC to consolidate ten Oracle WebLogic application server instances and one database server instance to achieve this result.

  • This result demonstrated sub-second average response times for all SPECjEnterprise2010 transactions and represents JEE 5.0 transactions generated by 299,000 users.

  • The SPARC T5-8 server requires only 8 rack units, the same as the space of the IBM Power 780. In this configuration IBM has a hardware core density of 4 cores per rack unit which contrasts with the 16 cores per rack unit for the SPARC T5-8 server. This again demonstrates why performance-per-core is a poor predictor of characteristics relevant to customers.

  • The application server used Oracle Fusion Middleware components including the Oracle WebLogic 12.1 application server and Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.7.0_25. The database server was configured with Oracle Database 12c Release 1.

  • The SPARC T5-8 server is 2.8x faster than a non-virtualized IBM POWER7+ based server result (one server for application and one server for database), the IBM PowerLinux 7R2 achieved 13,161.07 SPECjEnterprise2010 EjOPS.

Performance Landscape

SPECjEnterprise2010 Performance Chart
Only Three Virtualized Results (App+DB on 1 Server) as of 9/23/2013
Submitter EjOPS* Chips per Server Java EE Server & DB Server
App DB
Oracle 36,571.36 5 3 1 x SPARC T5-8
8 chips, 128 cores, 3.6 GHz SPARC T5
Oracle WebLogic 12c (12.1.2)
Oracle Database 12c (12.1.0.1)
Oracle 27,843.57 4 4 1 x SPARC T5-8
8 chips, 128 cores, 3.6 GHz SPARC T5
Oracle WebLogic 12c (12.1.1)
Oracle Database 11g (11.2.0.3)
IBM 10,902.30 4 4 1 x IBM Power 780
8 chips, 32 cores, 4.42 GHz POWER7+
WebSphere Application Server V8.5
IBM DB2 Universal Database 10.1

* SPECjEnterprise2010 EjOPS (bigger is better)

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results.

Configuration Summary

Oracle Summary

Application and Database Server:

1 x SPARC T5-8 server, with
8 x 3.6 GHz SPARC T5 processors
2 TB memory
9 x 10 GbE dual-port NIC
6 x 8 Gb dual-port HBA
Oracle Solaris 11.1 SRU 10.5
Oracle VM Server for SPARC
Oracle WebLogic Server 12c (12.1.2)
Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.7.0_25
Oracle Database 12c (12.1.0.1)

Storage Servers:

6 x Sun Server X3-2L (12-Drive), with
2 x 2.4 GHz Intel Xeon
16 GB memory
1 x 8 Gb FC HBA
4 x Sun Flash Accelerator F40 PCI-E Card
Oracle Solaris 11.1

2 x Sun Storage 2540-M2 Array
12 x 600 GB 15K RPM SAS HDD

Switch Hardware:

1 x Sun Network 10 GbE 72-port Top of Rack (ToR) Switch

IBM Summary

Application and Database Server:

1 x IBM Power 780 server, with
8 x 4.42 GHz POWER7+ processors
786 GB memory
6 x 10 GbE dual-port NIC
3 x 8 Gb four-port HBA
IBM AIX V7.1 TL2
IBM WebSphere Application Server V8.5
IBM J9 VM (build 2.6, JRE 1.7.0 IBM J9 AIX ppc-32)
IBM DB2 10.1
IBM InfoSphere Optim pureQuery Runtime v3.1.1

Storage:

2 x DS5324 Disk System with
48 x 146 GB 15K E-DDM Disks

1 x v7000 Disk Controller with
16 x 400 GB SSD Disks

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The new SPECjEnterprise2010 benchmark has been re-designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems,
  • The web zone, servlets, and web services
  • The EJB zone
  • JPA 1.0 Persistence Model
  • JMS and Message Driven Beans
  • Transaction management
  • Database connectivity
Moreover, SPECjEnterprise2010 also heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second (SPECjEnterprise2010 EjOPS). The primary metric for the SPECjEnterprise2010 benchmark is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is NO price/performance metric in this benchmark.

Key Points and Best Practices

  • Ten Oracle WebLogic server instances on the SPARC T5-8 server were hosted in 10 separate Oracle Solaris Zones within a separate guest domain on 80 cores (5 cpu chips).
  • The database ran in a separate guest domain consisting of 47 cores (3 cpu chips). One core was reserved for the primary domain.
  • The Oracle WebLogic application servers were executed in the FX scheduling class to improve performance by reducing the frequency of context switches.
  • The Oracle log writer process was run in the FX scheduling class at processor priority 60 to use the Critical Thread feature.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Results from www.spec.org as of 9/23/2013. SPARC T5-8, 36,571.36 SPECjEnterprise2010 EjOPS (using Oracle VM for SPARC and 5+3 split); SPARC T5-8, 27,843.57 SPECjEnterprise2010 EjOPS (using Oracle Zones and 4+4 split); IBM Power 780, 10,902.30 SPECjEnterprise2010 EjOPS; IBM PowerLinux 7R2, 13,161.07 SPECjEnterprise2010 EjOPS. SPARC T5-8 server total hardware plus software list price is $1,487,792 from http://www.oracle.com as of 9/20/2013. IBM Power 780 server total hardware plus software cost of $1,934,162 based on public pricing from http://www.ibm.com as of 5/22/2013. IBM PowerLinux 7R2 server total hardware plus software cost of $819,451 based on whywebsphere.com/2013/04/29/weblogic-12c-on-oracle-sparc-t5-8-delivers-half-the-transactions-per-core-at-double-the-cost-of-the-websphere-on-ibm-power7/ retrieved 9/20/2013.

Wednesday Sep 25, 2013

Sun Server X4-2 Delivers Single App Server, 2-Chip x86 World Record SPECjEnterprise2010

Oracle's Sun Server X4-2 and Sun Server X4-2L servers, using the Intel Xeon E5-2697 v2 processor, produced a world record x86 two-chip single application server SPECjEnterprise2010 benchmark result of 11,259.88 SPECjEnterprise2010 EjOPS. The Sun Server X4-2 ran the application tier and the Sun Server X4-2L was used for the database tier.

  • The 2-socket Sun Server X4-2 demonstrated 16% better performance when compared to the 2-socket IBM X3650 M4 server result of 9,696.43 SPECjEnterprise2010 EjOPS.

  • This result used Oracle WebLogic Server 12c, Java HotSpot(TM) 64-Bit Server 1.7.0_40, Oracle Database 12c, and Oracle Linux.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results. The table below shows the top single application server, two-chip x86 results.

SPECjEnterprise2010 Performance Chart
as of 9/22/2013
Submitter EjOPS* Application Server Database Server
Oracle 11,259.88 1x Sun Server X4-2
2x 2.7 GHz Intel Xeon E5-2697 v2
Oracle WebLogic 12c (12.1.2)
1x Sun Server X4-2L
2x 2.7 GHz Intel Xeon E5-2697 v2
Oracle Database 12c (12.1.0.1)
IBM 9,696.43 1x IBM X3650 M4
2x 2.9 GHz Intel Xeon E5-2690
WebSphere Application Server V8.5
1x IBM X3650 M4
2x 2.9 GHz Intel Xeon E5-2690
IBM DB2 10.1
Oracle 8,310.19 1x Sun Server X3-2
2x 2.9 GHz Intel Xeon E5-2690
Oracle WebLogic 11g (10.3.6)
1x Sun Server X3-2L
2x 2.9 GHz Intel Xeon E5-2690
Oracle Database 11g (11.2.0.3)

* SPECjEnterprise2010 EjOPS, bigger is better.

Configuration Summary

Application Server:

1 x Sun Server X4-2
2 x 2.7 GHz Intel Xeon processor E5-2697 v2
256 GB memory
4 x 10 GbE NIC
Oracle Linux 5 Update 9 (kernel-2.6.39-400.124.1.el5uek)
Oracle WebLogic Server 12c (12.1.2)
Java HotSpot(TM) 64-Bit Server VM on Linux, version 1.7.0_40 (Java SE 7 Update 40)

Database Server:

1 x Sun Server X4-2L
2 x 2.7 GHz Intel Xeon E5-2697 v2
256 GB memory
1 x 10 GbE NIC
2 x FC HBA
3 x Sun StorageTek 2540 M2
Oracle Linux 5 Update 9 (kernel-2.6.39-400.124.1.el5uek)
Oracle Database 12c Enterprise Edition Release 12.1.0.1

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The SPECjEnterprise2010 benchmark has been designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems.

The workload consists of an end to end web based order processing domain, an RMI and Web Services driven manufacturing domain and a supply chain model utilizing document based Web Services. The application is a collection of Java classes, Java Servlets, Java Server Pages, Enterprise Java Beans, Java Persistence Entities (pojo's) and Message Driven Beans.

The SPECjEnterprise2010 benchmark heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second ("SPECjEnterprise2010 EjOPS"). This metric is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is no price/performance metric in this benchmark.

Key Points and Best Practices

  • Four Oracle WebLogic server instances were started using numactl binding 2 instances per chip.
  • Two Oracle database listener processes were started and each was bound to a separate chip.
  • Additional tuning information is in the report at http://spec.org.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Sun Server X4-2, 11,259.88 SPECjEnterprise2010 EjOPS; Sun Server X3-2, 8,310.19 SPECjEnterprise2010 EjOPS; IBM System X3650 M4, 9,696.43 SPECjEnterprise2010 EjOPS. Results from www.spec.org as of 9/22/2013.

Tuesday Sep 10, 2013

Oracle ZFS Storage ZS3-4 Delivers World Record SPC-2 Performance

The Oracle Storage ZS3-4 storage system delivered a world record performance result for the SPC-2 benchmark along with excellent price-performance.

  • The Oracle Storage ZS3-4 storage system delivered an overall score of 17,244.22 SPC-2 MBPS™ and a SPC-2 price-performance of $22.53 on the SPC-2 benchmark.

  • This is over a 1.6X generational improvement in performance and over a 1.5X generational improvement in price-performance than over Oracle's Sun ZFS Storage 7420 SPC-2 Benchmark results.

  • The Oracle ZFS Storage ZS3-4 storage system has 6.8X better overall throughput and nearly 1.2X better price-performance than the IBM DS3524 Express turbo, which is IBM's best overall price-performance score on the SPC-2 benchmark.

  • The Oracle ZFS Storage ZS3-4 storage system has over 1.1X overall throughput and 5.8X better price-performance than the IBM DS8870, which is IBM's best overall performance score on the SPC-2 benchmark.

  • The Oracle ZFS Storage ZS3-4 storage system has over 1.3X overall throughput and 3.9X better price-performance than the HP StorageWorks P9500XP Disk Array on the SPC-2 benchmark.

Performance Landscape

SPC-2 Performance Chart (in decreasing performance order)

System SPC-2
MB/s
$/SPC-2
MB/s
ASU
Capacity
(GB)
TSC Price Data
Protection
Level
Date Results
Identifier
Oracle ZFS Storage ZS3-4 17,244.22 $22.53 31,611 $388,472 Mirroring 09/10/13 B00067
Fujitsu DX8700 S2 16,039 $79.51 71,404 $1,275,163 Mirroring 12/03/12 B00063
IBM DS8870 15,424 $131.21 30,924 $2,023,742 RAID-5 10/03/12 B00062
IBM SAN VC v6.4 14,581 $129.14 74,492 $1,883,037 RAID-5 08/01/12 B00061
NEC Storage M700 14,409 $25.13 53,550 $361,613 Mirroring 08/19/12 B00066
Hitachi VSP 13,148 $95.38 129,112 $1,254,093 RAID-5 07/27/12 B00060
HP StorageWorks P9500 13,148 $88.34 129,112 $1,161,504 RAID-5 03/07/12 B00056
Sun ZFS Storage 7420 10,704 $35.24 31,884 $377,225 Mirroring 04/12/12 B00058
IBM DS8800 9,706 $270.38 71,537 $2,624,257 RAID-5 12/01/10 B00051
HP XP24000 8,725 $187.45 18,401 $1,635,434 Mirroring 09/08/08 B00035

SPC-2 MB/s = the Performance Metric
$/SPC-2 MB/s = the Price-Performance Metric
ASU Capacity = the Capacity Metric
Data Protection = Data Protection Metric
TSC Price = Total Cost of Ownership Metric
Results Identifier = A unique identification of the result Metric

SPC-2 Price-Performance Chart (in increasing price-performance order)

System SPC-2
MB/s
$/SPC-2
MB/s
ASU
Capacity
(GB)
TSC Price Data
Protection
Level
Date Results
Identifier
SGI InfiniteStorage 5600 8,855.70 $15.97 28,748 $141,393 RAID6 03/06/13 B00065
Oracle ZFS Storage ZS3-4 17,244.22 $22.53 31,611 $388,472 Mirroring 09/10/13 B00067
Sun Storage J4200 548.80 $22.92 11,995 $12,580 Unprotected 07/10/08 B00033
NEC Storage M700 14,409 $25.13 53,550 $361,613 Mirroring 08/19/12 B00066
Sun Storage J4400 887.44 $25.63 23,965 $22,742 Unprotected 08/15/08 B00034
Sun StorageTek 2530 672.05 $26.15 1,451 $17,572 RAID5 08/16/07 B00026
Sun StorageTek 2530 663.51 $26.48 854 $17,572 Mirroring 08/16/07 B00025
Fujitsu ETERNUS DX80 1,357.55 $26.70 4,681 $36,247 Mirroring 03/15/10 B00050
IBM DS3524 Express Turbo 2,510 $26.76 14,374 $67,185 RAID-5 12/31/10 B00053
Fujitsu ETERNUS DX80 S2 2,685.50 $28.48 17,231 $76,475 Mirroring 08/19/11 B00055

SPC-2 MB/s = the Performance Metric
$/SPC-2 MB/s = the Price-Performance Metric
ASU Capacity = the Capacity Metric
Data Protection = Data Protection Metric
TSC Price = Total Cost of Ownership Metric
Results Identifier = A unique identification of the result Metric

Complete SPC-2 benchmark results may be found at http://www.storageperformance.org/results/benchmark_results_spc2.

Configuration Summary

Storage Configuration:

Oracle ZFS Storage ZS3-4 storage system in clustered configuration
2 x Oracle ZFS Storage ZS3-4 controllers, each with
4 x 2.4 GHz 10-core Intel Xeon processors
1024 GB memory
16 x Sun Disk shelves, each with
24 x 300 GB 15K RPM SAS-2 drives

Benchmark Description

SPC Benchmark-2 (SPC-2): Consists of three distinct workloads designed to demonstrate the performance of a storage subsystem during the execution of business critical applications that require the large-scale, sequential movement of data. Those applications are characterized predominately by large I/Os organized into one or more concurrent sequential patterns. A description of each of the three SPC-2 workloads is listed below as well as examples of applications characterized by each workload.

  • Large File Processing: Applications in a wide range of fields, which require simple sequential process of one or more large files such as scientific computing and large-scale financial processing.
  • Large Database Queries: Applications that involve scans or joins of large relational tables, such as those performed for data mining or business intelligence.
  • Video on Demand: Applications that provide individualized video entertainment to a community of subscribers by drawing from a digital film library.

SPC-2 is built to:

  • Provide a level playing field for test sponsors.
  • Produce results that are powerful and yet simple to use.
  • Provide value for engineers as well as IT consumers and solution integrators.
  • Is easy to run, easy to audit/verify, and easy to use to report official results.

See Also

Disclosure Statement

SPC-2 and SPC-2 MBPS are registered trademarks of Storage Performance Council (SPC). Results as of September 10, 2013, for more information see www.storageperformance.org. Oracle ZFS Storage ZS3-4 B00067, Fujitsu ET 8700 S2 B00063, IBM DS8870 B00062, IBM S.V.C 6.4 B00061, NEC Storage M700 B00066, Hitachi VSP B00060, HP P9500 XP Disk Array B00056, IBM DS8800 B00051.

Monday Jul 01, 2013

Quick Note about Blog Posting from John

Please have a look at John's Blog if you are interested in SPEC performance and how people use it in different ways to come to different conclusions.

Wednesday Jun 12, 2013

SPARC T5-4 Produces World Record Single Server TPC-H @3000GB Benchmark Result

Oracle's SPARC T5-4 server delivered world record single server performance of 409,721 QphH@3000GB with price/performance of $3.94/QphH@3000GB on the TPC-H @3000GB benchmark. This result shows that the 4-chip SPARC T5-4 server is significantly faster than the 8-chip server results from IBM (POWER7 based) and HP (Intel x86 based).

This result demonstrates a complete data warehouse solution that shows the performance both of individual and concurrent query processing streams, faster loading, and refresh of the data during business operations. The SPARC T5-4 server delivers superior performance and cost efficiency when compared to the IBM POWER7 result.

  • The SPARC T5-4 server with four SPARC T5 processors is 2.1 times faster than the IBM Power 780 server with eight POWER7 processors and 2.5 times faster than the HP ProLiant DL980 G7 server with eight x86 processors on the TPC-H @3000GB benchmark. The SPARC T5-4 server also delivered better performance per core than these eight processor systems from IBM and HP.

  • The SPARC T5-4 server with four SPARC T5 processors is 2.1 times faster than the IBM Power 780 server with eight POWER7 processors on the TPC-H @3000GB benchmark.

  • The SPARC T5-4 server costs 38% less per $/QphH@3000GB compared to the IBM Power 780 server with the TPC-H @3000GB benchmark.

  • The SPARC T5-4 server took 2 hours, 6 minutes, 4 seconds for data loading while the IBM Power 780 server took 2.8 times longer.

  • The SPARC T5-4 server executed the first refresh function (RF1) in 19.4 seconds, the IBM Power 780 server took 7.6 times longer.

  • The SPARC T5-4 server with four SPARC T5 processors is 2.5 times faster than the HP ProLiant DL980 G7 server with the same number of cores on the TPC-H @3000GB benchmark.

  • The SPARC T5-4 server took 2 hours, 6 minutes, 4 seconds for data loading while the HP ProLiant DL980 G7 server took 4.1 times longer.

  • The SPARC T5-4 server executed the first refresh function (RF1) in 19.4 seconds, the HP ProLiant DL980 G7 server took 8.9 times longer.

  • The SPARC T5-4 server delivered 6% better performance than the SPARC Enterprise M9000-64 server and 2.1 times better than the SPARC Enterprise M9000-32 server on the TPC-H @3000GB benchmark.

Performance Landscape

The table lists the leading TPC-H @3000GB results for non-clustered systems.

TPC-H @3000GB, Non-Clustered Systems
System
Processor
P/C/T – Memory
Composite
(QphH)
$/perf
($/QphH)
Power
(QppH)
Throughput
(QthH)
Database Available
SPARC T5-4
3.6 GHz SPARC T5
4/64/512 – 2048 GB
409,721.8 $3.94 345,762.7 485,512.1 Oracle 11g R2 09/24/13
SPARC Enterprise M9000
3.0 GHz SPARC64 VII+
64/256/256 – 1024 GB
386,478.3 $18.19 316,835.8 471,428.6 Oracle 11g R2 09/22/11
SPARC T4-4
3.0 GHz SPARC T4
4/32/256 – 1024 GB
205,792.0 $4.10 190,325.1 222,515.9 Oracle 11g R2 05/31/12
SPARC Enterprise M9000
2.88 GHz SPARC64 VII
32/128/256 – 512 GB
198,907.5 $15.27 182,350.7 216,967.7 Oracle 11g R2 12/09/10
IBM Power 780
4.1 GHz POWER7
8/32/128 – 1024 GB
192,001.1 $6.37 210,368.4 175,237.4 Sybase 15.4 11/30/11
HP ProLiant DL980 G7
2.27 GHz Intel Xeon X7560
8/64/128 – 512 GB
162,601.7 $2.68 185,297.7 142,685.6 SQL Server 2008 10/13/10

P/C/T = Processors, Cores, Threads
QphH = the Composite Metric (bigger is better)
$/QphH = the Price/Performance metric in USD (smaller is better)
QppH = the Power Numerical Quantity
QthH = the Throughput Numerical Quantity

The following table lists data load times and refresh function times during the power run.

TPC-H @3000GB, Non-Clustered Systems
Database Load & Database Refresh
System
Processor
Data Loading
(h:m:s)
T5
Advan
RF1
(sec)
T5
Advan
RF2
(sec)
T5
Advan
SPARC T5-4
3.6 GHz SPARC T5
02:06:04 1.0x 19.4 1.0x 22.4 1.0x
IBM Power 780
4.1 GHz POWER7
05:51:50 2.8x 147.3 7.6x 133.2 5.9x
HP ProLiant DL980 G7
2.27 GHz Intel Xeon X7560
08:35:17 4.1x 173.0 8.9x 126.3 5.6x

Data Loading = database load time
RF1 = power test first refresh transaction
RF2 = power test second refresh transaction
T5 Advan = the ratio of time to T5 time

Complete benchmark results found at the TPC benchmark website http://www.tpc.org.

Configuration Summary and Results

Hardware Configuration:

SPARC T5-4 server
4 x SPARC T5 processors (3.6 GHz total of 64 cores, 512 threads)
2 TB memory
2 x internal SAS (2 x 300 GB) disk drives

External Storage:

12 x Sun Storage 2540-M2 array with Sun Storage 2501-M2 expansion trays, each with
24 x 15K RPM 300 GB drives, 2 controllers, 2 GB cache
2 x Brocade 6510 Fibre Channel Switches (48 x 16 Gbs port each)

Software Configuration:

Oracle Solaris 11.1
Oracle Database 11g Release 2 Enterprise Edition

Audited Results:

Database Size: 3000 GB (Scale Factor 3000)
TPC-H Composite: 409,721.8 QphH@3000GB
Price/performance: $3.94/QphH@3000GB
Available: 09/24/2013
Total 3 year Cost: $1,610,564
TPC-H Power: 345,762.7
TPC-H Throughput: 485,512.1
Database Load Time: 2:06:04

Benchmark Description

The TPC-H benchmark is a performance benchmark established by the Transaction Processing Council (TPC) to demonstrate Data Warehousing/Decision Support Systems (DSS). TPC-H measurements are produced for customers to evaluate the performance of various DSS systems. These queries and updates are executed against a standard database under controlled conditions. Performance projections and comparisons between different TPC-H Database sizes (100GB, 300GB, 1000GB, 3000GB, 10000GB, 30000GB and 100000GB) are not allowed by the TPC.

TPC-H is a data warehousing-oriented, non-industry-specific benchmark that consists of a large number of complex queries typical of decision support applications. It also includes some insert and delete activity that is intended to simulate loading and purging data from a warehouse. TPC-H measures the combined performance of a particular database manager on a specific computer system.

The main performance metric reported by TPC-H is called the TPC-H Composite Query-per-Hour Performance Metric (QphH@SF, where SF is the number of GB of raw data, referred to as the scale factor). QphH@SF is intended to summarize the ability of the system to process queries in both single and multiple user modes. The benchmark requires reporting of price/performance, which is the ratio of the total HW/SW cost plus 3 years maintenance to the QphH. A secondary metric is the storage efficiency, which is the ratio of total configured disk space in GB to the scale factor.

Key Points and Best Practices

  • Twelve of Oracle's Sun Storage 2540-M2 arrays with Sun Storage 2501-M2 expansion trays were used for the benchmark. Each contains 24 15K RPM drives and is connected to a single dual port 16Gb FC HBA using 2 ports through a Brocade 6510 Fibre Channel switch.

  • The SPARC T5-4 server achieved a peak IO rate of 33 GB/sec from the Oracle database configured with this storage.

  • Oracle Solaris 11.1 required very little system tuning.

  • Some vendors try to make the point that storage ratios are of customer concern. However, storage ratio size has more to do with disk layout and the increasing capacities of disks – so this is not an important metric when comparing systems.

  • The SPARC T5-4 server and Oracle Solaris efficiently managed the system load of two thousand Oracle Database parallel processes.

  • Six Sun Storage 2540-M2/2501-M2 arrays were mirrored to another six Sun Storage 2540-M2/25001-M2 arrays on which all of the Oracle database files were placed. IO performance was high and balanced across all the arrays.

  • The TPC-H Refresh Function (RF) simulates periodical refresh portion of Data Warehouse by adding new sales and deleting old sales data. Parallel DML (parallel insert and delete in this case) and database log performance are a key for this function and the SPARC T5-4 server outperformed both the IBM POWER7 server and HP ProLiant DL980 G7 server. (See the RF columns above.)

See Also

Disclosure Statement

TPC-H, QphH, $/QphH are trademarks of Transaction Processing Performance Council (TPC). For more information, see www.tpc.org, results as of 6/7/13. Prices are in USD. SPARC T5-4 www.tpc.org/3288; SPARC T4-4 www.tpc.org/3278; SPARC Enterprise M9000 www.tpc.org/3262; SPARC Enterprise M9000 www.tpc.org/3258; IBM Power 780 www.tpc.org/3277; HP ProLiant DL980 www.tpc.org/3285. 

Wednesday May 01, 2013

SPARC T5-8 Delivers Best Single System SPECjEnterprise2010 Benchmark, Beats IBM

Oracle produced a world record single-server SPECjEnterprise2010 benchmark result of 27,843.57 SPECjEnterprise2010 EjOPS using one of Oracle's SPARC T5-8 servers for both the application and the database tier. This result directly compares the 8-chip SPARC T5-8 server (8 SPARC T5 processors) to the 8-chip IBM Power 780 server (8 POWER7+ processor).

  • The 8-chip SPARC T5 processor based server is 2.6x faster than the 8-chip IBM POWER7+ processor based server.

  • Both Oracle and IBM used virtualization to provide 4-chips for application and 4-chips for database.

  • The server cost/performance for the SPARC T5 processor based server was 6.9x better than the server cost/performance of the IBM POWER7+ processor based server. The cost/performance of the SPARC T5-8 server is $10.72 compared to the IBM Power 780 at $73.83.

  • The total configuration cost/performance (hardware+software) for the SPARC T5 processor based server was 3.6x better than the IBM POWER7+ processor based server. The cost/performance of the SPARC T5-8 server is $56.21 compared to the IBM Power 780 at $199.42. The IBM system had 1.6x better performance per core, but this did not reduce the total software and hardware cost to the customer. As shown by this comparison, performance-per-core is a poor predictor of characteristics relevant to customers.

  • The total IBM hardware plus software cost was $2,174,152 versus the total Oracle hardware plus software cost of $1,565,092. At this price IBM could only provide 768 GB of memory while Oracle was able to deliver 2 TB in the SPARC T5-8 server.

  • The SPARC T5-8 server requires only 8 rack units, the same as the space of the IBM Power 780. In this configuration IBM has a hardware core density of 4 cores per rack unit which contrasts with the 16 cores per rack unit for the SPARC T5-8 server. This again demonstrates why performance-per-core is a poor predictor of characteristics relevant to customers.

  • The virtualized SPARC T5 processor based server ran the application tier servers on 4 chips using Oracle Solaris Zones and the database tier in a 4-chip Oracle Solaris Zone. The virtualized IBM POWER7+ processor based server ran the application in a 4-chip LPAR and the database in a 4-chip LPAR.

  • The SPARC T5-8 server ran the Oracle Solaris 11.1 operating system and used Oracle Solaris Zones to consolidate eight Oracle WebLogic application server instances and one database server instance to achieve this result. The IBM system used LPARS and AIX V7.1.

  • This result demonstrated less than 1 second average response times for all SPECjEnterprise2010 transactions and represents JEE 5.0 transactions generated by 227,500 users.

  • The application server used Oracle Fusion Middleware components including the Oracle WebLogic 12.1 application server and Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.7.0_15. The database server was configured with Oracle Database 11g Release 2.

  • IBM has a non-virtualized result (one server for application and one server for database). The IBM PowerLinux 7R2 achieved 13,161.07 SPECjEnterprise2010 EjOPS which means it was 2.1x slower than the SPARC T5-8 server. The total configuration cost/performance (hardware+software) for the SPARC T5 processor based server was 11% better than the IBM POWER7+ processor based server. The cost/performance of the SPARC T5-8 server is $56.21 compared to the IBM PowerLinux 7R2 at $62.26. As shown by this comparison, performance-per-core is a poor predictor of characteristics relevant to customers.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results.

SPECjEnterprise2010 Performance Chart
Only Two Virtualized Results (App+DB on 1 Server) as of 5/1/2013
Submitter EjOPS* Java EE Server & DB Server
Oracle 27,843.57 1 x SPARC T5-8
8 chips, 128 cores, 3.6 GHz SPARC T5
Oracle WebLogic 12c (12.1.1)
Oracle Database 11g (11.2.0.3)
IBM 10,902.30 1 x IBM Power 780
8 chips, 32 cores, 4.42 GHz POWER7+
WebSphere Application Server V8.5
IBM DB2 Universal Database 10.1

* SPECjEnterprise2010 EjOPS (bigger is better)

Configuration Summary

Oracle Summary

Application and Database Server:

1 x SPARC T5-8 server, with
8 x 3.6 GHz SPARC T5 processors
2 TB memory
5 x 10 GbE dual-port NIC
6 x 8 Gb dual-port HBA
Oracle Solaris 11.1 SRU 4.5
Oracle WebLogic Server 12c (12.1.1)
Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.7.0_15
Oracle Database 11g (11.2.0.3)

Storage Servers:

6 x Sun Server X3-2L (12-Drive), with
2 x 2.4 GHz Intel Xeon
16 GB memory
1 x 8 Gb FC HBA
4 x Sun Flash Accelerator F40 PCI-E Card
Oracle Solaris 11.1

2 x Sun Storage 2540-M2 Array
12 x 600 GB 15K RPM SAS HDD

Switch Hardware:

1 x Sun Network 10 GbE 72-port Top of Rack (ToR) Switch

IBM Summary

Application and Database Server:

1 x IBM Power 780 server, with
8 x 4.42 GHz POWER7+ processors
786 GB memory
6 x 10 GbE dual-port NIC
3 x 8 Gb four-port HBA
IBM AIX V7.1 TL2
IBM WebSphere Application Server V8.5
IBM J9 VM (build 2.6, JRE 1.7.0 IBM J9 AIX ppc-32)
IBM DB2 10.1
IBM InfoSphere Optim pureQuery Runtime v3.1.1

Storage:

2 x DS5324 Disk System with
48 x 146GB 15K E-DDM Disks

1 x v7000 Disk Controller with
16 x 400GB SSD Disks

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The new SPECjEnterprise2010 benchmark has been re-designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems,
  • The web zone, servlets, and web services
  • The EJB zone
  • JPA 1.0 Persistence Model
  • JMS and Message Driven Beans
  • Transaction management
  • Database connectivity
Moreover, SPECjEnterprise2010 also heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second (SPECjEnterprise2010 EjOPS). The primary metric for the SPECjEnterprise2010 benchmark is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is NO price/performance metric in this benchmark.

Key Points and Best Practices

  • Eight Oracle WebLogic server instances on the SPARC T5-8 server were hosted in 8 separate Oracle Solaris Zones to demonstrate consolidation of multiple application servers. The 8 zones were bound to 4 resource pools using 64 cores (4 cpu chips).
  • The database ran in a separate Oracle Solaris Zone bound to a resource pool consisting 64 cores (4 cpu chips). The database shadow processes were run in the FX scheduling class and bound to one of four cpu chips using the plgrp command.
  • The Oracle WebLogic application servers were executed in the FX scheduling class to improve performance by reducing the frequency of context switches.
  • The Oracle log writer process was run in the FX scheduling class at processor priority 60 to use the Critical Thread feature.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Results from www.spec.org as of 5/1/2013. SPARC T5-8, 27,843.57 SPECjEnterprise2010 EjOPS; IBM Power 780, 10,902.30 SPECjEnterprise2010 EjOPS; IBM PowerLinux 7R2, 13,161.07 SPECjEnterprise2010 EjOPS. Oracle server only hardware list price is $298,494 and total hardware plus software list price is $1,565,092 from http://www.oracle.com as of  5/22/2013. IBM server only hardware list price is $804,931 and total hardware plus software cost of $2,174,152 based on public pricing from http://www.ibm.com as of 5/22/2013. IBM PowerLinux 7R2 server total hardware plus software cost of $819,451 based on public pricing from http://www.ibm.com as of 5/22/2013.

Tuesday Mar 26, 2013

SPARC T5-8 Delivers SPECjEnterprise2010 Benchmark World Record Performance

Oracle produced a world record SPECjEnterprise2010 benchmark result of 57,422.17 SPECjEnterprise2010 EjOPS using Oracle's SPARC T5-8 server in the application tier and another SPARC T5-8 server for the database tier.

  • The SPARC T5-8 server demonstrated 3.4x better performance compared to an 8-socket IBM Power 780 server result of 16,646.34 SPECjEnterprise2010 EjOPS. The SPARC T5-8 is 3.7x less expensive for the application server hardware list cost compared to the IBM configuration.

  • The SPARC T5 processor demonstrated 1.7x better performance per core compared to the POWER7 processor used in the IBM Power 780 SPECjEnterprise2010 result.

  • The SPARC T5-8 server demonstrated 2.2x better performance compared to the Cisco UCS B440 M2 Blade Server result of 26,118.67 SPECjEnterprise2010 EjOPS.

  • The SPARC T5-8 servers used in the application and database tiers ran the Oracle Solaris 11.1 operating system.

  • The SPARC T5-8 server for the application tier used Oracle Solaris Zones to consolidate sixteen Oracle WebLogic Server instances to achieve this result.

  • This result demonstrated less than 1 second response time for all SPECjEnterprise2010 transactions, while demonstrating a sustained load of Java EE 5 transactions equivalent to 468,000 users.

  • The SPARC T5-8 application server used Oracle Fusion Middleware components including the Oracle WebLogic 12.1 application server and Oracle JDK 7 Update 15. The SPARC T5-8 database server was configured with Oracle Database 11g Release 2.

  • This result used six Sun Server X3-2L systems each configured with 4 x 400 GB Sun Flash Accelerator F40 PCIe Card devices as storage servers for the database files.

  • This result represents the best performance/socket for a single system in the application tier of 7,177.77 SPECjEnterprise2010 EjOPS per socket.

  • A single SPARC T5-8 server in the application tier producing 57,422.17 SPECjEnterprise2010 EjOPS can replace a total of 4x SPARC T4-4 servers that obtained 40,104.86 SPECjEnterprise2010 EjOPS. A single SPARC T5-8 server in the application tier producing 57,422.17 SPECjEnterprise2010 EjOPS can replace 6x SPARC T3-4 servers where each SPARC T3-4 server obtained 9,456.28 SPECjEnterprise2010 EjOPS.

  • Oracle Fusion Middleware provides a family of complete, integrated, hot pluggable and best-of-breed products known for enabling enterprise customers to create and run agile and intelligent business applications. Oracle WebLogic Server's on-going, record-setting Java application server performance demonstrates why so many customers rely on Oracle Fusion Middleware as their foundation for innovation.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results.

SPECjEnterprise2010 Performance Chart
as of 3/26/2013
Submitter EjOPS* Java EE Server DB Server
Oracle 57,422.17 1 x SPARC T5-8
8 chips, 128 cores, 3.6 GHz SPARC T5
Oracle WebLogic 12c (12.1.1)
1 x SPARC T5-8
8 chips, 128 cores, 3.6 GHz SPARC T5
Oracle Database 11g (11.2.0.3)
Oracle 40,104.86 4 x SPARC T4-4
4 chips, 32 cores, 3.0 GHz SPARC T4
Oracle WebLogic 11g (10.3.5)
2 x SPARC T4-4
4 chips, 32 cores, 3.0 GHz SPARC T4
Oracle Database 11g (11.2.0.2)
Oracle 27,150.05 1x Sun Server X2-8
8x 2.4 GHz Intel Xeon E7-8870
Oracle WebLogic 12c
1x Sun Server X2-4
4x 2.4 GHz Intel Xeon E7-4870
Oracle Database 11g (11.2.0.2)
Cisco 26,118.67 2 x Cisco UCS B440 M2
4 chips, 40 cores, 2.4 GHz Xeon E7-4870
Oracle WebLogic 11g (10.3.5)
1 x Cisco UCS C460 M2
4 chips, 40 cores, 2.4 GHz Xeon E7-4870
Oracle Database 11g (11.2.0.2)
IBM 16,646.34 1 x IBM Power 780
8 chips, 64 cores, 3.86 GHz POWER7
WebSphere Application Server V7.0
1 x IBM Power 750 Express
4 chips, 32 cores, 3.55 GHz POWER7
IBM DB2 Universal Database 9.7

* SPECjEnterprise2010 EjOPS (bigger is better)

Configuration Summary

Application Server:

1 x SPARC T5-8 server, with
8 x 3.6 GHz SPARC T5 processors
2 TB memory
8 x 10 GbE dual-port NIC
Oracle Solaris 11.1 SRU 4.5
Oracle WebLogic Server 12c (12.1.1)
Oracle JDK 7 Update 15

Database Server:

1 x SPARC T5-8 server, with
8 x 3.6 GHz SPARC T5 processors
2 TB memory
5 x 10 GbE dual-port NIC
6 x 8 Gb FC dual-port HBA
Oracle Solaris 11.1 SRU 4.5
Oracle Database 11g Enterprise Edition Release 11.2.0.3

Storage Servers:

6 x Sun Server X3-2L (12-Drive), with
2 x 2.4 GHz Intel Xeon
16 GB memory
1 x 8 Gb FC HBA
4 x Sun Flash Accelerator F40 PCI-E Card
Oracle Solaris 11.1

2 x Sun Storage 2540-M2 Array
12 x 600 GB 15K RPM SAS HDD

Switch Hardware:

1 x Sun Network 10 GbE 72-port Top of Rack (ToR) Switch

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The new SPECjEnterprise2010 benchmark has been re-designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems,
  • The web zone, servlets, and web services
  • The EJB zone
  • JPA 1.0 Persistence Model
  • JMS and Message Driven Beans
  • Transaction management
  • Database connectivity
Moreover, SPECjEnterprise2010 also heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second (SPECjEnterprise2010 EjOPS). The primary metric for the SPECjEnterprise2010 benchmark is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is NO price/performance metric in this benchmark.

Key Points and Best Practices

  • Sixteen Oracle WebLogic server instances on the SPARC T5-8 server were hosted in 16 separate Oracle Solaris Zones to demonstrate consolidation of multiple application servers.
  • Each Oracle Solaris Zone was bound to a separate processor set, each contained total 58 hardware strands. This was done to improve performance by using the physical memory closest to the processors to reduce memory access latency. The default set was used for network and disk interrupt handling.
  • The Oracle WebLogic application servers were executed in the FX scheduling class to improve performance by reducing the frequency of context switches.
  • The Oracle database processes were run in 8 processor sets using psrset(1M) and executed in the FX scheduling class. This improved performance by reducing memory access latency and reducing context switches.
  • The Oracle log writer process was run in a separate processor set containing a single core and run in the RT scheduling class. This insured that the log writer had the most efficient use of CPU resources.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Results from www.spec.org as of 3/26/2013. SPARC T5-8, 57,422.17 SPECjEnterprise2010 EjOPS; SPARC T4-4, 40,104.86 SPECjEnterprise2010 EjOPS; Sun Server X2-8, 27,150.05 SPECjEnterprise2010 EjOPS; Cisco UCS B440 M2, 26,118.67 SPECjEnterprise2010 EjOPS; IBM Power 780, 16,646.34 SPECjEnterprise2010 EjOPS. SPARC T3-4 9456.28 SPECjEnterprise2010 EjOPS.

SPARC T5-8 (SPARC T5-8 Server base package, 8xSPARC T5 16-core processors, 128x16GB-1066 DIMMS, 2x600GB 10K RPM 2.5. SAS-2 HDD, 4x Power Cables) List Price $268,742. IBM Power 780 (IBM Power 780:9179 Model MHB, 8x3.8GHz 16-core, 64x one processor activation, 4xCEC Enclosure with IBM Bezel, I/O Backplane and System Midplane,16x 0/32GB DDR3 Memory (4x8GB) DIMMS-1066MHz Power7 CoD Memory, 12x Activation of 1 GB DDR3 Power7 Memory, 5x Activation of 100GB DDR3 Power7 Memory, 1x Disk/Media Backplane. 2x 146.8GB SAS 15K RPM 2.5. HDD (AIX/Linux only), 4x AC Power Supply 1725W) List Price $992,023. Source: Oracle.com and IBM.com, collected 03/18/2013.

SPARC T5-8 Realizes SAP SD Two-Tier Benchmark World Record for 8 Chip Systems

Oracle's SPARC T5-8 server produced a world record result for systems with 8 processors on the two-tier SAP Sales and Distribution (SD) Standard Application Benchmark.

  • The SPARC T5-8 server achieved 40,000 users with running the two-tier SAP Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement package 5 for SAP ERP 6.0.

  • The SPARC T5-8 server is 57% faster than the IBM Power 760 8-chip running SAP Enhancement Package 5 for SAP ERP 6.0.

  • The SPARC T5-8 server delivers 5% more SAP users per chip than the IBM Power 780 12-chip running SAP Enhancement Package 5 for SAP ERP 6.0.

  • The SPARC T5-8 server solution was run with Oracle Solaris 11 and used Oracle Database 11g.

Performance Landscape

SAP-SD 2-Tier Performance Table (in decreasing performance order). SAP ERP 6.0 Enhancement Pack 5 for SAP ERP 6.0 results (New version of the benchmark as of May 2012).

System OS
Database
Users SAPS SAP
ERP/ECC
Release
Date
SPARC T5-8 Server
8x SPARC T5 @3.6 GHz, 2 TB
Solaris 11
Oracle 11g
40,000 220,950 EHP5 for SAP
ERP 6.0
25-Mar-13
IBM Power 760
8xPOWER7+ @3.41 GHz, 1024 GB
AIX 7.1
DB2 10
25,488 139,220 EHP5 for SAP
ERP 6.0
5-Feb-13

SAP ERP 6.0 Enhancement Pack 4 for SAP ERP 6.0 Results
(Old version of the benchmark, obsolete at the end of April, 2012)

System OS
Database
Users SAPS SAP
ERP/ECC
Release
Date
IBM Power 795
32xPOWER7 @4 GHz, 4 TB
AIX 7.1
DB2 9.7
126,063 688,630 EHP4 for SAP
ERP 6.0
15-Nov-10
SPARC Enterprise Server M9000
64xSPARC64 VII @2.88 GHz, 1152 GB
Solaris 10
Oracle 10g
32,000 175,600 EHP4 for SAP
ERP 6.0
18-Nov-09

Complete benchmark results may be found at the SAP benchmark website http://www.sap.com/benchmark.

Configuration Summary and Results

Hardware Configuration:

1 x SPARC T5-8 server with
8 x 3.6 GHz SPARC T5 processors (total of 8 processors / 128 cores / 1024 threads)
2 TB memory
1 x Sun ZFS Storage 7420 appliance with
72 x 600 GB 15K RPM 3.5" SAS-2 disk
32 x 32 GB memory
1 x Sun Fire X4270 M2 server configured as a COMSTAR device with
10 x 2 TB 7.2K 3.5" SAS disk
18 x 8 GB memory

Software Configuration:

Oracle Solaris 11
SAP enhancement package 5 for SAP ERP 6.0
Oracle Database 11g Release 2

Certified Results (published by SAP)

Performance:
40,000 benchmark users
SAP Certification:
2013008

Benchmark Description

The SAP Standard Application SD (Sales and Distribution) Benchmark is a two-tier ERP business test that is indicative of full business workloads of complete order processing and invoice processing, and demonstrates the ability to run both the application and database software on a single system. The SAP Standard Application SD Benchmark represents the critical tasks performed in real-world ERP business environments.

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

Two-tier SAP Sales and Distribution (SD) Standard Application benchmarks SAP Enhancement package 5 for SAP ERP 6.0 as of 3/26/13:

SPARC T5-8 (8 processors, 128 cores, 1024 threads) 40,000 SAP SD users, 8 x 3.6 GHz SPARC T5, 2 TB memory, Oracle Database 11g, Oracle Solaris 11, Cert# 2013008. IBM Power 760 (8 processors, 48 cores, 192 threads) 25,488 SAP SD users, 8 x 3.41 GHz IBM POWER7+, 1024 GB memory, DB2 10, AIX 7.1, Cert#2013004.

Two-tier SAP Sales and Distribution (SD) Standard Application benchmarks SAP Enhancement package 4 for SAP ERP 6.0 as of 4/30/12:

IBM Power 795 (32 processors, 256 cores, 1024 threads) 126,063 SAP SD users, 32 x 4 GHz IBM POWER7, 4 TB memory, DB2 9.7, AIX7.1, Cert#2010046. SPARC Enterprise Server M9000 (64 processors, 256 cores, 512 threads) 32,000 SAP SD users, 64 x 2.88 GHz SPARC64 VII, 1152 GB memory, Oracle Database 10g, Oracle Solaris 10, Cert# 2009046.

SAP, R/3, reg TM of SAP AG in Germany and other countries. More info www.sap.com/benchmark

SPARC M5-32 Produces SAP SD Two-Tier Benchmark World Record for SAP Enhancement Package 5 for SAP ERP 6.0

Oracle's SPARC M5-32 server produced a world record result on the two-tier SAP Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement package 5 for SAP ERP 6.0.

  • The SPARC M5-32 server achieved 85,050 users running the two-tier SAP Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement package 5 for SAP ERP 6.0.

  • The SPARC M5-32 solution was run with Oracle Solaris 11 and used the Oracle Database 11g.

Performance Landscape

SAP-SD 2-Tier Performance Table (in decreasing performance order). SAP ERP 6.0 Enhancement Pack 5 for SAP ERP 6.0 results (new version of the benchmark as of May, 2012).

System OS
Database
Users SAPS SAP
ERP/ECC
Release
Date
SPARC M5-32 Server
32x SPARC M5 @3.6 GHz, 4 TB
Solaris 11
Oracle 11g
85,050 472,600 EHP5 for SAP
ERP 6.0
25-Mar-13
IBM Power 780
12xPOWER7+ @3.72 GHz, 1536 GB
AIX 7.1
DB2 10
57,024 311,720 EHP5 for SAP
ERP 6.0
3-Oct-12
IBM Power 760
8xPOWER7+ @3.41 GHz, 1024 GB
AIX 7.1
DB2 10
25,488 139,220 EHP5 for SAP
ERP 6.0
5-Feb-13

SAP ERP 6.0 Enhancement Pack 4 for SAP ERP 6.0 Results
(Old version of the benchmark, obsolete at the end of April, 2012)

System OS
Database
Users SAPS SAP
ERP/ECC
Release
Date
IBM Power 795
32xPOWER7 @4 GHz, 4 TB
AIX 7.1
DB2 9.7
126,063 688,630 EHP4 for SAP
ERP 6.0
15-Nov-10
SPARC Enterprise Server M9000
64xSPARC64 VII @2.88 GHz, 1152 GB
Solaris 10
Oracle 10g
32,000 175,600 EHP4 for SAP
ERP 6.0
18-Nov-09

Complete benchmark results may be found at the SAP benchmark website http://www.sap.com/benchmark.

Configuration Summary and Results

Hardware Configuration:

1 x SPARC M5-32 server with
32 x 3.6 GHz SPARC M5 processors (total of 32 processors / 192 cores / 1536 threads)
4 TB memory
1 x Sun Storage 2540-M2 (12 x 300 GB 5K RPM 3.5" SAS-2 disk & 2 GB cache)
Flash Storage

Software Configuration:

Oracle Solaris 11
SAP enhancement package 5 for SAP ERP 6.0
Oracle Database 11g Release 2

Certified Results (published by SAP)

Performance: 85,050 benchmark users
SAP Certification: 2013009

Benchmark Description

The SAP Standard Application SD (Sales and Distribution) Benchmark is a two-tier ERP business test that is indicative of full business workloads of complete order processing and invoice processing, and demonstrates the ability to run both the application and database software on a single system. The SAP Standard Application SD Benchmark represents the critical tasks performed in real-world ERP business environments.

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

Two-tier SAP Sales and Distribution (SD) standard application benchmarks, SAP Enhancement package 5 for SAP ERP 6.0 as of 3/26/13:

SPARC M5-32 (32 processors, 192 cores, 1536 threads) 85,050 SAP SD users, 32 x 3.6 GHz SPARC M5, 4 TB memory, Oracle Database 11g, Oracle Solaris 11, Cert# 2013009. IBM Power 780 (12 processors, 96 cores, 384 threads) 57,024 SAP SD users, 12 x 3.72 GHz IBM POWER7+, 1536 GB memory, DB210, AIX7.1, Cert#2012033. IBM Power 760 (8 processors, 48 cores, 192 threads) 25,488 SAP SD users, 8 x 3.41 GHz IBM POWER7+, 1024 GB memory, DB2 10, AIX 7.1, Cert#2013004.

Two-tier SAP Sales and Distribution (SD) standard application benchmarks, SAP Enhancement package 4 for SAP ERP 6.0 as of 3/26/13:

IBM Power 795 (32 processors, 256 cores, 1024 threads) 126,063 SAP SD users, 32 x 4 GHz IBM POWER7, 4 TB memory, DB2 9.7, AIX7.1, Cert#2010046. SPARC Enterprise Server M9000 (64 processors, 256 cores, 512 threads) 32,000 SAP SD users, 64 x 2.88 GHz SPARC64 VII, 1152 GB memory, Oracle Database 10g, Oracle Solaris 10, Cert# 2009046.

SAP, R/3, reg TM of SAP AG in Germany and other countries. More info www.sap.com/benchmark

SPARC T5 Systems Deliver SPEC CPU2006 Rate Benchmark Multiple World Records

Oracle's SPARC T5 processor based systems delivered world record performance on the SPEC CPU2006 rate benchmarks. This was accomplished with Oracle Solaris 11.1 and Oracle Solaris Studio 12.3 software.

SPARC T5-8

  • The SPARC T5-8 server delivered world record SPEC CPU2006 rate benchmark results for systems with eight processors.

  • The SPARC T5-8 server achieved scores of 3750 SPECint_rate2006, 3490 SPECint_rate_base2006, 3020 SPECfp_rate2006, and 2770 SPECfp_rate_base2006.

  • The SPARC T5-8 server beat the 8 processor IBM Power 760 with POWER7+ processors by 1.7x on the SPECint_rate2006 benchmark and 2.2x on the SPECfp_rate2006 benchmark.

  • The SPARC T5-8 server beat the 8 processor IBM Power 780 with POWER7 processors by 35% on the SPECint_rate2006 benchmark and 14% on the SPECfp_rate2006 benchmark.

  • The SPARC T5-8 server beat the 8 processor HP DL980 G7 with Intel Xeon E7-4870 processors by 1.7x on the SPECint_rate2006 benchmark and 2.1x on the SPECfp_rate2006 benchmark.

SPARC T5-1B

  • The SPARC T5-1B server module delivered world record SPEC CPU2006 rate benchmark results for systems with one processor.

  • The SPARC T5-1B server module achieved scores of 467 SPECint_rate2006, 436 SPECint_rate_base2006, 369 SPECfp_rate2006, and 350 SPECfp_rate_base2006.

  • The SPARC T5-1B server module beat the 1 processor IBM Power 710 Express with a POWER7 processor by 62% on the SPECint_rate2006 benchmark and 49% on the SPECfp_rate2006 benchmark.

  • The SPARC T5-1B server module beat the 1 processor NEC Express5800/R120d-1M with an Intel Xeon E5-2690 processor by 31% on the SPECint_rate2006 benchmark. The SPARC T5-1B server module beat the 1 processor Huawei RH2288 V2 with an Intel Xeon E5-2690 processor by 44% on the SPECfp_rate2006 benchmark.

  • The SPARC T5-1B server module beat the 1 processor Supermicro A+ 1012G-MTF with an AMD Operton 6386 SE processor by 51% on the SPECint_rate2006 benchmark and 65% on the SPECfp_rate2006 benchmark.

Performance Landscape

Complete benchmark results are at the SPEC website, SPEC CPU2006 Results. The tables below provide the new Oracle results, as well as, select results from other vendors.

SPEC CPU2006 Rate Results – Eight Processors
System Processor ch/co/th * Peak Base
SPECint_rate2006
SPARC T5-8 SPARC T5, 3.6 GHz 8/128/1024 3750 3490
IBM Power 780 POWER7, 3.92 GHz 8/64/256 2770 2420
HP DL980 G7 Xeon E7-4870, 2.4 GHz 8/80/160 2180 2070
IBM Power 760 POWER7+, 3.42 GHz 8/48/192 2170 1480
Dell PowerEdge C6145 Opteron 6180 SE, 2.5 GHz 8/96/96 1670 1440
SPECfp_rate2006
SPARC T5-8 SPARC T5, 3.6 GHz 8/128/1024 3020 2770
IBM Power 780 POWER7, 3.92 GHz 8/64/256 2640 2410
HP DL980 G7 Xeon E7-4870, 2.4 GHz 8/80/160 1430 1380
IBM Power 760 POWER7+, 3.42 GHz 8/48/192 1400 1130
Dell PowerEdge C6145 Opteron 6180 SE, 2.5 GHz 8/96/96 1310 1200

* ch/co/th — chips / cores / threads enabled

SPEC CPU2006 Rate Results – One Processor
System Processor ch/co/th * Peak Base
SPECint_rate2006
SPARC T5-1B SPARC T5, 3.6 GHz 1/16/128 467 436
NEC Express5800/R120d-1M Xeon E5-2690, 2.9 GHz 1/8/16 357 343
Supermicro A+ 1012G-MTF Opteron 6386 SE, 2.8 GHz 1/16/16 309 269
IBM Power 710 Express POWER7, 3.556 GHz 1/8/32 289 255
SPECfp_rate2006
SPARC T5-1B SPARC T5, 3.6 GHz 1/16/128 369 350
Huawei RH2288 V2 Xeon E5-2690, 2.9 GHz 1/8/16 257 250
IBM Power 710 Express POWER7, 3.556 GHz 1/8/32 248 229
Supermicro A+ 1012G-MTF Opteron 6386 SE, 2.8 GHz 1/16/16 223 199

* ch/co/th — chips / cores / threads enabled

Configuration Summary

Systems Under Test:

SPARC T5-8
8 x 3.6 GHz SPARC T5 processors
4 TB memory (128 x 32 GB dimms)
2 TB on 8 x 600 GB 10K RPM SAS disks, arranged as 4 x 2-way mirrors
Oracle Solaris 11.1 (SRU 4.6)
Oracle Solaris Studio 12.3 1/13 PSE

SPARC T5-1B
1 x 3.6 GHz SPARC T5 processor
256 GB memory (16 x 16 GB dimms)
157 GB on 2 x 300 GB 10K RPM SAS disks (mirrored)
Oracle Solaris 11.1 (SRU 3.4)
Oracle Solaris Studio 12.3 1/13 PSE

Benchmark Description

SPEC CPU2006 is SPEC's most popular benchmark. It measures:

  • Speed — single copy performance of chip, memory, compiler
  • Rate — multiple copy (throughput)

The benchmark is also divided into integer intensive applications and floating point intensive applications:

  • integer: 12 benchmarks derived from real applications such as perl, gcc, XML processing, and pathfinding
  • floating point: 17 benchmarks derived from real applications, including chemistry, physics, genetics, and weather.

It is also divided depending upon the amount of optimization allowed:

  • base: optimization is consistent per compiled language, all benchmarks must be compiled with the same flags per language.
  • peak: specific compiler optimization is allowed per application.

The overall metrics for the benchmark which are commonly used are:

  • SPECint_rate2006, SPECint_rate_base2006: integer, rate
  • SPECfp_rate2006, SPECfp_rate_base2006: floating point, rate
  • SPECint2006, SPECint_base2006: integer, speed
  • SPECfp2006, SPECfp_base2006: floating point, speed

See Also

Disclosure Statement

SPEC and the benchmark names SPECfp and SPECint are registered trademarks of the Standard Performance Evaluation Corporation. Results as of March 26, 2013 from www.spec.org and this report. SPARC T5-8: 3750 SPECint_rate2006, 3490 SPECint_rate_base2006, 3020 SPECfp_rate2006, 2770 SPECfp_rate_base2006; SPARC T5-1B: 467 SPECint_rate2006, 436 SPECint_rate_base2006, 369 SPECfp_rate2006, 350 SPECfp_rate_base2006.

Tuesday May 01, 2012

SPARC T4 Servers Running Oracle Solaris 11 and Oracle RAC Deliver World Record on PeopleSoft HRMS 9.1

Oracle's SPARC T4-4 server running Oracle's PeopleSoft HRMS Self-Service 9.1 benchmark achieved world record performance with 18,000 interactive users. This was accomplished using a high availability configuration using Oracle Real Application Clusters (RAC) 11g Release 2 software for the database tier running on Oracle Solaris 11. The benchmark configuration included the SPARC T4-4 server for the application tier, a SPARC T4-2 server for the web tier and two SPARC T4-2 servers for the database tier.

  • The combination of the SPARC T4 servers running PeopleSoft HRSS 9.1 benchmark supports 4.5x the number of users an IBM pSeries 570 running PeopleSoft HRSS 8.9, with an average response time 40 percent better than IBM.

  • This result was obtained with two SPARC T4-2 servers running the database service using Oracle Real Application Clusters 11g Release 2 software in a high availability configuration.

  • The two SPARC T4-2 servers in the database tier used Oracle Solaris 11, and Oracle RAC 11g Release 2 software with database shared disk storage managed by Oracle Automatic Storage Management (ASM).

  • The average CPU utilization on one SPARC T4-4 server in the application tier handling 18,000 users is 54 percent, showing significant headroom for growth.

  • The SPARC T4 server for the application tier used Oracle Solaris Containers on Oracle Solaris 10, which provides a flexible, scalable and manageable virtualized environment.

  • The Peoplesoft HRMS Self-Service benchmark demonstrates better performance on Oracle hardware and software, engineered to work together, than Oracle software on IBM.

Performance Landscape

PeopleSoft HRMS Self-Service 9.1 Benchmark
Systems Processors Users Ave Response -
Search (sec)
Ave Response -
Save (sec)
SPARC T4-2 (web)
SPARC T4-4 (app)
2 x SPARC T4-2 (db)
2 x SPARC T4, 2.85 GHz
4 x SPARC T4, 3.0 GHz
2 x (2 x SPARC T4, 2.85 GHz)
18,000 1.048 0.742
SPARC T4-2 (web)
SPARC T4-4 (app)
SPARC T4-4 (db)
2 x SPARC T4, 2.85 GHz
4 x SPARC T4, 3.0 GHz
4 x SPARC T4, 3.0 GHz
15,000 1.01 0.63
PeopleSoft HRMS Self-Service 8.9 Benchmark
IBM Power 570 (web/app)
IBM Power 570 (db)
12 x POWER5, 1.9 GHz
4 x POWER5, 1.9 GHz
4,000 1.74 1.25
IBM p690 (web)
IBM p690 (app)
IBM p690 (db)
4 x POWER4, 1.9 GHz
12 x POWER4, 1.9 GHz
6 x 4392 MIPS/Gen1
4,000 1.35 1.01

The main differences between version 9.1 and version 8.9 of the benchmark are:

  • the database expanded from 100K employees and 20K managers to 500K employees and 100K managers,
  • the manager data was expanded,
  • a new transaction, "Employee Add Profile," was added, the percent of users executing it is less then 2%, and the transaction has a heavier footprint,
  • version 9.1 has a different benchmark metric (Average Response Search/Save time for x number of users) versus single user search/save time,
  • newer versions of the PeopleSoft application and PeopleTools software are used.

Configuration Summary

Application Server:

1 x SPARC T4-4 server
4 x SPARC T4 processors 3.0 GHz
512 GB main memory
5 x 300 GB SAS internal disks,
2 x 100 GB internal SSDs
1 x 300 GB internal SSD
Oracle Solaris 10 8/11
PeopleSoft PeopleTools 8.51.02
PeopleSoft HCM 9.1
Oracle Tuxedo, Version 10.3.0.0, 64-bit, Patch Level 031
Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.6.0_20

Web Server:

1 x SPARC T4-2 server
2 x SPARC T4 processors 2.85 GHz
256 GB main memory
2 x 300 GB SAS internal disks
1 x 100 GB internal SSD
Oracle Solaris 10 8/11
PeopleSoft PeopleTools 8.51.02
Oracle WebLogic Server 11g (10.3.3)
Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.6.0_20

Database Server:

2 x SPARC T4-2 servers, each with
2 x SPARC T4 processors 2.85 GHz
128 GB main memory
3 x 300 GB SAS internal disks
Oracle Solaris 11 11/11
Oracle Database 11g Release 2
Oracle Real Application Clusters

Database Storage:

Data
1 x Sun Storage F5100 Flash Array (80 flash modules)
1 x COMSTAR Sun Fire X4470 M2 server
4 x Intel Xeon X7550 processors 2.0 GHz
128 GB main memory
Oracle Solaris 11 11/11
Redo
2 x COMSTAR Sun Fire X4275 servers, each with
1 x Intel Xeon E5540 processor 2.53 GHz
6 GB main memory)
12 x 2 TB SAS disks
Oracle Solaris 11 Express 2010.11

Connectivity:

1 x 8-port 10GbE switch
1 x 24-port 1GbE switch
1 x 32-port Brocade FC switch

Benchmark Description

The purpose of the PeopleSoft HRMS Self-Service 9.1 benchmark is to measure comparative online performance of the selected processes in PeopleSoft Enterprise HCM 9.1 with Oracle Database 11g. The benchmark kit is an Oracle standard benchmark kit run by all platform vendors to measure the performance. It is an OLTP benchmark with no dependency on remote COBOL calls, there is no batch workload, and DB SQLs are moderately complex. The results are certified by Oracle and a white paper is published.

PeopleSoft defines a business transaction as a series of HTML pages that guide a user through a particular scenario. Users are defined as corporate Employees, Managers and HR administrators. The benchmark consists of 14 scenarios which emulate users performing typical HCM transactions such as viewing paychecks, promoting and hiring employees, updating employee profiles and other typical HCM application transactions.

All of these transactions are well defined in the PeopleSoft HR Self-Service 9.1 benchmark kit. This benchmark metric is the Weighted Average Response search/save time for all users.

Key Points and Best Practices

  • The combined processing power of two SPARC T4-2 servers running the highly available Oracle RAC database can provide greater throughput and Oracle RAC scalability than is available from a single server.

  • All database data files/recovery files and Oracle Clusterware files were created with Oracle Automatic Storage Management (Oracle ASM) volume manager and file system which resulted in equivalent performance of conventional volume managers, file systems, and raw devices, but with the added benefit of the ease of management provided by Oracle ASM integrated storage management solution.

  • Five Oracle PeopleSoft Domains with 200 application servers (40 per each Domain) on the SPARC T4-4 server were hosted in two separate Oracle Solaris Containers for a total of 10 Domains/400 application servers processes to demonstrate consolidation of multiple application servers, ease of administration and load balancing.

  • Each Oracle Solaris Container was bound to a separate processor set, each containing 124 virtual processors. The default set (composed of 4 virtual processors from first and third processor socket, total of 8 virtual processors) was used for network and disk interrupt handling. This was done to improve performance by reducing memory access latency by using the physical memory closest to the processors and offload I/O interrupt handling to default set virtual processors, freeing up processing resources for application server virtual processors.

See Also

Disclosure Statement

Oracle's PeopleSoft HRMS 9.1 benchmark, www.oracle.com/us/solutions/benchmark/apps-benchmark/peoplesoft-167486.html, results 5/1/2012.

Sunday Apr 15, 2012

Sun ZFS Storage 7420 Appliance Delivers Top High-End Price/Performance Result for SPC-2 Benchmark

Oracle's Sun ZFS Storage 7420 appliance delivered leading high-end price/performance on the SPC Benchmark 2 (SPC-2).

  • The Sun ZFS Storage 7420 appliance delivered a result of 10,704 SPC-2 MB/s at $35.24 $/SPC-2 MB/s on the SPC-2 benchmark.

  • The Sun ZFS Storage 7420 appliance beats the IBM DS8800 result by over 10% on SPC-2 MB/s and has 7.7x better $/SPC-2 MB/s.

  • The Sun ZFS Storage 7420 appliance achieved the best price/performance for the top 18 posted unique performance results on the SPC-2 benchmark.

Performance Landscape

SPC-2 Performance Chart (in decreasing performance order)

System SPC-2
MB/s
$/SPC-2
MB/s
ASU
Capacity
(GB)
TSC Price Data
Protection
Level
Date Results
Identifier
HP StorageWorks P9500 13,148 $88.34 129,112 $1,161,504 RAID-5 03/07/12 B00056
Sun ZFS Storage 7420 10,704 $35.24 31,884 $377,225 Mirroring 04/12/12 B00058
IBM DS8800 9,706 $270.38 71,537 $2,624,257 RAID-5 12/01/10 B00051
HP XP24000 8,725 $187.45 18,401 $1,635,434 Mirroring 09/08/08 B00035
Hitachi Storage Platform V 8,725 $187.49 18,401 $1,635,770 Mirroring 09/08/08 B00036
TMS RamSan-630 8,323 $49.37 8,117 $410,927 RAID-5 05/10/11 B00054
IBM XIV 7,468 $152.34 154,619 $1,137,641 RAID-1 10/19/11 BE00001
IBM DS8700 7,247 $277.22 32,642 $2,009,007 RAID-5 11/30/09 B00049
IBM SAN Vol Ctlr 4.2 7,084 $463.66 101,155 $3,284,767 RAID-5 07/12/07 B00024
Fujitsu ETERNUS DX440 S2 5,768 $66.50 42,133 $383,576 Mirroring 04/12/12 B00057
IBM DS5300 5,634 $74.13 16,383 $417,648 RAID-5 10/21/09 B00045
Sun Storage 6780 5,634 $47.03 16,383 $264,999 RAID-5 10/28/09 B00047
IBM DS5300 5,544 $75.33 14,043 $417,648 RAID-6 10/21/09 B00046
Sun Storage 6780 5,544 $47.80 14,043 $264,999 RAID-6 10/28/09 B00048
IBM DS5300 4,818 $93.80 16,383 $451,986 RAID-5 09/25/08 B00037
Sun Storage 6780 4,818 $53.61 16,383 $258,329 RAID-5 02/02/09 B00039
IBM DS5300 4,676 $96.67 14,043 $451,986 RAID-6 09/25/08 B00038
Sun Storage 6780 4,676 $55.25 14,043 $258,329 RAID-6 02/03/09 B00040
IBM SAN Vol Ctlr 4.1 4,544 $400.78 51,265 $1,821,301 RAID-5 09/12/06 B00011
IBM SAN Vol Ctlr 3.1 3,518 $563.93 20,616 $1,983,785 Mirroring 12/14/05 B00001
Fujitsu ETERNUS8000 1100 3,481 $238.93 4,570 $831,649 Mirroring 03/08/07 B00019
IBM DS8300 3,218 $539.38 15,393 $1,735,473 Mirroring 12/14/05 B00006
IBM Storwize V7000 3,133 $71.32 29,914 $223,422 RAID-5 12/13/10 B00052

SPC-2 MB/s = the Performance Metric
$/SPC-2 MB/s = the Price/Performance Metric
ASU Capacity = the Capacity Metric
Data Protection = Data Protection Metric
TSC Price = Total Cost of Ownership Metric
Results Identifier = A unique identification of the result Metric

Complete SPC-2 benchmark results may be found at http://www.storageperformance.org.

Configuration Summary

Storage Configuration:

Sun ZFS Storage 7420 appliance in clustered configuration
2 x Sun ZFS Storage 7420 controllers, each with
4 x 2.0 GHz Intel Xeon X7550 processors
512 GB memory, 64 x 8 GB 1066 MHz DDR3 DIMMs
16 x Sun Disk shelves, each with
24 x 300 GB 15K RPM SAS-2 drives

Server Configuration:

1 x Sun Fire X4470 server, with
4 x 2.4 GHz Intel Xeon E7-4870 processors
512 GB memory
8 x 8 Gb FC connections to the Sun ZFS Storage 7420 appliance
Oracle Solaris 11 11/11

2 x Sun Fire X4470 servers, each with
4 x 2.4 GHz Intel Xeon E7-4870 processors
256 GB memory
8 x 8 Gb FC connections to the Sun ZFS Storage 7420 appliance
Oracle Solaris 11 11/11

Benchmark Description

SPC Benchmark-2 (SPC-2): Consists of three distinct workloads designed to demonstrate the performance of a storage subsystem during the execution of business critical applications that require the large-scale, sequential movement of data. Those applications are characterized predominately by large I/Os organized into one or more concurrent sequential patterns. A description of each of the three SPC-2 workloads is listed below as well as examples of applications characterized by each workload.

  • Large File Processing: Applications in a wide range of fields, which require simple sequential process of one or more large files such as scientific computing and large-scale financial processing.
  • Large Database Queries: Applications that involve scans or joins of large relational tables, such as those performed for data mining or business intelligence.
  • Video on Demand: Applications that provide individualized video entertainment to a community of subscribers by drawing from a digital film library.

SPC-2 is built to:

  • Provide a level playing field for test sponsors.
  • Produce results that are powerful and yet simple to use.
  • Provide value for engineers as well as IT consumers and solution integrators.
  • Is easy to run, easy to audit/verify, and easy to use to report official results.

See Also

Disclosure Statement

SPC-2, SPC-2 MB/s, $/SPC-2 MB/s are registered trademarks of Storage Performance Council (SPC). Results as of April 12, 2012, for more information see www.storageperformance.org. Sun ZFS Storage 7420 Appliance http://www.storageperformance.org/results/benchmark_results_spc2#b00058; IBM DS8800 http://www.storageperformance.org/results/benchmark_results_spc2#b00051.

Thursday Mar 29, 2012

Sun Server X2-8 (formerly Sun Fire X4800 M2) Posts World Record x86 SPECjEnterprise2010 Result

Oracle's Sun Server X2-8 (formerly Sun Fire X4800 M2) using the Intel Xeon E7-8870 processor and Sun Server X2-4 using the Intel Xeon E7-4870 processor, produced a world record single application server SPECjEnterprise2010 benchmark result of 27,150.05 SPECjEnterprise2010 EjOPS. The Sun Server X2-8 ran the application tier and the Sun Server X2-4 was used for the database tier.

  • The Sun Server X2-8 demonstrated 63% better performance compared to IBM P780 server result of 16,646.34 SPECjEnterprise2010 EjOPS.

  • The Sun Server X2-8 demonstrated 4% better performance than the Cisco UCS B440 M2 result, both results used the same number of processors.

  • This result used Oracle WebLogic Server 12c, Java HotSpot(TM) 64-Bit Server 1.7.0_02, and Oracle Database 11g.

  • This result was produced using Oracle Linux.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results. The table below compares against the best results from IBM and Cisco.

SPECjEnterprise2010 Performance Chart
as of 7/11/2012
Submitter EjOPS* Application Server Database Server
Oracle 27,150.05 1x Sun Server X2-8
8x 2.4 GHz Intel Xeon E7-8870
Oracle WebLogic 12c
1x Sun Server X2-4
4x 2.4 GHz Intel Xeon E7-4870
Oracle Database 11g (11.2.0.2)
Cisco 26,118.67 2x UCS B440 M2 Blade Server
4x 2.4 GHz Intel Xeon E7-4870
Oracle WebLogic 11g (10.3.5)
1x UCS C460 M2 Blade Server
4x 2.4 GHz Intel Xeon E7-4870
Oracle Database 11g (11.2.0.2)
IBM 16,646.34 1x IBM Power 780
8x 3.86 GHz POWER 7
WebSphere Application Server V7
1x IBM Power 750 Express
4x 3.55 GHz POWER 7
IBM DB2 9.7 Workgroup Server Edition FP3a

* SPECjEnterprise2010 EjOPS, bigger is better.

Configuration Summary

Application Server:

1 x Sun Server X2-8

8 x 2.4 GHz Intel Xeon processor E7-8870
256 GB memory
4 x 10 GbE NIC
2 x FC HBA
Oracle Linux 5 Update 6
Oracle WebLogic Server Standard Edition Release 12.1.1
Java HotSpot(TM) 64-Bit Server VM on Linux, version 1.7.0_02 (Java SE 7 Update 2)

Database Server:

1 x Sun Server X2-4
4 x 2.4 GHz Intel Xeon E7-4870
512 GB memory
4 x 10 GbE NIC
2 x FC HBA
2 x Sun StorageTek 2540 M2
4 x Sun Fire X4270 M2
4 x Sun Storage F5100 Flash Array
Oracle Linux 5 Update 6
Oracle Database 11g Enterprise Edition Release 11.2.0.2

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The SPECjEnterprise2010 benchmark has been designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems.

The workload consists of an end to end web based order processing domain, an RMI and Web Services driven manufacturing domain and a supply chain model utilizing document based Web Services. The application is a collection of Java classes, Java Servlets, Java Server Pages, Enterprise Java Beans, Java Persistence Entities (pojo's) and Message Driven Beans.

The SPECjEnterprise2010 benchmark heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second ("SPECjEnterprise2010 EjOPS"). This metric is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is no price/performance metric in this benchmark.

Key Points and Best Practices

  • Sixteen Oracle WebLogic server instances were started using numactl, binding 2 instances per chip.
  • Eight Oracle database listener processes were started, binding 2 instances per chip using taskset.
  • Additional tuning information is in the report at http://spec.org.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Sun Server X2-8, 27,150.05 SPECjEnterprise2010 EjOPS; IBM Power 780, 16,646.34 SPECjEnterprise2010 EjOPS; Cisco UCS B440 M2, 26,118.67 SPECjEnterprise2010 EjOPS. Results from www.spec.org as of 7/11/2012.

Wednesday Nov 30, 2011

SPARC T4-4 Beats 8-CPU IBM POWER7 on TPC-H @3000GB Benchmark

Oracle's SPARC T4-4 server delivered a world record TPC-H @3000GB benchmark result for systems with four processors. This result beats eight processor results from IBM (POWER7) and HP (x86). The SPARC T4-4 server also delivered better performance per core than these eight processor systems from IBM and HP. Comparisons below are based upon system to system comparisons, highlighting Oracle's complete software and hardware solution.

This database world record result used Oracle's Sun Storage 2540-M2 arrays (rotating disk) connected to a SPARC T4-4 server running Oracle Solaris 11 and Oracle Database 11g Release 2 demonstrating the power of Oracle's integrated hardware and software solution.

  • The SPARC T4-4 server based configuration achieved a TPC-H scale factor 3000 world record for four processor systems of 205,792 QphH@3000GB with price/performance of $4.10/QphH@3000GB.

  • The SPARC T4-4 server with four SPARC T4 processors (total of 32 cores) is 7% faster than the IBM Power 780 server with eight POWER7 processors (total of 32 cores) on the TPC-H @3000GB benchmark.

  • The SPARC T4-4 server is 36% better in price performance compared to the IBM Power 780 server on the TPC-H @3000GB Benchmark.

  • The SPARC T4-4 server is 29% faster than the IBM Power 780 for data loading.

  • The SPARC T4-4 server is up to 3.4 times faster than the IBM Power 780 server for the Refresh Function.

  • The SPARC T4-4 server with four SPARC T4 processors is 27% faster than the HP ProLiant DL980 G7 server with eight x86 processors on the TPC-H @3000GB benchmark.

  • The SPARC T4-4 server is 52% faster than the HP ProLiant DL980 G7 server for data loading.

  • The SPARC T4-4 server is up to 3.2 times faster than the HP ProLiant DL980 G7 for the Refresh Function.

  • The SPARC T4-4 server achieved a peak IO rate from the Oracle database of 17 GB/sec. This rate was independent of the storage used, as demonstrated by the TPC-H @3000TB benchmark which used twelve Sun Storage 2540-M2 arrays (rotating disk) and the TPC-H @1000TB benchmark which used four Sun Storage F5100 Flash Array devices (flash storage). [*]

  • The SPARC T4-4 server showed linear scaling from TPC-H @1000GB to TPC-H @3000GB. This demonstrates that the SPARC T4-4 server can handle the increasingly larger databases required of DSS systems. [*]

  • The SPARC T4-4 server benchmark results demonstrate a complete solution of building Decision Support Systems including data loading, business questions and refreshing data. Each phase usually has a time constraint and the SPARC T4-4 server shows superior performance during each phase.

[*] The TPC believes that comparisons of results published with different scale factors are misleading and discourages such comparisons.

Performance Landscape

The table lists the leading TPC-H @3000GB results for non-clustered systems.

TPC-H @3000GB, Non-Clustered Systems
System
Processor
P/C/T – Memory
Composite
(QphH)
$/perf
($/QphH)
Power
(QppH)
Throughput
(QthH)
Database Available
SPARC Enterprise M9000
3.0 GHz SPARC64 VII+
64/256/256 – 1024 GB
386,478.3 $18.19 316,835.8 471,428.6 Oracle 11g R2 09/22/11
SPARC T4-4
3.0 GHz SPARC T4
4/32/256 – 1024 GB
205,792.0 $4.10 190,325.1 222,515.9 Oracle 11g R2 05/31/12
SPARC Enterprise M9000
2.88 GHz SPARC64 VII
32/128/256 – 512 GB
198,907.5 $15.27 182,350.7 216,967.7 Oracle 11g R2 12/09/10
IBM Power 780
4.1 GHz POWER7
8/32/128 – 1024 GB
192,001.1 $6.37 210,368.4 175,237.4 Sybase 15.4 11/30/11
HP ProLiant DL980 G7
2.27 GHz Intel Xeon X7560
8/64/128 – 512 GB
162,601.7 $2.68 185,297.7 142,685.6 SQL Server 2008 10/13/10

P/C/T = Processors, Cores, Threads
QphH = the Composite Metric (bigger is better)
$/QphH = the Price/Performance metric in USD (smaller is better)
QppH = the Power Numerical Quantity
QthH = the Throughput Numerical Quantity

The following table lists data load times and refresh function times during the power run.

TPC-H @3000GB, Non-Clustered Systems
Database Load & Database Refresh
System
Processor
Data Loading
(h:m:s)
T4
Advan
RF1
(sec)
T4
Advan
RF2
(sec)
T4
Advan
SPARC T4-4
3.0 GHz SPARC T4
04:08:29 1.0x 67.1 1.0x 39.5 1.0x
IBM Power 780
4.1 GHz POWER7
05:51:50 1.5x 147.3 2.2x 133.2 3.4x
HP ProLiant DL980 G7
2.27 GHz Intel Xeon X7560
08:35:17 2.1x 173.0 2.6x 126.3 3.2x

Data Loading = database load time
RF1 = power test first refresh transaction
RF2 = power test second refresh transaction
T4 Advan = the ratio of time to T4 time

Complete benchmark results found at the TPC benchmark website http://www.tpc.org.

Configuration Summary and Results

Hardware Configuration:

SPARC T4-4 server
4 x SPARC T4 3.0 GHz processors (total of 32 cores, 128 threads)
1024 GB memory
8 x internal SAS (8 x 300 GB) disk drives

External Storage:

12 x Sun Storage 2540-M2 array storage, each with
12 x 15K RPM 300 GB drives, 2 controllers, 2 GB cache

Software Configuration:

Oracle Solaris 11 11/11
Oracle Database 11g Release 2 Enterprise Edition

Audited Results:

Database Size: 3000 GB (Scale Factor 3000)
TPC-H Composite: 205,792.0 QphH@3000GB
Price/performance: $4.10/QphH@3000GB
Available: 05/31/2012
Total 3 year Cost: $843,656
TPC-H Power: 190,325.1
TPC-H Throughput: 222,515.9
Database Load Time: 4:08:29

Benchmark Description

The TPC-H benchmark is a performance benchmark established by the Transaction Processing Council (TPC) to demonstrate Data Warehousing/Decision Support Systems (DSS). TPC-H measurements are produced for customers to evaluate the performance of various DSS systems. These queries and updates are executed against a standard database under controlled conditions. Performance projections and comparisons between different TPC-H Database sizes (100GB, 300GB, 1000GB, 3000GB, 10000GB, 30000GB and 100000GB) are not allowed by the TPC.

TPC-H is a data warehousing-oriented, non-industry-specific benchmark that consists of a large number of complex queries typical of decision support applications. It also includes some insert and delete activity that is intended to simulate loading and purging data from a warehouse. TPC-H measures the combined performance of a particular database manager on a specific computer system.

The main performance metric reported by TPC-H is called the TPC-H Composite Query-per-Hour Performance Metric (QphH@SF, where SF is the number of GB of raw data, referred to as the scale factor). QphH@SF is intended to summarize the ability of the system to process queries in both single and multiple user modes. The benchmark requires reporting of price/performance, which is the ratio of the total HW/SW cost plus 3 years maintenance to the QphH. A secondary metric is the storage efficiency, which is the ratio of total configured disk space in GB to the scale factor.

Key Points and Best Practices

  • Twelve Sun Storage 2540-M2 arrays were used for the benchmark. Each Sun Storage 2540-M2 array contains 12 15K RPM drives and is connected to a single dual port 8Gb FC HBA using 2 ports. Each Sun Storage 2540-M2 array showed 1.5 GB/sec for sequential read operations and showed linear scaling, achieving 18 GB/sec with twelve Sun Storage 2540-M2 arrays. These were stand alone IO tests.

  • The peak IO rate measured from the Oracle database was 17 GB/sec.

  • Oracle Solaris 11 11/11 required very little system tuning.

  • Some vendors try to make the point that storage ratios are of customer concern. However, storage ratio size has more to do with disk layout and the increasing capacities of disks – so this is not an important metric in which to compare systems.

  • The SPARC T4-4 server and Oracle Solaris efficiently managed the system load of over one thousand Oracle Database parallel processes.

  • Six Sun Storage 2540-M2 arrays were mirrored to another six Sun Storage 2540-M2 arrays on which all of the Oracle database files were placed. IO performance was high and balanced across all the arrays.

  • The TPC-H Refresh Function (RF) simulates periodical refresh portion of Data Warehouse by adding new sales and deleting old sales data. Parallel DML (parallel insert and delete in this case) and database log performance are a key for this function and the SPARC T4-4 server outperformed both the IBM POWER7 server and HP ProLiant DL980 G7 server. (See the RF columns above.)

See Also

Disclosure Statement

TPC-H, QphH, $/QphH are trademarks of Transaction Processing Performance Council (TPC). For more information, see www.tpc.org. SPARC T4-4 205,792.0 QphH@3000GB, $4.10/QphH@3000GB, available 5/31/12, 4 processors, 32 cores, 256 threads; IBM Power 780 QphH@3000GB, 192,001.1 QphH@3000GB, $6.37/QphH@3000GB, available 11/30/11, 8 processors, 32 cores, 128 threads; HP ProLiant DL980 G7 162,601.7 QphH@3000GB, $2.68/QphH@3000GB available 10/13/10, 8 processors, 64 cores, 128 threads.

Monday Oct 03, 2011

SPARC T4-4 Servers Set World Record on SPECjEnterprise2010, Beats IBM POWER7, Cisco x86

Oracle produced a world record SPECjEnterprise2010 benchmark result of 40,104.86 SPECjEnterprise2010 EjOPS using four of Oracle's SPARC T4-4 servers in the application tier and two more SPARC T4-4 servers for the database server.

  • The four SPARC T4-4 server configuration (sixteen SPARC T4 processors total, 3.0 GHz) demonstrated 2.4x better performance compared to the IBM Power 780 server (eight POWER7 processors, 3.86 THz) result of 16,646.34 SPECjEnterprise2010 EjOPS.

  • In the database tier, two SPARC T4-4 servers with a total of eight SPARC T4 processors at 3.0 GHz, processed 2.4x more transactions compared to the IBM result of 16,646.34 SPECjEnterprise2010 EjOPS which used four POWER7 processors at 3.55 GHz.

  • The four SPARC T4-4 server configuration demonstrated 1.5x better performance compared to the Cisco UCS B440 M2 Blade Server result of 26,118.67 SPECjEnterprise2010 EjOPS.

  • The four SPARC T4-4 server configuration demonstrated 2.3x better performance compared to the Cisco UCS B440 M1 Blade Server result of 17,301.86 SPECjEnterprise2010 EjOPS.

  • This result demonstrated less than 1 second average response times for all SPECjEnterprise2010 transactions and 90% of all transaction times took less than 1 second.

  • This result demonstrated a sustained Java EE 5 transaction load generated by approximately 320,000 users.

  • This result using 16 Oracle WebLogic 10.3.5 server instances demonstrated 4.8x better performance per application server instance when compared to the IBM result which used 32 WebSphere instances.

  • The SPARC T4-4 servers delivered a 6.7x price/performance advantage over the IBM Power 780 for the servers used in the application tier (see disclosure statement below for details). This price/performance advantage in the application tier was accomplished with a SPARC T4-4 server configuration with 2 TB of total memory compared to the IBM solution with 0.5 TB of memory.

  • The SPARC T4-4 servers had a 1.9x advantage over IBM in performance per space for the application tier (see disclosure statement below for details) even though the Oracle solution had four servers.

  • The four SPARC T4-4 servers used for the application tier used Oracle Solaris Containers to consolidate four Oracle WebLogic application server instances on each server to achieve this result.

  • The two SPARC T4-4 servers used for the database tier hosted Oracle Database 11g Release 2 and Oracle RAC cluster software using Oracle Automatic Storage Management (ASM).

  • Oracle Fusion Middleware provides a family of complete, integrated, hot pluggable and best-of-breed products known for enabling enterprise customers to create and run agile and intelligent business applications. Oracle WebLogic Server's on-going, record-setting Java application server performance demonstrates why so many customers rely on Oracle Fusion Middleware as their foundation for innovation.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results.

SPECjEnterprise2010 Performance Chart
as of 10/11/2011
Submitter EjOPS* Java EE Server DB Server
Oracle 40,104.86 4 x SPARC T4-4
4 chips, 32 cores, 3.0 GHz SPARC T4
Oracle WebLogic 11g (10.3.5)
2 x SPARC T4-4
4 chips, 32 cores, 3.0 GHz SPARC T4
Oracle 11g DB 11.2.0.2
Cisco 26,118.67 2 x Cisco UCS B440 M2
4 chips, 40 cores, 2.4 GHz Xeon E7-4870
Oracle WebLogic 11g (10.3.5)
1 x Cisco UCS C460 M2
4 chips, 40 cores, 2.4 GHz Xeon E7-4870
Oracle 11g DB 11.2.0.2
Cisco 17,301.86 2 x Cisco UCS B440 M1
4 chips, 32 cores, 2.26 GHz Xeon X7560
Oracle WebLogic 10.3.4
1 x Cisco UCS C460 M1
4 chips, 32 cores, 2.26 GHz Xeon X7560
Oracle 11g DB 11.2.0.2
IBM 16,646.34 1 x IBM Power 780
8 chips, 64 cores, 3.86 GHz POWER7
WebSphere Application Server V7.0
1 x IBM Power 750 Express
4 chips, 32 cores, 3.55 GHz POWER7
IBM DB2 Universal Database 9.7

* SPECjEnterprise2010 EjOPS (bigger is better)

Configuration Summary

Application Servers:

4 x SPARC T4-4 servers, each with
4 x 3.0 GHz SPARC T4 processors
512 GB memory
2 x 10GbE NIC
Oracle Solaris 10 8/11
Oracle WebLogic Server 11g Release 1 (10.3.5)
Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.6.0_26 (Java SE 6 Update 26)

Database Servers:

2 x SPARC T4-4 servers, each with
4 x 3.0 GHz SPARC T4 processors
1024 GB memory
2 x 10GbE NIC
4 x 8Gb FC HBA
Oracle Solaris 10 8/11
Oracle Database 11g Enterprise Edition Release 11.2.0.2
Oracle Real Application Clusters 11g Release 2

Storage Servers:

8 x Sun Fire X4270 M2 (12-Drive)
1 x 3.0 GHz Intel Xeon
8 GB memory
1 x 8Gb FC HBA
Oracle Solaris 11 Express 2010.11
8 x Sun Storage F5100 Flash Arrays

Switch Hardware:

2 x Sun Network 10GbE 72-port Top of Rack (ToR) Switch
1 x Brocade 5300 80-port Fiber Channel Switch

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The new SPECjEnterprise2010 benchmark has been re-designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems,
  • The web container, servlets, and web services
  • The EJB container
  • JPA 1.0 Persistence Model
  • JMS and Message Driven Beans
  • Transaction management
  • Database connectivity
Moreover, SPECjEnterprise2010 also heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second (SPECjEnterprise2010 EjOPS). The primary metric for the SPECjEnterprise2010 benchmark is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is NO price/performance metric in this benchmark.

Key Points and Best Practices

  • Four Oracle WebLogic server instances on each SPARC T4-4 server were hosted in 4 separate Oracle Solaris Containers to demonstrate consolidation of multiple application servers.
  • Each Oracle Solaris Container was bound to a separate processor set, each contained 7 cores (total 56 threads). This was done to improve performance by reducing memory access latency by using the physical memory closest to the processors. The default set was used for network and disk interrupt handling.
  • The Oracle WebLogic application servers were executed in the FX scheduling class to improve performance by reducing the frequency of context switches.
  • The Oracle database processes were run in 2 processor sets using psrset(1M) and executed in the FX scheduling class. This improved performance by reducing memory access latency and reducing context switches.
  • The Oracle log writer process was run in a separate processor set containing 2 threads and run in the RT scheduling class. This insured that the log writer had the most efficient use of CPU resources.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Results from www.spec.org as of 10/11/2011. SPARC T4-4, 40,104.86 SPECjEnterprise2010 EjOPS; Cisco UCS B440 M2, 26,118.67 SPECjEnterprise2010 EjOPS; Cisco UCS B440 M1, 17,301.86 SPECjEnterprise2010 EjOPS; IBM Power 780, 16,646.34 SPECjEnterprise2010 EjOPS.

SPECjEnterprise2010 models contemporary Java-based applications that run on large Java EE (Java Enterprise Edition) servers, backed by network infrastructure and database servers. Focusing on the critical Java EE server hardware & OS, the IBM result includes a Java EE server with a list price of $1.30 million. The Oracle Java EE servers have a list price of $0.47 million. The Java EE server price versus delivered EjOPS is $77.97/EjOPS for IBM versus $11.67/EjOPS for Oracle. Oracle's $/perf advantage is 6.7x better than IBM ($77.97/$11.67).

Pricing details for IBM, IBM p780 512GB based on public pricing at http://tpc.org/results/FDR/TPCH/TPC-H_1TB_IBM780_Sybase-FDR.pdf. Adjusted hardware costs to license all 64 cores. AIX pricing at: http://www-304.ibm.com/easyaccess3/fileserve?contentid=214347 and AIX Standard Edition V7.1 per processor (5765-G98-0017 64*2,600=$166,400). This gives application tier hardware & OS Price/perf: $77.97/EjOPS (1297956/16646.34)

Pricing details for Oracle, four SPARC T4-4 512 GB, HW acquisition price from Oracle's price list: $467,856 http://www.oracle.com. This gives application tier hardware & OS Price/perf: $11.67/EjOPS (467856/40104.86)

The Oracle application tier servers occupy 20U of space, 40,140.86/20=2005 EjOPS/U. The IBM application tier server occupies 16U of space, 16,646.34/16=1040 EjOPS/U. 2005/1040=1.9x

SPARC T4-4 Beats IBM POWER7 and HP Itanium on TPC-H @1000GB Benchmark

Oracle's SPARC T4-4 server configured with SPARC-T4 processors, Oracle's Sun Storage F5100 Flash Array storage, Oracle Solaris, and Oracle Database 11g Release 2 achieved a TPC-H benchmark performance result of 201,487 QphH@1000GB with price/performance of $4.60/QphH@1000GB.

  • The SPARC T4-4 server benchmark results demonstrate a complete solution of building Decision Support Systems including data loading, business questions and refreshing data. Each phase usually has a time constraint and the SPARC T4-4 server shows superior performance during each phase.

  • The SPARC T4-4 server is 22% faster than the 8-socket IBM POWER7 server with the same number of cores. The SPARC T4-4 server has over twice the performance per socket compared to the IBM POWER7 server.

  • The SPARC T4-4 server achieves 33% better price/performance than the IBM POWER7 server.

  • The SPARC T4-4 server is up to 4 times faster than the IBM POWER7 server for the Refresh Function.

  • The SPARC T4-4 server is 44% faster than the HP Superdome 2 server. The SPARC T4-4 server has 5.7x the performance per socket of the HP Superdome 2 server.

  • The SPARC T4-4 server is 62% better on price/performance than the HP Itanium server.

  • The SPARC T4-4 server is up to 3.7 times faster than the HP Itanium server for the Refresh Function.

  • The SPARC T4-4 server delivers nearly the same performance as Oracle's SPARC Enterprise M8000 server, but with 52% better price/performance on the TPC-H @1000GB benchmark.

  • Oracle used Storage Redundancy Level 3 as defined by the TPC-H 2.14.2 specification which is the strictest level.

  • This TPC-H result demonstrates that the SPARC T4-4 server can deliver the performance while running the increasingly larger databases required of DSS systems. The server measured more than 16 GB/sec of IO throughput through Oracle Database 11g Release 2 software while maintaining the high cpu load.

Performance Landscape

The table below lists published non-cluster results from comparable enterprise class systems from Oracle, IBM and HP. Each system was configured with 512 GB of memory.

TPC-H @1000GB

System
CPU type
Proc/Core/Thread
Composite
(QphH)
$/perf
($/QphH)
Power
(QppH)
Throughput
(QthH)
Database Available
SPARC Enterprise M8000
3 GHz SPARC64 VII+
16 / 64 / 128
209,533.6 $9.53 177,845.9 246,867.2 Oracle 11g 09/22/11
SPARC T4-4
3 GHz SPARC-T4
4 / 32 / 256
201,487.0 $4.60 181,760.6 223,354.2 Oracle 11g 10/30/11
IBM Power 780
4.14 GHz POWER7
8 / 32 / 128
164,747.2 $6.85 170,206.4 159,463.1 Sybase 03/31/11
HP Superdome 2
1.73 GHz Intel Itanium 9350
16 / 64 / 64
140,181.1 $12.15 139,181.0 141,188.3 Oracle 11g 10/20/10

QphH = the Composite Metric (bigger is better)
$/QphH = the Price/Performance metric (smaller is better)
QppH = the Power Numerical Quantity
QthH = the Throughput Numerical Quantity

Complete benchmark results found at the TPC benchmark website http://www.tpc.org.

Configuration Summary and Results

Hardware Configuration:

SPARC T4-4 server
4 x SPARC-T4 3.0 GHz processors (total of 32 cores, 128 threads)
512 GB memory
8 x internal SAS (8 x 300 GB) disk drives

External Storage:

4 x Sun Storage F5100 Flash Array storage, each with
80 x 24 GB Flash Modules

Software Configuration:

Oracle Solaris 10 8/11
Oracle Database 11g Release 2 Enterprise Edition

Audited Results:

Database Size: 1000 GB (Scale Factor 1000)
TPC-H Composite: 201,487 QphH@1000GB
Price/performance: $4.60/QphH@1000GB
Available: 10/30/2011
Total 3 Year Cost: $925,525
TPC-H Power: 181,760.6
TPC-H Throughput: 223,354.2
Database Load Time: 1:22:39

Benchmark Description

The TPC-H benchmark is a performance benchmark established by the Transaction Processing Council (TPC) to demonstrate Data Warehousing/Decision Support Systems (DSS). TPC-H measurements are produced for customers to evaluate the performance of various DSS systems. These queries and updates are executed against a standard database under controlled conditions. Performance projections and comparisons between different TPC-H Database sizes (100GB, 300GB, 1000GB, 3000GB and 10000GB) are not allowed by the TPC.

TPC-H is a data warehousing-oriented, non-industry-specific benchmark that consists of a large number of complex queries typical of decision support applications. It also includes some insert and delete activity that is intended to simulate loading and purging data from a warehouse. TPC-H measures the combined performance of a particular database manager on a specific computer system.

The main performance metric reported by TPC-H is called the TPC-H Composite Query-per-Hour Performance Metric (QphH@SF, where SF is the number of GB of raw data, referred to as the scale factor). QphH@SF is intended to summarize the ability of the system to process queries in both single and multi user modes. The benchmark requires reporting of price/performance, which is the ratio of QphH to total HW/SW cost plus 3 years maintenance.

Key Points and Best Practices

  • Four Sun Storage F5100 Flash Array devices were used for the benchmark. Each F5100 device contains 80 flash modules (FMODs). Twenty (20) FMODs from each F5100 device were connected to a single SAS 6 Gb HBA. A single F5100 device showed 4.16 GB/sec for sequential read and demonstrated linear scaling of 16.62 GB/sec with 4 x F5100 devices.

  • The IO rate from the Oracle database was over 16 GB/sec.

  • Oracle Solaris 10 8/11 required very little system tuning.

  • The SPARC T4-4 server and Oracle Solaris efficiently managed the system load of over one thousand Oracle parallel processes.

  • The Oracle database files for tables and indexes were managed by Oracle Automatic Storage Manager (ASM) with 4M stripe. Two F5100 devices were mirrored to another 2 F5100 devices under ASM. IO performance was high and balanced across all the FMODs.
  • The Oracle redo log files were mirrored across the F5100 devices using Oracle Solaris Volume Manager with 128K stripe.
  • Parallel degree on tables and indexes was set to 128. This setting worked the best for performance.
  • TPC-H Refresh Function simulates periodical Refresh portion of Data Warehouse by adding new sales and deleting old sales data. Parallel DML (parallel insert and delete in this case) and database log performance are a key for this function and the SPARC T4-4 server outperformed both HP Superdome 2 and IBM POWER7 servers.

See Also

Disclosure Statement

TPC-H, QphH, $/QphH are trademarks of Transaction Processing Performance Council (TPC). For more information, see www.tpc.org. SPARC T4-4 201,487 QphH@1000GB, $4.60/QphH@1000GB, avail 10/30/2011, 4 processors, 32 cores, 256 threads; SPARC Enterprise M8000 209,533.6 QphH@1000GB, $9.53/QphH@1000GB, avail 09/22/11, 16 processors, 64 cores, 128 threads; IBM Power 780 QphH@1000GB, 164,747.2 QphH@1000GB, $6.85/QphH@1000GB, avail 03/31/11, 8 processors, 32 cores, 128 threads; HP Integrity Superdome 2 140,181.1 QphH@1000GB, $12.15/QphH@1000GB avail 10/20/10, 16 processors, 64, cores, 64 threads.

Thursday Sep 29, 2011

SPARC T4 Processor Outperforms IBM POWER7 and Intel (Westmere AES-NI) on OpenSSL AES Encryption Test

Oracle's SPARC T4 processor is faster than the Intel Xeon X5690 (with AES-NI) and the IBM POWER7.

  • On single-thread OpenSSL encryption, the 2.85 GHz SPARC T4 processor is 4.3 times faster than the 3.5 GHz IBM POWER7 processor.

  • On single-thread OpenSSL encryption, the 2.85 GHz SPARC T4 processor is 17% faster than the 3.46 GHz Intel Xeon X5690 processor.

The SPARC T4 processor has Encryption Instruction Accelerators for encryption and decryption for AES and many other ciphers. The Intel Xeon X5690 processor has AES-NI instructions which accelerate only AES ciphers. The IBM POWER7 does not have cryptographic instructions, but cryptographic coprocessors are available.

Performance Landscape

The table below shows results when running the OpenSSL speed command with the AES-256-CBC cipher. The reported results are for a message size of 8192 bytes. Results are reported for a single thread and for running on all available hardware threads (no over subscribing).

OpenSSL Performance with
AES-256-CBC Encryption
Processor Performance (MB/sec)
1 Thread Maximum Throughput
(at number of threads)
SPARC T4, 2.85 GHz 769 11,967 (64)
Intel Xeon X5690, 3.46 GHz 660 7,362 (12)
IBM POWER7, 3.5 GHz 179 2,860 (est*)

(est*) The performance of the IBM POWER7 is estimated at 16 times the rate of the single thread performance. The estimate is considered an upper bound on expected performance for this processor.

Configuration Summary

SPARC Configuration:

SPARC T4-1 server
1 x SPARC T4 processors, 2.85 GHz
64 GB memory
Oracle Solaris 11

Intel Configuration:

Sun Fire X4270 M2 server
1 x Intel Xeon X5690 processors, 3.46 GHz
24 GB memory
Oracle Solaris 11

Software Configuration:

OpenSSL 1.0.0.d
gcc 3.4.3

Benchmark Description

The in-memory SSL performance was measured with the openssl command. openssl has an option for measuring the speed of various ciphers and message sizes. The actual command used to measure the speed of AES-256-CBC was:

openssl speed -multi {number of threads} -evp aes-256-cbc

openssl runs for several minutes and measures the speed, in units of MB/sec, of the specified cipher for messages of sizes 16 bytes to 8192 bytes.

Key Points and Best Practices

  • The Encryption Instruction Accelerators are accessed through a platform independent API for cryptographic engines.
  • The OpenSSL libraries use the API. The default is to not use the Encryption Instruction Accelerators.
  • Cryptography is compute intensive. Using all available threads streams, both the SPARC T4 processor and the Intel Xeon processor were able to saturate the memory bandwidth of the respective systems.

See Also

Disclosure Statement

Copyright 2011, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 9/26/2011.

About

BestPerf is the source of Oracle performance expertise. In this blog, Oracle's Strategic Applications Engineering group explores Oracle's performance results and shares best practices learned from working on Enterprise-wide Applications.

Index Pages
Search

Archives
« May 2016
SunMonTueWedThuFriSat
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
    
       
Today