Monday Oct 26, 2015

SAP Two-Tier Standard Sales and Distribution SD Benchmark: SPARC T7-2 World Record 2 Processors

Oracle's SPARC T7-2 server produces a world record result for 2-processors on the SAP two-tier Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement Package 5 for SAP ERP 6.0 (2 chips / 64 cores / 512 threads).

  • The SPARC T7-2 server achieved 30,800 SAP SD benchmark users running the two-tier SAP Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement Package 5 for SAP ERP 6.0.

  • The SPARC T7-2 server achieved 1.9 times more users than the Dell PowerEdge R730 server result.

  • The SPARC T7-2 server achieved 1.5 times more users than the IBM Power System S824 server result.

  • The SPARC T7-2 server achieved 1.9 times more users than the HP ProLiant DL380 Gen9 server result.

  • The SPARC T7-2 server result was run with Oracle Solaris 11 and used Oracle Database 12c.

Performance Landscape

SAP-SD 2-tier performance table in decreasing performance order for leading two-processor systems and four-processor IBM Power System S824 server, with SAP ERP 6.0 Enhancement Package 5 for SAP ERP 6.0 results (current version of the benchmark as of May, 2012).

SAP SD Two-Tier Benchmark
System
Processor
OS
Database
Users Resp Time
(sec)
Version Cert#
SPARC T7-2
2 x SPARC M7 (2x 32core)
Oracle Solaris 11
Oracle Database 12c
30,800 0.96 EHP5 2015050
IBM Power S824
4 x POWER8 (4x 6core)
AIX 7
DB2 10.5
21,212 0.98 EHP5 2014016
Dell PowerEdge R730
2 x Intel E5-2699 v3 (2x 18core)
Red Hat Enterprise Linux 7
SAP ASE 16
16,500 0.99 EHP5 2014033
HP ProLiant DL380 Gen9
2 x Intel E5-2699 v3 (2x 18core)
Red Hat Enterprise Linux 6.5
SAP ASE 16
16,101 0.99 EHP5 2014032

Version – Version of SAP, EHP5 refers to SAP ERP 6.0 Enhancement Package 5 for SAP ERP 6.0

Number of cores presented are per chip, to get system totals, multiple by the number of chips.

Complete benchmark results may be found at the SAP benchmark website http://www.sap.com/benchmark.

Configuration Summary and Results

Database/Application Server:

1 x SPARC T7-2 server with
2 x SPARC M7 processors (4.13 GHz, total of 2 processors / 64 cores / 512 threads)
1 TB memory
Oracle Solaris 11.3
Oracle Database 12c

Database Storage:
3 x Sun Server X3-2L each with
2 x Intel Xeon Processors E5-2609 (2.4 GHz)
16 GB memory
4 x Sun Flash Accelerator F40 PCIe Card
12 x 3 TB SAS disks
Oracle Solaris 11

REDO log Storage:
1 x Pillar FS-1 Flash Storage System, with
2 x FS1-2 Controller (Netra X3-2)
2 x FS1-2 Pilot (X4-2)
4 x DE2-24P Disk enclosure
96 x 300 GB 10000 RPM SAS Disk Drive Assembly

Certified Results (published by SAP)

Number of SAP SD benchmark users: 30,800
Average dialog response time: 0.96 seconds
Throughput:
  Fully processed order line items per hour: 3,372,000
  Dialog steps per hour: 10,116,000
  SAPS: 168,600
Average database request time (dialog/update): 0.022 sec / 0.047 sec
SAP Certification: 2015050

Benchmark Description

The SAP Standard Application SD (Sales and Distribution) Benchmark is an ERP business test that is indicative of full business workloads of complete order processing and invoice processing, and demonstrates the ability to run both the application and database software on a single system. The SAP Standard Application SD Benchmark represents the critical tasks performed in real-world ERP business environments.

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

Two-tier SAP Sales and Distribution (SD) standard application benchmarks, SAP Enhancement Package 5 for SAP ERP 6.0 as of 10/23/15:

SPARC T7-2 (2 processors, 64 cores, 512 threads) 30,800 SAP SD users, 2 x 4.13 GHz SPARC M7, 1 TB memory, Oracle Database 12c, Oracle Solaris 11, Cert# 2015050.
IBM Power System S824 (4 processors, 24 cores, 192 threads) 21,212 SAP SD users, 4 x 3.52 GHz POWER8, 512 GB memory, DB2 10.5, AIX 7, Cert#2014016
Dell PowerEdge R730 (2 processors, 36 cores, 72 threads) 16,500 SAP SD users, 2 x 2.3 GHz Intel Xeon Processor E5-2699 v3 256 GB memory, SAP ASE 16, RHEL 7, Cert#2014033
HP ProLiant DL380 Gen9 (2 processors, 36 cores, 72 threads) 16,101 SAP SD users, 2 x 2.3 GHz Intel Xeon Processor E5-2699 v3 256 GB memory, SAP ASE 16, RHEL 6.5, Cert#2014032

SAP, R/3, reg TM of SAP AG in Germany and other countries. More info www.sap.com/benchmark

SPARC T7-1 Delivers 1-Chip World Records for SPEC CPU2006 Rate Benchmarks

This page has been updated on November 19, 2015. The SPARC T7-1 server results have been published at www.spec.org.

Oracle's SPARC T7-1 server delivered world record SPEC CPU2006 rate benchmark results for systems with one chip. This was accomplished with Oracle Solaris 11.3 and Oracle Solaris Studio 12.4 software.

  • The SPARC T7-1 server achieved world record scores of 1200 SPECint_rate2006, 1120 SPECint_rate_base2006, 832 SPECfp_rate2006, and 801 SPECfp_rate_base2006.

  • The SPARC T7-1 server beat the 1 chip Fujitsu CELSIUS C740 with an Intel Xeon Processor E5-2699 v3 by 1.7x on the SPECint_rate2006 benchmark. The SPARC T7-1 server beat the 1 chip NEC Express5800/R120f-1M with an Intel Xeon Processor E5-2699 v3 by 1.8x on the SPECfp_rate2006 benchmark.

  • The SPARC T7-1 server beat the 1 chip IBM Power S812LC server with a POWER8 processor by 1.9 times on the SPECint_rate2006 benchmark and by 1.8 times on the SPECfp_rate2006 benchmark.

  • The SPARC T7-1 server beat the 1 chip Fujitsu SPARC M10-4S with a SPARC64 X+ processor by 2.2x on the SPECint_rate2006 benchmark and by 1.6x on the SPECfp_rate2006 benchmark.

  • The SPARC T7-1 server improved upon the previous generation SPARC platform which used the SPARC T5 processor by 2.5 on the SPECint_rate2006 benchmark and by 2.3 on the SPECfp_rate2006 benchmark.

The SPEC CPU2006 benchmarks are derived from the compute-intensive portions of real applications, stressing chip, memory hierarchy, and compilers. The benchmarks are not intended to stress other computer components such as networking, the operating system, or the I/O system. Note that there are many other SPEC benchmarks, including benchmarks that specifically focus on Java computing, enterprise computing, and network file systems.

Performance Landscape

Complete benchmark results are at the SPEC website. The tables below provide the new Oracle results, as well as select results from other vendors.

Presented are single chip SPEC CPU2006 rate results. Only the best results published at www.spec.org per chip type are presented (best Intel, IBM, Fujitsu, Oracle chips).

SPEC CPU2006 Rate Results – One Chip
System Chip Peak Base
  SPECint_rate2006
SPARC T7-1 SPARC M7 (4.13 GHz, 32 cores) 1200 1120
Fujitsu CELSIUS C740 Intel E5-2699 v3 (2.3 GHz, 18 cores) 715 693
IBM Power S812LC POWER8 (2.92 GHz, 10 cores) 642 482
Fujitsu SPARC M10-4S SPARC64 X+ (3.7 GHz, 16 cores) 546 479
SPARC T5-1B SPARC T5 (3.6 GHz, 16 cores) 489 441
IBM Power 710 Express POWER7 (3.55 GHz, 8 cores) 289 255
  SPECfp_rate2006
SPARC T7-1 SPARC M7 (4.13 GHz, 32 cores) 832 801
NEC Express5800/R120f-1M Intel E5-2699 v3 (2.3 GHz, 18 cores) 474 460
IBM Power S812LC POWER8 (2.92 GHz, 10 cores) 468 394
Fujitsu SPARC M10-4S SPARC64 X+ (3.7 GHz, 16 cores) 462 418
SPARC T5-1B SPARC T5 (3.6 GHz, 16 cores) 369 350
IBM Power 710 Express POWER7 (3.55 GHz, 8 cores) 248 229

The following table highlights the performance of the single-chip SPARC M7 processor based server to the best published two-chip POWER8 processor based server.

SPEC CPU2006 Rate Results
Comparing One SPARC M7 Chip to Two POWER8 Chips
System Chip Peak Base
  SPECint_rate2006
SPARC T7-1 1 x SPARC M7 (4.13 GHz, 32core) 1200 1120
IBM Power S822LC 2 x POWER8 (2.92 GHz, 2x 10core) 1100 853
  SPECfp_rate2006
SPARC T7-1 1 x SPARC M7 (4.13 GHz, 32 cores) 832 801
IBM Power S822LC 2 x POWER8 (2.92 GHz, 2x 10core) 888 745

Configuration Summary

System Under Test:

SPARC T7-1
1 x SPARC M7 processor (4.13 GHz)
512 GB memory (16 x 32 GB dimms)
800 GB on 4 x 400 GB SAS SSD (mirrored)
Oracle Solaris 11.3
Oracle Solaris Studio 12.4 with 4/15 Patch Set

Benchmark Description

SPEC CPU2006 is SPEC's most popular benchmark. It measures:

  • Speed — single copy performance of chip, memory, compiler
  • Rate — multiple copy (throughput)

The benchmark is also divided into integer intensive applications and floating point intensive applications:

  • integer: 12 benchmarks derived from applications such as artificial intelligence chess playing, artificial intelligence go playing, quantum computer simulation, perl, gcc, XML processing, and pathfinding
  • floating point: 17 benchmarks derived from applications, including chemistry, physics, genetics, and weather.

It is also divided depending upon the amount of optimization allowed:

  • base: optimization is consistent per compiled language, all benchmarks must be compiled with the same flags per language.
  • peak: specific compiler optimization is allowed per application.

The overall metrics for the benchmark which are commonly used are:

  • SPECint_rate2006, SPECint_rate_base2006: integer, rate
  • SPECfp_rate2006, SPECfp_rate_base2006: floating point, rate
  • SPECint2006, SPECint_base2006: integer, speed
  • SPECfp2006, SPECfp_base2006: floating point, speed

Key Points and Best Practices

  • Jobs were bound using pbind.

See Also

Disclosure Statement

SPEC and the benchmark names SPECfp and SPECint are registered trademarks of the Standard Performance Evaluation Corporation. Results as of November 19, 2015 from www.spec.org.
SPARC T7-1: 1200 SPECint_rate2006, 1120 SPECint_rate_base2006, 832 SPECfp_rate2006, 801 SPECfp_rate_base2006; SPARC T5-1B: 489 SPECint_rate2006, 440 SPECint_rate_base2006, 369 SPECfp_rate2006, 350 SPECfp_rate_base2006; Fujitsu SPARC M10-4S: 546 SPECint_rate2006, 479 SPECint_rate_base2006, 462 SPECfp_rate2006, 418 SPECfp_rate_base2006. IBM Power 710 Express: 289 SPECint_rate2006, 255 SPECint_rate_base2006, 248 SPECfp_rate2006, 229 SPECfp_rate_base2006; Fujitsu CELSIUS C740: 715 SPECint_rate2006, 693 SPECint_rate_base2006; NEC Express5800/R120f-1M: 474 SPECfp_rate2006, 460 SPECfp_rate_base2006; IBM Power S822LC: 1100 SPECint_rate2006, 853 SPECint_rate_base2006, 888 SPECfp_rate2006, 745 SPECfp_rate_base2006; IBM Power S812LC: 642 SPECint_rate2006, 482 SPECint_rate_base2006, 468 SPECfp_rate2006, 394 SPECfp_rate_base2006.

SPARC T7-4 Delivers 4-Chip World Record for SPEC OMP2012

Oracle's SPARC T7-4 server delivered world record performance on the SPEC OMP2012 benchmark for systems with four chips. This was accomplished with Oracle Solaris 11.3 and Oracle Solaris Studio 12.4 software.

  • The SPARC T7-4 server delivered world record for systems with four chips of 27.9 SPECompG_peak2012 and 26.4 SPECompG_base2012.

  • The SPARC T7-4 server beat the four chip HP ProLiant DL580 Gen9 with Intel Xeon Processor E7-8890 v3 by 29% on the SPECompG_peak2012 benchmark.

  • This SPEC OMP2012 benchmark result demonstrates that the SPARC M7 processor performs well on floating-point intensive technical computing and modeling workloads.

Performance Landscape

Complete benchmark results are at the SPEC website, SPEC OMP2012 Results. The table below provides the new Oracle result as well as the previous best four chip results.

SPEC OMP2012 Results
Four Chip Results
System Processor Peak Base
SPARC T7-4 SPARC M7, 4.13 GHz 27.9 26.4
HP ProLiant DL580 Gen9 Intel Xeon E7-8890 v3, 2.5 GHz 21.5 20.4
Cisco UCS C460 M4 Intel Xeon E7-8890 v3, 2.5 GHz -- 20.8

Configuration Summary

Systems Under Test:

SPARC T7-4
4 x 4.13 GHz SPARC M7 processors
2 TB memory (64 x 32 GB dimms)
4 x 600 GB SAS 10,000 RPM HDD (mirrored)
Oracle Solaris 11.3 (11.3.0.30.0)
Oracle Solaris Studio 12.4 with 4/15 Patch Set

Benchmark Description

The following was taken from the SPEC website.

SPEC OMP2012 focuses on compute intensive performance, which means these benchmarks emphasize the performance of:

  • the computer processor (CPU),
  • the memory architecture,
  • the parallel support libraries, and
  • the compilers.

It is important to remember the contribution of the latter three components. SPEC OMP performance intentionally depends on more than just the processor.

SPEC OMP2012 contains a suite that focuses on parallel computing performance using the OpenMP parallelism standard.

The suite can be used to measure along the following vector:

  • Compilation method: Consistent compiler options across all programs of a given language (the base metrics) and, optionally, compiler options tuned to each program (the peak metrics).

SPEC OMP2012 is not intended to stress other computer components such as networking, the operating system, graphics, or the I/O system. Note that there are many other SPEC benchmarks, including benchmarks that specifically focus on graphics, distributed Java computing, webservers, and network file systems.

Key Points and Best Practices

  • Jobs were bound using OMP_PLACES.

See Also

Disclosure Statement

SPEC and the benchmark name SPEC OMP are registered trademarks of the Standard Performance Evaluation Corporation. Results as of November 11, 2015 from www.spec.org. SPARC T7-4 (4 chips, 128 cores, 1024 threads): 27.9 SPECompG_peak2012, 26.4 SPECompG_base2012; HP ProLiant DL580 Gen9 (4 chips, 72 cores, 144 threads): 21.5 SPECompG_peak2012, 20.4 SPECompG_base2012; Cisco UCS C460 M7 (4 chips, 72 cores, 144 threads): 20.8 SPECompG_base2012.

Friday Mar 20, 2015

Oracle ZFS Storage ZS4-4 Shows 1.8x Generational Performance Improvement on SPC-2 Benchmark

The Oracle ZFS Storage ZS4-4 appliance delivered 1.8x improved performance and 1.3x improved price performance over the previous generation Oracle ZFS Storage ZS3-4 appliance as shown by the SPC-2 benchmark.

  • Running the SPC-2 benchmark, the Oracle ZFS Storage ZS4-4 appliance delivered SPC-2 Price-Performance of $17.09 and an overall score of 31,486.23 SPC-2 MBPS.

  • The Oracle ZFS Storage continues its strong price performance by occupying the three of the top five SPC-2 price performance.

  • Oracle holds the three of the top four performance results on the SPC-2 benchmark for HDD based systems.

  • The Oracle ZFS Storage ZS4-4 appliance has a 7.6x price-performance advantage over the IBM DS8870 and 2x performance advantage as measured by the SPC-2 benchmark.

  • The Oracle ZFS Storage ZS4-4 appliance has a 5.0x performance advantage over the new Fujitsu DX200 S3 as measured by the SPC-2 benchmark.

  • The Oracle ZFS Storage ZS4-4 appliance has a 4.6x price-performance advantage over the Fujitsu ET8700 S2 and 1.9x performance advantage as shown by the SPC-2 benchmark.

  • The Oracle ZFS Storage ZS4-4 appliance has a 4.6x price-performance advantage over the Hitachi Virtual Storage Platform (VSP) and 1.96x performance advantage as measured by the SPC-2 benchmark.

  • The Oracle ZFS Storage ZS4-4 appliance has a 1.6x price-performance advantage over the HP XP7 disk array as shown by the SPC-2 benchmark (HP even discounted their hardware 63%).

Performance Landscape

SPC-2 Price-Performance

Below is a table of the top SPC-2 Price-Performance results for HDD storage based systems, presented in increasing price-performance order (as of 03/17/2015). The complete set of results may be found at SPC2 top 10 Price-Performance list.

System SPC-2
MBPS
$/SPC-2
MBPS
Results
Identifier
Oracle ZFS Storage ZS3-2 16,212.66 $12.08 BE00002
Fujitsu Eternus DX200 S3 6,266.50 $15.42 B00071
SGI InfiniteStorage 5600 8,855.70 $15.97 B00065
Oracle ZFS Storage ZS4-4 31,486.23 $17.09 B00072
Oracle ZFS Storage ZS3-4 17,244.22 $22.53 B00067
NEC Storage M700 14,408.89 $25.10 B00066
Sun StorageTek 2530 663.51 $26.48 B00026
HP XP7 storage 43,012.53 $28.30 B00070
Fujitsu ETERNUS DX80 S2 2,685.50 $28.48 B00055
SGI InfiniteStorage 5500-SP 4,064.49 $28.57 B00059
Hitachi Unified Storage VM 11,274.83 $32.64 B00069

SPC-2 MBPS = the Performance Metric
$/SPC-2 MBPS = the Price-Performance Metric
Results Identifier = A unique identification of the result

SPC-2 Performance

The following table list the top SPC-2 -Performance results for HDD storage based systems, presented in decreasing performance order (as of 03/17/2015). The complete set of results may be found at the SPC2 top 10 Performance list.

HDD Based Systems SPC-2
MBPS
$/SPC-2
MBPS
TSC Price Results
Identifier
HP XP7 storage 43,012.52 $28.30 $1,217,462 B00070
Oracle ZFS Storage ZS4-4 31,486.23 $17.09 $538,050 B00072
Oracle ZFS Storage ZS3-4 17,244.22 $22.53 $388,472 B00067
Oracle ZFS Storage ZS3-2 16,212.66 $12.08 $195,915 BE00002
Fujitsu ETERNUS DX8870 S2 16,038.74 $79.51 $1,275,163 B00063
IBM System Storage DS8870 15,423.66 $131.21 $2,023,742 B00062
IBM SAN VC v6.4 14,581.03 $129.14 $1,883,037 B00061
Hitachi Virtual Storage Platform (VSP) 13,147.87 $95.38 $1,254,093 B00060
HP StorageWorks P9500 XP Storage Array 13,147.87 $88.34 $1,161,504 B00056

SPC-2 MBPS = the Performance Metric
$/SPC-2 MBPS = the Price-Performance Metric
TSC Price = Total Cost of Ownership Metric
Results Identifier = A unique identification of the result Metric

Complete SPC-2 benchmark results may be found at
http://www.storageperformance.org/results/benchmark_results_spc2.

Configuration Summary

Storage Configuration:

Oracle ZFS Storage ZS4-4 storage system in clustered configuration
2 x Oracle ZFS Storage ZS4-4 controllers with
8 x Intel Xeon processors
3 TB memory
24 x Oracle Storage Drive Enclosure DE2-24P, each with
24 x 300 GB 10K RPM SAS-2 drives

Benchmark Description

SPC Benchmark 2 (SPC-2): Consists of three distinct workloads designed to demonstrate the performance of a storage subsystem during the execution of business critical applications that require the large-scale, sequential movement of data. Those applications are characterized predominately by large I/Os organized into one or more concurrent sequential patterns. A description of each of the three SPC-2 workloads is listed below as well as examples of applications characterized by each workload.

  • Large File Processing: Applications in a wide range of fields, which require simple sequential process of one or more large files such as scientific computing and large-scale financial processing.
  • Large Database Queries: Applications that involve scans or joins of large relational tables, such as those performed for data mining or business intelligence.
  • Video on Demand: Applications that provide individualized video entertainment to a community of subscribers by drawing from a digital film library.

SPC-2 is built to:

  • Provide a level playing field for test sponsors.
  • Produce results that are powerful and yet simple to use.
  • Provide value for engineers as well as IT consumers and solution integrators.
  • Is easy to run, easy to audit/verify, and easy to use to report official results.

See Also

Disclosure Statement

SPC-2 and SPC-2 MBPS are registered trademarks of Storage Performance Council (SPC). Results as of March 17, 2015, for more information see www.storageperformance.org.

Oracle ZFS Storage ZS4-4 - B00072, Oracle ZFS Storage ZS3-2 - BE00002, Oracle ZFS Storage ZS3-4 - B00067, Fujitsu ETERNUS DX80 S2, B00055, Fujitsu ETERNUS DX8870 S2 - B00063, Fujitsu ETERNUS DX200 S3 - B00071, HP StorageWorks P9500 XP Storage Array - B00056, HP XP7 Storage Array - B00070, Hitachi Unified Storage VM - B00069, Hitachi Virtual Storage Platform (VSP) - B00060, IBM SAN VC v6.4 - B00061, IBM System Storage DS8870 - B00062, IBM XIV Storage System Gen3 - BE00001, NEC Storage M700 - B00066, SGI InfiniteStorage 5500-SP - B00059, SGI InfiniteStorage 5600 - B00065, Sun StorageTek 2530 - B00026.

Wednesday Jun 25, 2014

Oracle ZFS Storage ZS3-2 Delivers World Record Price-Performance on SPC-2/E

The Oracle ZFS Storage ZS3-2 appliance delivered a world record Price-Performance result, world record energy result and excellent overall performance for the SPC-2/E benchmark.

  • The Oracle ZFS Storage ZS3-2 appliance delivered the top SPC-2 Price-Performance of $12.08 and it delivered an overall score of 16,212.66 SPC-2 MBPS for the SPC-2/E benchmark.

  • The Oracle ZFS Storage ZS3-2 appliance produced the top Performance-Energy SPC-2/E benchmark result of 3.67 SPC2 MBPS / watt.

  • Oracle holds the top two performance results on the SPC-2 benchmark for HDD based systems.

  • The Oracle ZFS Storage ZS3-2 appliance has an 11x price-performance advantage over the IBM DS8870.

  • The Oracle ZFS Storage ZS3-2 appliance has an 8x price-performance advantage over the Hitachi Virtual Storage Platform (VSP).

  • The Oracle ZFS Storage ZS3-2 appliance has an 7.3x price-performance advantage over the HP P9500 XP disk array.

Performance Landscape

SPC-2 Price-Performance

Below is a table of the top SPC-2 Price-Performance results for HDD storage based systems, presented in increasing price-performance order (as of 06/25/2014). The complete set of results may be found at SPC2 top 10 Price-Performance list.

System SPC-2
MBPS
$/SPC-2
MBPS
Results
Identifier
Oracle ZFS Storage ZS3-2 16,212.66 $12.08 BE00002
SGI InfiniteStorage 5600 8,855.70 $15.97 B00065
Oracle ZFS Storage ZS3-4 17,244.22 $22.53 B00067
NEC Storage M700 14,408.89 $25.10 B00066
Sun StorageTek 2530 663.51 $26.48 B00026
Fujitsu ETERNUS DX80 S2 2,685.50 $28.48 B00055
SGI InfiniteStorage 5500-SP 4,064.49 $28.57 B00059
Hitachi Unified Storage VM 11,274.83 $32.64 B00069

SPC-2 MBPS = the Performance Metric
$/SPC-2 MBPS = the Price-Performance Metric
Results Identifier = A unique identification of the result

SPC-2/E Results

The table below list all SPC-2/E results. The SPC-2/E benchmark extends the SPC-2 benchmark by additionally measuring power consumption during the SPC-2 benchmark run.

System SPC-2
MBPS
$/SPC-2
MBPS
TSC Price SPC2 MBPS /
watt
Results
Identifier
Oracle ZFS Storage ZS3-2 16,212.66 $12.08 $195,915 3.67 BE00002
IBM XIV Storage System Gen3 7,467.99 $152.34 $1,137,641 0.81 BE00001

SPC-2 MBPS = the Performance Metric
$/SPC-2 MBPS = the Price-Performance Metric
TSC Price = Total Cost of Ownership Metric
SPC2 MBPS / watt = Number of SPC2 MB/second produced per watt consumed. Higher is Better.
Results Identifier = A unique identification of the result

SPC-2 Performance

The following table list the top SPC-2 -Performance results for HDD storage based systems, presented in decreasing performance order (as of 06/25/2014). The complete set of results may be found at the SPC2 top 10 Performance list.

System SPC-2
MBPS
$/SPC-2
MBPS
TSC Price Results
Identifier
Oracle ZFS Storage ZS3-4 17,244.22 $22.53 $388,472 B00067
Oracle ZFS Storage ZS3-2 16,212.66 $12.08 $195,915 BE00002
Fujitsu ETERNUS DX8870 S2 16,038.74 $79.51 $1,275,163 B00063
IBM System Storage DS8870 15,423.66 $131.21 $2,023,742 B00062
IBM SAN VC v6.4 14,581.03 $129.14 $1,883,037 B00061
Hitachi Virtual Storage Platform (VSP) 13,147.87 $95.38 $1,254,093 B00060
HP StorageWorks P9500 XP Storage Array 13,147.87 $88.34 $1,161,504 B00056

SPC-2 MBPS = the Performance Metric
$/SPC-2 MBPS = the Price-Performance Metric
TSC Price = Total Cost of Ownership Metric
Results Identifier = A unique identification of the result Metric

Complete SPC-2 benchmark results may be found at
http://www.storageperformance.org/results/benchmark_results_spc2.

Configuration Summary

Storage Configuration:

Oracle ZFS Storage ZS3-2 storage system in clustered configuration
2 x Oracle ZFS Storage ZS3-2 controllers, each with
4 x 2.1 GHz 8-core Intel Xeon processors
512 GB memory
12 x Sun Disk shelves, each with
24 x 300 GB 10K RPM SAS-2 drives

Benchmark Description

SPC Benchmark 2 (SPC-2): Consists of three distinct workloads designed to demonstrate the performance of a storage subsystem during the execution of business critical applications that require the large-scale, sequential movement of data. Those applications are characterized predominately by large I/Os organized into one or more concurrent sequential patterns. A description of each of the three SPC-2 workloads is listed below as well as examples of applications characterized by each workload.

  • Large File Processing: Applications in a wide range of fields, which require simple sequential process of one or more large files such as scientific computing and large-scale financial processing.
  • Large Database Queries: Applications that involve scans or joins of large relational tables, such as those performed for data mining or business intelligence.
  • Video on Demand: Applications that provide individualized video entertainment to a community of subscribers by drawing from a digital film library.

SPC-2 is built to:

  • Provide a level playing field for test sponsors.
  • Produce results that are powerful and yet simple to use.
  • Provide value for engineers as well as IT consumers and solution integrators.
  • Is easy to run, easy to audit/verify, and easy to use to report official results.

SPC Benchmark 2/Energy (SPC-2/E): consists of the complete set of SPC-2 performance measurement and reporting plus the measurement and reporting of energy use. This benchmark extension provides measurement and reporting to complete storage configurations, complementing SPC-2C/E, which focuses on storage component configurations.

See Also

Disclosure Statement

SPC-2 and SPC-2 MBPS are registered trademarks of Storage Performance Council (SPC). Results as of June 25, 2014, for more information see www.storageperformance.org.

Fujitsu ETERNUS DX80 S2, B00055, Fujitsu ETERNUS DX8870 S2 - B00063, HP StorageWorks P9500 XP Storage Array - B00056, Hitachi Unified Storage VM - B00069, Hitachi Virtual Storage Platform (VSP) - B00060, IBM SAN VC v6.4 - B00061, IBM System Storage DS8870 - B00062, IBM XIV Storage System Gen3 - BE00001, NEC Storage M700 - B00066, Oracle ZFS Storage ZS3-2 - BE00002, Oracle ZFS Storage ZS3-4 - B00067, SGI InfiniteStorage 5500-SP - B00059, SGI InfiniteStorage 5600 - B00065, Sun StorageTek 2530 - B00026.

Thursday Mar 27, 2014

SPARC M6-32 Produces SAP SD Two-Tier Benchmark World Record for 32-Processor Systems

Oracle's SPARC M6-32 server produced a world record result for 32-processors on the SAP two-tier Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement Package 5 for SAP ERP 6.0 (32 chips / 384 cores / 3072 threads).

  • SPARC M6-32 server achieved 140,000 SAP SD benchmark users with a low average dialog response time of 0.58 seconds running the SAP two-tier Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement package 5 for SAP ERP 6.0.

  • The SPARC M6-32 delivered 2.5 times more users than the IBM Power 780 result using SAP Enhancement Package 5 for SAP ERP 6.0. The IBM result also had 1.7 times worse average dialog response time compared to the SPARC M6-32 server result.

  • The SPARC M6-32 delivered 3.0 times more users than the Fujitsu PRIMEQUEST 2800E (with Intel Xeon E7-8890 v2 processors) result. The Fujitsu result also had 1.7 times worse average dialog response time compared to the SPARC M6-32 server result.

  • The SPARC M6-32 server solution was run with Oracle Solaris 11 and used Oracle Database 11g.

Performance Landscape

SAP-SD 2-Tier Performance Table (in decreasing performance order). With SAP ERP 6.0 Enhancement Package 4 for SAP ERP 6.0 (Old version of the benchmark, obsolete at the end of April, 2012), and SAP ERP 6.0 Enhancement Package 5 for SAP ERP 6.0 results (current version of the benchmark as of May, 2012).

System
Processor
Ch / Co / Th — Memory
OS
Database
Users Resp Time
(sec)
Version Cert#
Fujitsu SPARC M10-4S
SPARC64 X @3.0 GHz
40 / 640 / 1280 — 10 TB
Solaris 11
Oracle 11g
153,000 0.87 EHP5 2013014
SPARC M6-32 Server
SPARC M6 @3.6 GHz
32 / 384 / 3072 — 16 TB
Solaris 11
Oracle 11g
140,000 0.58 EHP5 2014008
IBM Power 795
POWER7 @4 GHz
32 / 256 / 1024 — 4 TB
AIX 7.1
DB2 9.7
126,063 0.98 EHP4 2010046
IBM Power 780
POWER7+ @3.72 GHz
12 / 96 / 834 — 1536 GB
AIX 7.1
DB2 10
57,024 0.98 EHP5 2012033
Fujitsu PRIMEQUEST 2800E
Intel Xeon E7-8890 v2 @2.8 GHz
8 / 120 / 240 — 1024 GB
Windows Server 2012 SE
SQL Server 2012
47,500 0.97 EHP5 2014003
IBM Power 760
POWER7+ @3.41 GHz
8 / 48 / 192 — 1024 GB
AIX 7.1
DB2 10
25,488 0.99 EHP5 2013004

Version – Version of SAP, EHP5 refers to SAP ERP 6.0 Enhancement Package 5 for SAP ERP 6.0 and EHP4 refers to SAP ERP 6.0 Enhancement Package 4 for SAP ERP 6.0

Ch / Co / Th – Total chips, coreas and threads

Complete benchmark results may be found at the SAP benchmark website http://www.sap.com/benchmark.

Configuration Summary and Results

Hardware Configuration:

1 x SPARC M6-32 server with
32 x 3.6 GHz SPARC M6 processors (total of 32 processors / 384 cores / 3072 threads)
16 TB memory
6 x Sun Server X3-2L each with
2 x Intel Xeon E5-2609 2.4 GHz Processors
16 GB Memory
4 x Flash Accelerator F40
12 x 3 TB SAS disks
2 x Sun Server X3-2L each with
2 x Intel Xeon E5-2609 2.4 GHz Processors
16 GB Memory
1 x 8-Port 6Gbps SAS-2 RAID PCI Express HBA
12 x 3 TB SAS disks

Software Configuration:

Oracle Solaris 11
SAP Enhancement Package 5 for SAP ERP 6.0
Oracle Database 11g Release 2

Certified Results (published by SAP)

Number of SAP SD benchmark users:
140,000
Average dialog response time:
0.58 seconds
Throughput:

  Fully processed order line items per hour:
15,878,670
  Dialog steps per hour:
47,636,000
  SAPS:
793,930
Average database request time (dialog/update):
0.020 sec / 0.041 sec
SAP Certification:
2014008

Benchmark Description

The SAP Standard Application SD (Sales and Distribution) Benchmark is an ERP business test that is indicative of full business workloads of complete order processing and invoice processing, and demonstrates the ability to run both the application and database software on a single system. The SAP Standard Application SD Benchmark represents the critical tasks performed in real-world ERP business environments.

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

Two-tier SAP Sales and Distribution (SD) standard application benchmarks, SAP Enhancement Package 5 for SAP ERP 6.0 as of 3/26/14:

SPARC M6-32 (32 processors, 384 cores, 3072 threads) 140,000 SAP SD users, 32 x 3.6 GHz SPARC M6, 16 TB memory, Oracle Database 11g, Oracle Solaris 11, Cert# 2014008. Fujitsu SPARC M10-4S (40 processors, 640 cores, 1280 threads) 153,000 SAP SD users, 40 x 3.0 GHz SPARC65 X, 10 TB memory, Oracle Database 11g, Oracle Solaris 11, Cert# 2013014. IBM Power 780 (12 processors, 96 cores, 384 threads) 57,024 SAP SD users, 12 x 3.72 GHz IBM POWER7+, 1536 GB memory, DB210, AIX7.1, Cert#2012033. Fujitsu PRIMEQUEST 2800E (8 processors, 120 cores, 240 threads) 47,500 SAP SD users, 8 x 2.8 GHz Intel Xeon Processor E7-8890 v2, 1024 GB memory, SQL Server 2012, Windows Server 2012 Standard Edition, Cert# 2014003. IBM Power 760 (8 processors, 48 cores, 192 threads) 25,488 SAP SD users, 8 x 3.41 GHz IBM POWER7+, 1024 GB memory, DB2 10, AIX 7.1, Cert#2013004.

Two-tier SAP Sales and Distribution (SD) standard application benchmarks, SAP Enhancement Package 4 for SAP ERP 6.0 as of 3/26/14:

IBM Power 795 (32 processors, 256 cores, 1024 threads) 126,063 SAP SD users, 32 x 4 GHz IBM POWER7, 4 TB memory, DB2 9.7, AIX7.1, Cert#2010046.

SAP, R/3, reg TM of SAP AG in Germany and other countries. More info www.sap.com/benchmark

Wednesday Mar 05, 2014

SPARC T5-2 Delivers World Record 2-Socket SPECvirt_sc2010 Benchmark

Oracle's SPARC T5-2 server delivered a world record two-chip SPECvirt_sc2010 result of 4270 @ 264 VMs, establishing performance superiority in virtualized environments of the SPARC T5 processors with Oracle Solaris 11, which includes as standard virtualization products Oracle VM for SPARC and Oracle Solaris Zones.

  • The SPARC T5-2 server has 2.3x better performance than an HP BL620c G7 blade server (with two Westmere EX processors) which used VMware ESX 4.1 U1 virtualization software (best SPECvirt_sc2010 result on two-chip servers using VMware software).

  • The SPARC T5-2 server has 1.6x better performance than an IBM Flex System x240 server (with two Sandy Bridge processors) which used Kernel-based Virtual Machines (KVM).

  • This is the first SPECvirt_sc2010 result using Oracle production level software: Oracle Solaris 11.1, Oracle WebLogic Server 10.3.6, Oracle Database 11g Enterprise Edition, Oracle iPlanet Web Server 7 and Oracle Java Development Kit 7 (JDK). The only exception for the Dovecot mail server.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECvirt_sc2010 Results. The following table highlights the leading two-chip results for the benchmark, bigger is better.

SPECvirt_sc2010
Leading Two-Chip Results
System Processor Result @ VMs Virtualization Software
SPARC T5-2 2 x SPARC T5, 3.6 GHz 4270 @ 264 Oracle VM Server for SPARC 3.0
Oracle Solaris Zones
IBM Flex System x240 2 x Intel E5-2690, 2.9 GHz 2741 @ 168 Red Hat Enterprise Linux 6.4 KVM
HP Proliant BL6200c G7 2 x Intel E7-2870, 2.4 GHz 1878 @ 120 VMware ESX 4.1 U1

Configuration Summary

System Under Test Highlights:

1 x SPARC T5-2 server, with
2 x 3.6 GHz SPARC T5 processors
1 TB memory
Oracle Solaris 11.1
Oracle VM Server for SPARC 3.0
Oracle iPlanet Web Server 7.0.15
Oracle PHP 5.3.14
Dovecot 2.1.17
Oracle WebLogic Server 11g (10.3.6)
Oracle Database 11g (11.2.0.3)
Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.7.0_51

Benchmark Description

The SPECvirt_sc2010 benchmark is SPEC's first benchmark addressing performance of virtualized systems. It measures the end-to-end performance of all system components that make up a virtualized environment.

The benchmark utilizes several previous SPEC benchmarks which represent common tasks which are commonly used in virtualized environments. The workloads included are derived from SPECweb2005, SPECjAppServer2004 and SPECmail2008. Scaling of the benchmark is achieved by running additional sets of virtual machines until overall throughput reaches a peak. The benchmark includes a quality of service criteria that must be met for a successful run.

Key Points and Best Practices

  • The SPARC T5 server running the Oracle Solaris 11.1, utilizes embedded virtualization products as the Oracle VM for SPARC and Oracle Solaris Zones, which provide a low overhead, flexible, scalable and manageable virtualization environment.

  • In order to provide a high level of data integrity and availability, all the benchmark data sets are stored on mirrored (RAID1) storage.

See Also

Disclosure Statement

SPEC and the benchmark name SPECvirt_sc are registered trademarks of the Standard Performance Evaluation Corporation. Results from www.spec.org as of 3/5/2014. SPARC T5-2, SPECvirt_sc2010 4270 @ 264 VMs; IBM Flex System x240, SPECvirt_sc2010 2741 @ 168 VMs; HP Proliant BL620c G7, SPECvirt_sc2010 1878 @ 120 VMs.

Friday Feb 14, 2014

SPARC M6-32 Delivers Oracle E-Business and PeopleSoft World Record Benchmarks, Linear Data Warehouse Scaling in a Virtualized Configuration

This result demonstrates how the combination of Oracle virtualization technologies for SPARC and Oracle's SPARC M6-32 server allow the deployment and concurrent high performance execution of multiple Oracle applications and databases sized for the Enterprise.

  • In an 8-chip Dynamic Domain (also known as PDom), the SPARC M6-32 server set a World Record E-Business 12.1.3 X-Large world record with 14,660 online users running five simultaneous E-Business modules.

  • In a second 8-chip Dynamic Domain, the SPARC M6-32 server set a World Record PeopleSoft HCM 9.1 HR Self-Service online supporting 35,000 users while simultaneously running a batch workload in 29.17 minutes. This was done with a database of 600,480 employees. Two other separate tests were run, one supporting 40,000 online users only and another a batch-only workload that was run in 18.27 min.

  • In a third Dynamic Domain with 16-chips on the SPARC M6-32 server, a data warehouse test was run that showed near-linear scaling.

  • On the SPARC M6-32 server, several critical applications instances were virtualized: an Oracle E-Business application and database, an Oracle's PeopleSoft application and database, and a Decision Support database instance using Oracle Database 12c.

  • In this Enterprise Virtualization benchmark a SPARC M6-32 server utilized all levels of Oracle Virtualization features available for SPARC servers. The 32-chip SPARC M6 based server was divided in three separate Dynamic Domains (also known as PDoms), available only on the SPARC Enterprise M-Series systems, which are completely electrically isolated and independent hardware partitions. Each PDom was subsequently split into multiple hypervisor-based Oracle VM for SPARC partitions (also known as LDoms), each one running its own Oracle Solaris kernel and managing its own CPUs and I/O resources. The hardware resources allocated to each Oracle VM for SPARC partition were then organized in various Oracle Solaris Zones, to further refine application tier isolation and resources management. The three PDoms were dedicated to the enterprise applications as follows:

    • Oracle E-Business PDom: Oracle E-Business 12.1.3 Suite World Record Extra-Large benchmark, exercising five Online Modules: Customer Service, Human Resources Self Service, iProcurement, Order Management and Financial, with 14,660 users and an average user response time under 2 seconds.

    • PeopleSoft PDom: PeopleSoft Human Capital Management (HCM) 9.1 FP2 World Record Benchmark, using PeopleTools 8.52 and an Oracle Database 11g Release 2, with 35,000 users, at an average user Search Time of 1.46 seconds and Save Time of 0.93 seconds. An online run with 40,000 users, had an average user Search Time of 2.17 seconds and Save Time of 1.39 seconds, and a Payroll batch run completed in 29.17 minutes elapsed time for more than 500,000 employees.

    • Decision Support PDom: An Oracle Database 12c instance executing a Decision Support workload on about 30 billion rows of data and achieving linear scalability, i.e. on the 16 chips comprising the PDom, the workload ran 16x faster than on a single chip. Specifically, the 16-chip PDom processed about 320M rows/sec whereas a single chip could process about 20M rows/sec.

  • The SPARC M6-32 server is ideally suited for large-memory utilization. In this virtualized environment, three critical applications made use of 16 TB of physical memory. Each of the Oracle VM Server for SPARC environments utilized from 4 to 8 TB of memory, more than the limits of other virtualization solutions.

  • SPARC M6-32 Server Virtualization Layout Highlights

    • The Oracle E-Business application instances were run in a dedicated Dynamic Domain consisting of 8 SPARC M6 processors and 4 TB of memory. The PDom was split into four symmetric Oracle VM Server for SPARC (LDoms) environments of 2 chips and 1 TB of memory each, two dedicated to the Application Server tier and the other two to the Database Server tier. Each Logical Domain was subsequently divided into two Oracle Solaris Zones, for a total of eight, one for each E-Business Application server and one for each Oracle Database 11g instance.

    • The PeopleSoft application was run in a dedicated Dynamic Domain (PDom) consisting of 8 SPARC M6 processors and 4 TB of memory. The PDom was split into two Oracle VM Server for SPARC (LDoms) environments one of 6 chips and 3 TB of memory, reserved for the Web and Application Server tiers, and a second one of 2 chips and 1 TB of memory, reserved for the Database tier. Two PeopleSoft Application Servers, a Web Server instance, and a single Oracle Database 11g instance were each executed in their respective and exclusive Oracle Solaris Zone.

    • The Oracle Database 12c Decision Support workload was run in a Dynamic Domain consisting of 16 SPARC M6 processors and 8 TB of memory.

  • All the Oracle Applications and Database instances were running at high level of performance and concurrently in a virtualized environment. Running three Enterprise level application environments on a single SPARC M6-32 server offers centralized administration, simplified physical layout, high availability and security features (as each PDom and LDom runs its own Oracle Solaris operating system copy physically and logically isolated from the other environments), enabling the coexistence of multiple versions Oracle Solaris and application software on a single physical server.

  • Dynamic Domains and Oracle VM Server for SPARC guests were configured with independent direct I/O domains, allowing for fast and isolated I/O paths, providing secure and high performance I/O access.

Performance Landscape

Oracle E-Business Test using Oracle Database 11g
SPARC M6-32 PDom, 8 SPARC M6 Processors, 4 TB Memory
Total Online Users Weighted Average
Response Time (sec)
90th Percentile
Response Time (s)
14,660 0.81 0.88
Multiple Online Modules X-Large Configuration (HR Self-Service, Order Management, iProcurement, Customer Service, Financial)

PeopleSoft HR Self-Service Online Plus Payroll Batch using Oracle Database 11g
SPARC M6-32 PDom, 8 SPARC M6 Processors, 4 TB Memory
HR Self-Service Payroll Batch
Elapsed (min)
Online Users Average User
Search / Save
Time (sec)
Transactions
per Second
35,000 1.46 / 0.93 116 29.17

HR Self-Service Only Payroll Batch Only
Elapsed (min)
40,000 2.17 / 1.39 132 18.27

Oracle Database 12c Decision Support Query Test
SPARC M6-32 PDom, 16 SPARC M6 Processors, 8 TB Memory
Parallelism
Chips Used
Rows Processing Rate
(rows/s)
Scaling Normalized to 1 Chip
16 319,981,734 15.9
8 162,545,303 8.1
4 80,943,271 4.0
2 40,458,329 2.0
1 20,086,829 1.0

Configuration Summary

System Under Test:

SPARC M6-32 server with
32 x SPARC M6 processors (3.6 GHz)
16 TB memory

Storage Configuration:

6 x Sun Storage 2540-M2 each with
8 x Expansion Trays (each tray equipped with 12 x 300 GB SAS drives)
7 x Sun Server X3-2L each with
2 x Intel Xeon E5-2609 2.4 GHz Processors
16 GB Memory
4 x Sun Flash Accelerator F40 PCIe 400 GB cards
Oracle Solaris 11.1 (COMSTAR)
1 x Sun Server X3-2L with
2 x Intel Xeon E5-2609 2.4 GHz Processors
16 GB Memory
12 x 3 TB SAS disks
Oracle Solaris 11.1 (COMSTAR)

Software Configuration:

Oracle Solaris 11.1 (11.1.10.5.0), Oracle E-Business
Oracle Solaris 11.1 (11.1.10.5.0), PeopleSoft
Oracle Solaris 11.1 (11.1.9.5.0), Decision Support
Oracle Database 11g Release 2, Oracle E-Business and PeopleSoft
Oracle Database 12c Release 1, Decision Support
Oracle E-Business Suite 12.1.3
PeopleSoft Human Capital Management 9.1 FP2
PeopleSoft PeopleTools 8.52.03
Oracle Java SE 6u32
Oracle Tuxedo, Version 10.3.0.0, 64-bit, Patch Level 043
Oracle WebLogic Server 11g (10.3.4)

Oracle Dynamic Domains (PDoms) resources:


Oracle E-Business PeopleSoft Oracle DSS
Processors 8 8 16
Memory 4 TB 4 TB 8 TB
Oracle Solaris 11.1 (11.1.10.5.0) 11.1 (11.1.10.5.0) 11.1 (11.1.9.5.0)
Oracle Database 11g 11g 12c
Oracle VM for SPARC /
Oracle Solaris Zones
4 LDom / 8 Zones 2 LDom / 4 Zones None
Storage 7 x Sun Server X3-2L 1 x Sun Server X3-2L
(12 x 3 TB SAS )
2 x Sun Storage 2540-M2 / 2501 pairs
4 x Sun Storage 2540-M2/2501 pairs

Benchmark Description

This benchmark consists of three different applications running concurrently. It shows that large, enterprise workloads can be run on a single system and without performance impact between application environments.

The three workloads are:

  • Oracle E-Business Suite Online

    • This test simulates thousands of online users executing transactions typical of an internal Enterprise Resource Processing, including 5 application modules: Customer Service, Human Resources Self Service, Procurement, Order Management and Financial.

    • Each database tier uses a database instance of about 600 GB in size, and supporting thousands of application users, accessing hundreds of objects (tables, indexes, SQL stored procedures, etc.).

    • The application tier includes multiple web and application server instances, specifically Apache Web Server, Oracle Application Server 10g and Oracle Java SE 6u32.

  • PeopleSoft Human Capital Management

    • This test simulates thousands of online employees, managers and Human Resource administrators executing transactions typical of a Human Resources Self Service application for the Enterprise. Typical transactions are: viewing paychecks, promoting and hiring employees, updating employee profiles, etc.

    • The database tier uses a database instance of about 500 GB in size, containing information for 500,480 employees.

    • The application tier for this test includes web and application server instances, specifically Oracle WebLogic Server 11g, PeopleSoft Human Capital Management 9.1 and Oracle Java SE 6u32.

  • Decision Support Workload using the Oracle Database.

    • The query processes 30 billion rows stored in the Oracle Database, making heavy use of Oracle parallel query processing features. It performs multiple aggregations and summaries by reading and processing all the rows of the database.

Key Points and Best Practices

Oracle E-Business Environment

The Oracle E-Business Suite setup consisted 4 Oracle E-Business environments running 5 online Oracle E-Business modules simultaneously.

The Oracle E-Business environments were deployed on 4 Oracle VM for SPARC, respectively 2 for the Application tier and 2 for the Database tier. Each LDom included 2 SPARC M6 processor chips. The Application LDom was further split into 2 Oracle Solaris Zones, each one containing one Oracle E-Business Application instance. Similarly, on the Database tier, each LDom was further divided into 2 Oracle Solaris Zones, each containing an Oracle Database instance. Applications on the same LDom shared a 10 GbE network link to connect to the Database tier LDom. Each Application in a Zone was connected to its own dedicated Database Zone. The communication between the two Zones was implemented via Oracle Solaris 11 virtual network, which provides high performance, low latency transfers at memory speed using large frames (9000 bytes vs typical 1500 bytes frames).

The Oracle E-Business setup made use of the Oracle Database Shared Server feature in order to limit memory utilization, as well as the number of database Server processes. The Oracle Database configuration and optimization was substantially out-of-the-box, except for proper sizing the Oracle Database memory areas (System Global Area and Program Global Area).

In the Oracle E-Business Application LDom handling Customer Service and HR Self Service modules, 28 Forms servers and 8 OC4J application servers were hosted in the two separate Oracle Solaris Zones, for a total of 56 forms servers and 16 applications servers.

All the Oracle Database server processes and the listener processes were executed in the Oracle Solaris FX scheduler class.

PeopleSoft Environment

The PeopleSoft Application Oracle VM for SPARC had one Oracle Solaris Zone of 12 cores containing the web tier and two Oracle Solaris Zones of 57 cores total containing the Application tier. The Database tier was contained in an Oracle VM for SPARC consisting of one Oracle Solaris Zone of 24 cores. One core, in the Application Oracle VM, was dedicated to network and disk interrupt handling.

All database data files, recovery files and Oracle Clusterware files for the PeopleSoft test were created with the Oracle Automatic Storage Management (Oracle ASM) volume manager for the added benefit of the ease of management provided by Oracle ASM integrated storage management solution.

In the application tier, 5 PeopleSoft domains with 350 application servers (70 per each domain) were hosted in the two separate Oracle Solaris Zones for a total of 10 domains with 700 application server processes.

All PeopleSoft Application processes and Web Server JVM instances were executed in the Oracle Solaris FX scheduler class.

Oracle Decision Support Environment

The decision support workload showed how the combination of a large memory (8 TB) and a large number of processors (16 chips comprising 1536 virtual CPUs) together with Oracle parallel query facility can linearly increase the performance of certain decision support queries as the number of CPUs increase.

The large memory was used to cache the entire 30 billion row Oracle table in memory. There are a number of ways to accomplish this. The method deployed in this test was to allocate sufficient memory for Oracle's "keep cache" and direct the table to the "keep cache."

To demonstrate scalability, it was necessary to ensure that the number of Oracle parallel servers was always equal to the number of available virtual CPUs. This was accomplished by the combination of providing a degree of parallelism hint to the query and setting both parallel_max_servers and parallel_min_servers to the number of virtual CPUs.

The number of virtual CPUs for each stage of the scalability test was adjusted using the psradm command available in Oracle Solaris.

See Also

Disclosure Statement

Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. PeopleSoft results as of 02/14/2014. Other results as of 09/22/2013.

Oracle E-Business Suite R12 extra-large multiple-online module benchmark, SPARC M6-32, SPARC M6, 3.6 GHz, 8 chips, 96 cores, 768 threads, 4 TB memory, 14,660 online users, average response time 0.81 sec, 90th percentile response time 0.88 sec, Oracle Solaris 11.1, Oracle Solaris Zones, Oracle VM for SPARC, Oracle E-Business Suite 12.1.3, Oracle Database 11g Release 2, Results as of 9/22/2013.

Monday Nov 25, 2013

World Record Single System TPC-H @10000GB Benchmark on SPARC T5-4

Oracle's SPARC T5-4 server delivered world record single server performance of 377,594 QphH@10000GB with price/performance of $4.65/QphH@10000GB USD on the TPC-H @10000GB benchmark. This result shows that the 4-chip SPARC T5-4 server is significantly faster than the 8-chip server results from HP (Intel x86 based).

  • The SPARC T5-4 server with four SPARC T5 processors is 2.4 times faster than the HP ProLiant DL980 G7 server with eight x86 processors.

  • The SPARC T5-4 server delivered 4.8 times better performance per chip and 3.0 times better performance per core than the HP ProLiant DL980 G7 server.

  • The SPARC T5-4 server has 28% better price/performance than the HP ProLiant DL980 G7 server (for the price/QphH metric).

  • The SPARC T5-4 server with 2 TB memory is 2.4 times faster than the HP ProLiant DL980 G7 server with 4 TB memory (for the composite metric).

  • The SPARC T5-4 server took 9 hours, 37 minutes, 54 seconds for data loading while the HP ProLiant DL980 G7 server took 8.3 times longer.

  • The SPARC T5-4 server accomplished the refresh function in around a minute, the HP ProLiant DL980 G7 server took up to 7.1 times longer to do the same function.

This result demonstrates a complete data warehouse solution that shows the performance both of individual and concurrent query processing streams, faster loading, and refresh of the data during business operations. The SPARC T5-4 server delivers superior performance and cost efficiency when compared to the HP result.

Performance Landscape

The table lists the leading TPC-H @10000GB results for non-clustered systems.

TPC-H @10000GB, Non-Clustered Systems
System
Processor
P/C/T – Memory
Composite
(QphH)
$/perf
($/QphH)
Power
(QppH)
Throughput
(QthH)
Database Available
SPARC T5-4
3.6 GHz SPARC T5
4/64/512 – 2048 GB
377,594.3 $4.65 342,714.1 416,024.4 Oracle 11g R2 11/25/13
HP ProLiant DL980 G7
2.4 GHz Intel Xeon E7-4870
8/80/160 – 4096 GB
158,108.3 $6.49 185,473.6 134,780.5 SQL Server 2012 04/15/13

P/C/T = Processors, Cores, Threads
QphH = the Composite Metric (bigger is better)
$/QphH = the Price/Performance metric in USD (smaller is better)
QppH = the Power Numerical Quantity (bigger is better)
QthH = the Throughput Numerical Quantity (bigger is better)

The following table lists data load times and average refresh function times.

TPC-H @10000GB, Non-Clustered Systems
Database Load & Database Refresh
System
Processor
Data Loading
(h:m:s)
T5
Advan
RF1
(sec)
T5
Advan
RF2
(sec)
T5
Advan
SPARC T5-4
3.6 GHz SPARC T5
09:37:54 8.3x 58.8 7.1x 62.1 6.4x
HP ProLiant DL980 G7
2.4 GHz Intel Xeon E7-4870
79:28:23 1.0x 416.4 1.0x 394.9 1.0x

Data Loading = database load time
RF1 = throughput average first refresh transaction
RF2 = throughput average second refresh transaction
T5 Advan = the ratio of time to the SPARC T5-4 server time

Complete benchmark results found at the TPC benchmark website http://www.tpc.org.

Configuration Summary and Results

Server Under Test:

SPARC T5-4 server
4 x SPARC T5 processors (3.6 GHz total of 64 cores, 512 threads)
2 TB memory
2 x internal SAS (2 x 300 GB) disk drives
12 x 16 Gb FC HBA

External Storage:

24 x Sun Server X4-2L servers configured as COMSTAR nodes, each with
2 x 2.5 GHz Intel Xeon E5-2609 v2 processors
4 x Sun Flash Accelerator F80 PCIe Cards, 800 GB each
6 x 4 TB 7.2K RPM 3.5" SAS disks
1 x 8 Gb dual port HBA

2 x 48 port Brocade 6510 Fibre Channel Switches

Software Configuration:

Oracle Solaris 11.1
Oracle Database 11g Release 2 Enterprise Edition

Audited Results:

Database Size: 10000 GB (Scale Factor 10000)
TPC-H Composite: 377,594.3 QphH@10000GB
Price/performance: $4.65/QphH@10000GB USD
Available: 11/25/2013
Total 3 year Cost: $1,755,709 USD
TPC-H Power: 342,714.1
TPC-H Throughput: 416,024.4
Database Load Time: 9:37:54

Benchmark Description

The TPC-H benchmark is a performance benchmark established by the Transaction Processing Council (TPC) to demonstrate Data Warehousing/Decision Support Systems (DSS). TPC-H measurements are produced for customers to evaluate the performance of various DSS systems. These queries and updates are executed against a standard database under controlled conditions. Performance projections and comparisons between different TPC-H Database sizes (100GB, 300GB, 1000GB, 3000GB, 10000GB, 30000GB and 100000GB) are not allowed by the TPC.

TPC-H is a data warehousing-oriented, non-industry-specific benchmark that consists of a large number of complex queries typical of decision support applications. It also includes some insert and delete activity that is intended to simulate loading and purging data from a warehouse. TPC-H measures the combined performance of a particular database manager on a specific computer system.

The main performance metric reported by TPC-H is called the TPC-H Composite Query-per-Hour Performance Metric (QphH@SF, where SF is the number of GB of raw data, referred to as the scale factor). QphH@SF is intended to summarize the ability of the system to process queries in both single and multiple user modes. The benchmark requires reporting of price/performance, which is the ratio of the total HW/SW cost plus 3 years maintenance to the QphH. A secondary metric is the storage efficiency, which is the ratio of total configured disk space in GB to the scale factor.

Key Points and Best Practices

  • COMSTAR (Common Multiprotocol SCSI Target) is the software framework that enables an Oracle Solaris host to serve as a SCSI Target platform. COMSTAR uses a modular approach to break the huge task of handling all the different pieces in a SCSI target subsystem into independent functional modules which are glued together by the SCSI Target Mode Framework (STMF). The modules implementing functionality at SCSI level (disk, tape, medium changer etc.) are not required to know about the underlying transport. And the modules implementing the transport protocol (FC, iSCSI, etc.) are not aware of the SCSI-level functionality of the packets they are transporting. The framework hides the details of allocation providing execution context and cleanup of SCSI commands and associated resources and simplifies the task of writing the SCSI or transport modules.

  • The SPARC T5-4 server achieved a peak IO rate of 37 GB/sec from the Oracle database configured with this storage.

  • Twelve COMSTAR nodes were mirrored to another twelve COMSTAR nodes on which all of the Oracle database files were placed. IO performance was high and balanced across all the nodes.

  • Oracle Solaris 11.1 required very little system tuning.

  • Some vendors try to make the point that storage ratios are of customer concern. However, storage ratio size has more to do with disk layout and the increasing capacities of disks – so this is not an important metric when comparing systems.

  • The SPARC T5-4 server and Oracle Solaris efficiently managed the system load of nearly two thousand Oracle Database parallel processes.

See Also

Disclosure Statement

TPC Benchmark, TPC-H, QphH, QthH, QppH are trademarks of the Transaction Processing Performance Council (TPC). Results as of 11/25/13, prices are in USD. SPARC T5-4 www.tpc.org/3293; HP ProLiant DL980 G7 www.tpc.org/3285.

Thursday Sep 26, 2013

SPARC M6-32 Delivers Oracle E-Business and PeopleSoft World Record Benchmarks, Linear Data Warehouse Scaling in a Virtualized Configuration

This result has been superceded.  Please see the latest result.

 This result demonstrates how the combination of Oracle virtualization technologies for SPARC and Oracle's SPARC M6-32 server allow the deployment and concurrent high performance execution of multiple Oracle applications and databases sized for the Enterprise.

  • In an 8-chip Dynamic Domain (also known as PDom), the SPARC M6-32 server set a World Record E-Business 12.1.3 X-Large world record with 14,660 online users running five simultaneous E-Business modules.

  • In a second 8-chip Dynamic Domain, the SPARC M6-32 server set a World Record PeopleSoft HCM 9.1 HR Self-Service online supporting 34,000 users while simultaneously running a batch workload in 29.7 minutes. This was done with a database of 600,480 employees. In a separate test running a batch-only workload was run in 21.2 min.

  • In a third Dynamic Domain with 16-chips on the SPARC M6-32 server, a data warehouse test was run that showed near-linear scaling.

  • On the SPARC M6-32 server, several critical applications instances were virtualized: an Oracle E-Business application and database, an Oracle's PeopleSoft application and database, and a Decision Support database instance using Oracle Database 12c.

  • In this Enterprise Virtualization benchmark a SPARC M6-32 server utilized all levels of Oracle Virtualization features available for SPARC servers. The 32-chip SPARC M6 based server was divided in three separate Dynamic Domains (also known as PDoms), available only on the SPARC Enterprise M-Series systems, which are completely electrically isolated and independent hardware partitions. Each PDom was subsequently split into multiple hypervisor-based Oracle VM for SPARC partitions (also known as LDoms), each one running its own Oracle Solaris kernel and managing its own CPUs and I/O resources. The hardware resources allocated to each Oracle VM for SPARC partition were then organized in various Oracle Solaris Zones, to further refine application tier isolation and resources management. The three PDoms were dedicated to the enterprise applications as follows:

    • Oracle E-Business PDom: Oracle E-Business 12.1.3 Suite World Record Extra-Large benchmark, exercising five Online Modules: Customer Service, Human Resources Self Service, iProcurement, Order Management and Financial, with 14,660 users and an average user response time under 2 seconds.

    • PeopleSoft PDom: PeopleSoft Human Capital Management (HCM) 9.1 FP2 World Record Benchmark, using PeopleTools 8.52 and an Oracle Database 11g Release 2, with 34,000 users, at an average user Search Time of 1.11 seconds and Save Time of 0.77 seconds, and a Payroll batch run completed in 29.7 minutes elapsed time for more than 500,000 employees.

    • Decision Support PDom: An Oracle Database 12c instance executing a Decision Support workload on about 30 billion rows of data and achieving linear scalability, i.e. on the 16 chips comprising the PDom, the workload ran 16x faster than on a single chip. Specifically, the 16-chip PDom processed about 320M rows/sec whereas a single chip could process about 20M rows/sec.

  • The SPARC M6-32 server is ideally suited for large-memory utilization. In this virtualized environment, three critical applications made use of 16 TB of physical memory. Each of the Oracle VM Server for SPARC environments utilized from 4 to 8 TB of memory, more than the limits of other virtualization solutions.

  • SPARC M6-32 Server Virtualization Layout Highlights

    • The Oracle E-Business application instances were run in a dedicated Dynamic Domain consisting of 8 SPARC M6 processors and 4 TB of memory. The PDom was split into four symmetric Oracle VM Server for SPARC (LDoms) environments of 2 chips and 1 TB of memory each, two dedicated to the Application Server tier and the other two to the Database Server tier. Each Logical Domain was subsequently divided into two Oracle Solaris Zones, for a total of eight, one for each E-Business Application server and one for each Oracle Database 11g instance.

    • The PeopleSoft application was run in a dedicated Dynamic Domain (PDom) consisting of 8 SPARC M6 processors and 4 TB of memory. The PDom was split into two Oracle VM Server for SPARC (LDoms) environments one of 6 chips and 3 TB of memory, reserved for the Web and Application Server tiers, and a second one of 2 chips and 1 TB of memory, reserved for the Database tier. Two PeopleSoft Application Servers, a Web Server instance, and a single Oracle Database 11g instance were each executed in their respective and exclusive Oracle Solaris Zone.

    • The Oracle Database 12c Decision Support workload was run in a Dynamic Domain consisting of 16 SPARC M6 processors and 8 TB of memory.

  • All the Oracle Applications and Database instances were running at high level of performance and concurrently in a virtualized environment. Running three Enterprise level application environments on a single SPARC M6-32 server offers centralized administration, simplified physical layout, high availability and security features (as each PDom and LDom runs its own Oracle Solaris operating system copy physically and logically isolated from the other environments), enabling the coexistence of multiple versions Oracle Solaris and application software on a single physical server.

  • Dynamic Domains and Oracle VM Server for SPARC guests were configured with independent direct I/O domains, allowing for fast and isolated I/O paths, providing secure and high performance I/O access.

Performance Landscape

Oracle E-Business Test using Oracle Database 11g
SPARC M6-32 PDom, 8 SPARC M6 Processors, 4 TB Memory
Total Online Users Weighted Average
Response Time (sec)
90th Percentile
Response Time (s)
14,660 0.81 0.88
Multiple Online Modules X-Large Configuration (HR Self-Service, Order Management, iProcurement, Customer Service, Financial)

PeopleSoft HR Self-Service Online Plus Payroll Batch using Oracle Database 11g
SPARC M6-32 PDom, 8 SPARC M6 Processors, 4 TB Memory
HR Self-Service Payroll Batch
Elapsed (min)
Online Users Average User
Search / Save
Time (sec)
Transactions
per Second
34,000 1.11 / 0.77 113 29.7

Payroll Batch Only
Elapsed (min)
21.17

Oracle Database 12c Decision Support Query Test
SPARC M6-32 PDom, 16 SPARC M6 Processors, 8 TB Memory
Parallelism
Chips Used
Rows Processing Rate
(rows/s)
Scaling Normalized to 1 Chip
16 319,981,734 15.9
8 162,545,303 8.1
4 80,943,271 4.0
2 40,458,329 2.0
1 20,086,829 1.0

Configuration Summary

System Under Test:

SPARC M6-32 server with
32 x SPARC M6 processors (3.6 GHz)
16 TB memory

Storage Configuration:

6 x Sun Storage 2540-M2 each with
8 x Expansion Trays (each tray equipped with 12 x 300 GB SAS drives)
7 x Sun Server X3-2L each with
2 x Intel Xeon E5-2609 2.4 GHz Processors
16 GB Memory
4 x Sun Flash Accelerator F40 PCIe 400 GB cards
Oracle Solaris 11.1 (COMSTAR)
1 x Sun Server X3-2L with
2 x Intel Xeon E5-2609 2.4 GHz Processors
16 GB Memory
12 x 3 TB SAS disks
Oracle Solaris 11.1 (COMSTAR)

Software Configuration:

Oracle Solaris 11.1 (11.1.10.5.0), Oracle E-Business
Oracle Solaris 11.1 (11.1.10.5.0), PeopleSoft
Oracle Solaris 11.1 (11.1.9.5.0), Decision Support
Oracle Database 11g Release 2, Oracle E-Business and PeopleSoft
Oracle Database 12c Release 1, Decision Support
Oracle E-Business Suite 12.1.3
PeopleSoft Human Capital Management 9.1 FP2
PeopleSoft PeopleTools 8.52.03
Oracle Java SE 6u32
Oracle Tuxedo, Version 10.3.0.0, 64-bit, Patch Level 043
Oracle WebLogic Server 11g (10.3.4)

Oracle Dynamic Domains (PDoms) resources:


Oracle E-Business PeopleSoft Oracle DSS
Processors 8 8 16
Memory 4 TB 4 TB 8 TB
Oracle Solaris 11.1 (11.1.10.5.0) 11.1 (11.1.10.5.0) 11.1 (11.1.9.5.0)
Oracle Database 11g 11g 12c
Oracle VM for SPARC /
Oracle Solaris Zones
4 LDom / 8 Zones 2 LDom / 4 Zones None
Storage 7 x Sun Server X3-2L 1 x Sun Server X3-2L
(12 x 3 TB SAS )
2 x Sun Storage 2540-M2 / 2501 pairs
4 x Sun Storage 2540-M2/2501 pairs

Benchmark Description

This benchmark consists of three different applications running concurrently. It shows that large, enterprise workloads can be run on a single system and without performance impact between application environments.

The three workloads are:

  • Oracle E-Business Suite Online

    • This test simulates thousands of online users executing transactions typical of an internal Enterprise Resource Processing, including 5 application modules: Customer Service, Human Resources Self Service, Procurement, Order Management and Financial.

    • Each database tier uses a database instance of about 600 GB in size, and supporting thousands of application users, accessing hundreds of objects (tables, indexes, SQL stored procedures, etc.).

    • The application tier includes multiple web and application server instances, specifically Apache Web Server, Oracle Application Server 10g and Oracle Java SE 6u32.

  • PeopleSoft Human Capital Management

    • This test simulates thousands of online employees, managers and Human Resource administrators executing transactions typical of a Human Resources Self Service application for the Enterprise. Typical transactions are: viewing paychecks, promoting and hiring employees, updating employee profiles, etc.

    • The database tier uses a database instance of about 500 GB in size, containing information for 500,480 employees.

    • The application tier for this test includes web and application server instances, specifically Oracle WebLogic Server 11g, PeopleSoft Human Capital Management 9.1 and Oracle Java SE 6u32.

  • Decision Support Workload using the Oracle Database.

    • The query processes 30 billion rows stored in the Oracle Database, making heavy use of Oracle parallel query processing features. It performs multiple aggregations and summaries by reading and processing all the rows of the database.

Key Points and Best Practices

Oracle E-Business Environment

The Oracle E-Business Suite setup consisted 4 Oracle E-Business environments running 5 online Oracle E-Business modules simultaneously. The Oracle E-Business environments were deployed on 4 Oracle VM for SPARC, respectively 2 for the Application tier and 2 for the Database tier. Each LDom included 2 SPARC M6 processor chips. The Application LDom was further split into 2 Oracle Solaris Zones, each one containing one Oracle E-Business Application instance. Similarly, on the Database tier, each LDom was further divided into 2 Oracle Solaris Zones, each containing an Oracle Database instance. Applications on the same LDom shared a 10 GbE network link to connect to the Database tier LDom. Each Application in a Zone was connected to its own dedicated Database Zone. The communication between the two Zones was implemented via Oracle Solaris 11 virtual network, which provides high performance, low latency transfers at memory speed using large frames (9000 bytes vs typical 1500 bytes frames).

The Oracle E-Business setup made use of the Oracle Database Shared Server feature in order to limit memory utilization, as well as the number of database Server processes. The Oracle Database configuration and optimization was substantially out-of-the-box, except for proper sizing the Oracle Database memory areas (System Global Area and Program Global Area).

In the Oracle E-Business Application LDom handling Customer Service and HR Self Service modules, 28 Forms servers and 8 OC4J application servers were hosted in the two separate Oracle Solaris Zones, for a total of 56 forms servers and 16 applications servers.

All the Oracle Database server processes and the listener processes were executed in the Oracle Solaris FX scheduler class.

PeopleSoft Environment

The PeopleSoft Application Oracle VM for SPARC had one Oracle Solaris Zone of 12 cores containing the web tier and two Oracle Solaris Zones of 28 cores each containing the Application tier. The Database tier was contained in an Oracle VM for SPARC consisting of one Oracle Solaris Zone of 24 cores. One and a half cores, in the Application Oracle VM, were dedicated to network and disk interrupt handling.

All database data files, recovery files and Oracle Clusterware files for the PeopleSoft test were created with the Oracle Automatic Storage Management (Oracle ASM) volume manager for the added benefit of the ease of management provided by Oracle ASM integrated storage management solution.

In the application tier, 5 PeopleSoft domains with 350 application servers (70 per each domain) were hosted in the two separate Oracle Solaris Zones for a total of 10 domains with 700 application server processes.

All PeopleSoft Application processes and Web Server JVM instances were executed in the Oracle Solaris FX scheduler class.

Oracle Decision Support Environment

The decision support workload showed how the combination of a large memory (8 TB) and a large number of processors (16 chips comprising 1536 virtual CPUs) together with Oracle parallel query facility can linearly increase the performance of certain decision support queries as the number of CPUs increase.

The large memory was used to cache the entire 30 billion row Oracle table in memory. There are a number of ways to accomplish this. The method deployed in this test was to allocate sufficient memory for Oracle's "keep cache" and direct the table to the "keep cache."

To demonstrate scalability, it was necessary to ensure that the number of Oracle parallel servers was always equal to the number of available virtual CPUs. This was accomplished by the combination of providing a degree of parallelism hint to the query and setting both parallel_max_servers and parallel_min_servers to the number of virtual CPUs.

The number of virtual CPUs for each stage of the scalability test was adjusted using the psradm command available in Oracle Solaris.

See Also

Disclosure Statement

Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 09/22/2013.

Oracle E-Business Suite R12 extra-large multiple-online module benchmark, SPARC M6-32, SPARC M6, 3.6 GHz, 8 chips, 96 cores, 768 threads, 4 TB memory, 14,660 online users, average response time 0.81 sec, 90th percentile response time 0.88 sec, Oracle Solaris 11.1, Oracle Solaris Zones, Oracle VM for SPARC, Oracle E-Business Suite 12.1.3, Oracle Database 11g Release 2, Results as of 9/20/2013.

SPARC T5-2 Server Beats x86 Server on Oracle Database Transparent Data Encryption

Database security is becoming increasingly important. Oracle Database Advanced Security Transparent Data Encryption (TDE) stops would-be attackers from bypassing the database and reading sensitive information from storage by enforcing data-at-rest encryption in the database layer. Oracle's SPARC T5-2 server outperformed x86 systems when running Oracle Database 12c with Transparent Data Encryption.

  • The SPARC T5-2 server sustained more than 8.0 GB/sec of read bandwidth while decrypting using Transparent Data Encryption (TDE) in Oracle Database 12c. This was the bandwidth available on the system and matched the rate for querying the non-encrypted data.

  • The SPARC T5-2 server achieves about 1.5x higher decryption rate per socket using Oracle Database 12c with TDE than a Sun Server X4-2 system.

  • The SPARC T5-2 server achieves more than double the decryption rate per socket using Oracle Database 12c with TDE than a Sun Server X3-2 system.

Performance Landscape

Table of Size 250 GB Encrypted with AES-128-CFB
Full Table Scan with Degree of Parallelism 128
System Chips Table Data Format SPARC T5-2 Advantage
Clear Encrypted
SPARC T5-2 2 8.4 GB/sec 8.3 GB/sec 1.0
Sun Server X4-2L 2 8.2 GB/sec 5.6 GB/sec 1.5

SPARC T5-2 1 8.4 GB/sec 4.2 GB/sec 1.0
Sun Server X4-2L 1 8.2 GB/sec 2.8 GB/sec 1.5
Sun Server X3-2L 1 8.2 GB/sec 2.0 GB/sec 2.1

Configuration Summary

Systems Under Test:

SPARC T5-2
2 x SPARC T5 processors, 3.6 GHz
256 GB memory
Oracle Solaris 11.1
Oracle Database 12c

Sun Server X3-2L
2 x Intel Xeon E5-2690 processor, 2.90 GHz
64 GB memory
Oracle Solaris 11.1
Oracle Database 12c

Sun Server X4-2L
2 x Intel Xeon E5-2697 v2 processor, 2.70 GHz
256 GB memory
Oracle Solaris 11.1
Oracle Database 12c

Storage:

Flash Storage

Benchmark Description

The purpose of the benchmark is to show the query performance of a database using data encryption to keep the data secure. The benchmark creates a 250 GB table. It is loaded both into a clear text (no encryption) tablespace and an AES-128 encrypted tablespace. Full table scans of the tables were timed.

Key Points and Best Practices

The Oracle Database feature, Transparent Data Encryption (TDE), simplifies the encryption of data within datafiles, preventing unauthorized access to it from the operating system. Transparent Data Encryption allows encryption of the entire contents of a tablespace.

With hardware acceleration of the encryption routines, the SPARC T5-2 server can achieve nearly the same query rate whether the table is encrypted or not up to a limit of about 4 GB/sec per chip.

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 23 September 2013.

Wednesday Sep 25, 2013

SPARC T5-8 Delivers World Record Oracle OLAP Perf Version 3 Benchmark Result on Oracle Database 12c

Oracle's SPARC T5-8 server delivered world record query performance for systems running Oracle Database 12c for the Oracle OLAP Perf Version 3 benchmark.

  • The query throughput on the SPARC T5-8 server is 1.7x higher than that of an 8-chip Intel Xeon E7-8870 server. Both systems had sub-second average response times.

  • The SPARC T5-8 server with the Oracle Database demonstrated the ability to support at least 700 concurrent users querying OLAP cubes (with no think time), processing 2.33 million analytic queries per hour with an average response time of less than 1 second per query. This performance was enabled by keeping the entire cube in-memory utilizing the 4 TB of memory on the SPARC T5-8 server.

  • Assuming a 60 second think time between query requests, the SPARC T5-8 server can support approximately 39,450 concurrent users with the same sub-second response time.

  • The workload uses a set of realistic Business Intelligence (BI) queries that run against an OLAP cube based on a 4 billion row fact table of sales data. The 4 billion rows are partitioned by month spanning 10 years.

  • The combination of the Oracle Database 12cwith the Oracle OLAP option running on a SPARC T5-8 server supports live data updates occurring concurrently with minimally impacted user query executions.

Performance Landscape

Oracle OLAP Perf Version 3 Benchmark
Oracle cube base on 4 billion fact table rows
10 years of data partitioned by month
System Queries/
hour
Users Average Response
Time (sec)
0 sec think time 60 sec think time
SPARC T5-8 2,329,000 700 39,450 <1 sec
8-chip Intel Xeon E7-8870 1,354,000 120 22,675 <1 sec

Configuration Summary

SPARC T5-8:

1 x SPARC T5-8 server with
8 x SPARC T5 processors, 3.6 GHz
4 TB memory
Data Storage and Redo Storage
Flash Storage
Oracle Solaris 11.1 (11.1.8.2.0)
Oracle Database 12c Release 1 (12.1.0.1) with Oracle OLAP option

Sun Server X2-8:

1 x Sun Server X2-8 with
8 x Intel Xeon E7-8870 processors, 2.4 GHz
1 TB memory
Data Storage and Redo Storage
Flash Storage
Oracle Solaris 10 10/12
Oracle Database 12c Release 1 (12.1.0.1) with Oracle OLAP option

Benchmark Description

The Oracle OLAP Perf Version 3 benchmark is a workload designed to demonstrate and stress the ability of the OLAP Option to deliver fast query, near real-time updates and rich calculations using a multi-dimensional model in the context of the Oracle data warehousing.

The bulk of the benchmark entails running a number of concurrent users, each issuing typical multidimensional queries against an Oracle cube. The cube has four dimensions: time, product, customer, and channel. Each query user issues approximately 150 different queries. One query chain may ask for total sales in a particular region (e.g South America) for a particular time period (e.g. Q4 of 2010) followed by additional queries which drill down into sales for individual countries (e.g. Chile, Peru, etc.) with further queries drilling down into individual stores, etc. Another query chain may ask for yearly comparisons of total sales for some product category (e.g. major household appliances) and then issue further queries drilling down into particular products (e.g. refrigerators, stoves. etc.), particular regions, particular customers, etc.

While the core of every OLAP Perf benchmark is real world query performance, the benchmark itself offers numerous execution options such as varying data set sizes, number of users, numbers of queries for any given user and cube update frequency. Version 3 of the benchmark is executed with a much larger number of query streams than previous versions and used a cube designed for near real-time updates. The results produced by version 3 of the benchmark are not directly comparable to results produced by previous versions of the benchmark.

The near real-time update capability is implemented along the following lines. A large Oracle cube, H, is built from a 4 billion row star schema, containing data up until the end of last business day. A second small cube, D, is then created which will contain all of today's new data coming in from outside the world. It will be updated every L minutes with the data coming in within the last L minutes. A third cube, R, joins cubes H and D for reporting purposes much like a view might join data from two tables. Calculations are installed into cube R. The use of a reporting cube which draws data from different storage cubes is a common practice.

Query users are never locked out of query operations while new data is added to the update cube. The point of the demonstration is to show that an Oracle OLAP system can be designed which results in data being no more than L minutes out of date, where L may be as low as just a few minutes. This is what is meant by near real-time analytics.

Key Points and Best Practices

  • Building and querying cubes with the Oracle OLAP option requires a large temporary tablespace. Normally temporary tablespaces would reside on disk storage. However, because the SPARC T5-8 server used in this benchmark had 4 TB of main memory, it was possible to use main memory for the OLAP temporary tablespace. This was accomplished by using a temporary, memory-based file system (TMPFS) for the temporary tablespace datafiles.

  • Since typical business intelligence users are often likely to issue similar queries, either with the same or different constants in the where clauses, setting the init.ora parameter "cursor_sharing" to "force" provides for additional query throughput and a larger number of potential users.

  • Assuming the normal Oracle Database initialization parameters (e.g. SGA, PGA, processes etc.) are appropriately set, out of the box performance for the Oracle OLAP workload should be close to what is reported here. Additional performance resulted from using memory for the OLAP temporary tablespace setting "cursor_sharing" to force.

  • Oracle OLAP Cube update performance was optimized by running update processes in the FX class with a priority greater than 0.

  • The maximum lag time between updates to the source fact table and data availability to query users (what was referred to as L in the benchmark description) was less than 3 minutes for the benchmark environment on the SPARC T5-8 server.

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 09/22/2013.

SPARC T5 Encryption Performance Tops Intel E5-2600 v2 Processor

The cryptography benchmark suite was developed by Oracle to measure security performance on important AES security modes. Oracle's SPARC T5 processor with it security software in silicon is faster than x86 servers that have the AES-NI instructions. In this test, the performance of on-processor encryption operations is measured (32 KB encryptions). Multiple threads are used to measure each processors maximum throughput. The SPARC T5-8 shows dramatically faster encryption.

  • A SPARC T5 processor running Oracle Solaris 11.1 is 2.7 times faster executing AES-CFB 256-bit key encryption (in cache) than the Intel E5-2697 v2 processor (with AES-NI) running Oracle Linux 6.3. AES-CFB encryption is used by Oracle Database for Transparent Data Encryption (TDE) which provides security for database storage.

  • On the AES-CFB 128-bit key encryption, the SPARC T5 processor is 2.5 times faster than the Intel E5-2697 v2 processor (with AES-NI) running Oracle Linux 6.3 for in-cache encryption. AES-CFB mode is used by Oracle Database for Transparent Data Encryption (TDE) which provides security for database storage.

  • The IBM POWER7+ has three hardware security units for 8-core processors, but IBM has not publicly shown any measured performance results on AES-CFB or other encryption modes.

Performance Landscape

Presented below are results for running encryption using the AES cipher with the CFB, CBC, CCM and GCM modes for key sizes of 128, 192 and 256. Decryption performance was similar and is not presented. Results are presented as MB/sec (10**6).

Encryption Performance – AES-CFB

Performance is presented for in-cache AES-CFB128 mode encryption. Multiple key sizes of 256-bit, 192-bit and 128-bit are presented. The encryption was performance on 32 KB of pseudo-random data (same data for each run).

AES-CFB
Microbenchmark Performance (MB/sec)
Processor GHz Chips Performance Software Environment
AES-256-CFB
SPARC T5 3.60 2 54,396 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2697 v2 2.70 2 19,960 Oracle Linux 6.3, IPP/AES-NI
Intel E5-2690 2.90 2 12,823 Oracle Linux 6.3, IPP/AES-NI
AES-192-CFB
SPARC T5 3.60 2 61,000 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2697 v2 2.70 2 23,217 Oracle Linux 6.3, IPP/AES-NI
Intel E5-2690 2.90 2 14,928 Oracle Linux 6.3, IPP/AES-NI
AES-128-CFB
SPARC T5 3.60 2 68,695 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2697 v2 2.70 2 27,740 Oracle Linux 6.3, IPP/AES-NI
Intel E5-2690 2.90 2 17,824 Oracle Linux 6.3, IPP/AES-NI

Encryption Performance – AES-GCM

Performance is presented for in-cache AES-GCM mode encryption with authentication. Multiple key sizes of 256-bit, 192-bit and 128-bit are presented. The encryption/authentication was performance on 32 KB of pseudo-random data (same data for each run).

AES-GCM
Microbenchmark Performance (MB/sec)
Processor GHz Chips Performance Software Environment
AES-256-GCM
SPARC T5 3.60 2 34,101 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2697 v2 2.70 2 15,338 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 13,520 Oracle Linux 6.3, IPP/AES-NI
AES-192-GCM
SPARC T5 3.60 2 36,852 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2697 v2 2.70 2 15,768 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 14,159 Oracle Linux 6.3, IPP/AES-NI
AES-128-GCM
SPARC T5 3.60 2 39,003 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2697 v2 2.70 2 16,405 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 14,877 Oracle Linux 6.3, IPP/AES-NI

Encryption Performance – AES-CCM

Performance is presented for in-cache AES-CCM mode encryption with authentication. Multiple key sizes of 256-bit, 192-bit and 128-bit are presented. The encryption/authentication was performance on 32 KB of pseudo-random data (same data for each run).

AES-CCM
Microbenchmark Performance (MB/sec)
Processor GHz Chips Performance Software Environment
AES-256-CCM
SPARC T5 3.60 2 29,431 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2697 v2 2.70 2 19,447 Oracle Linux 6.3, IPP/AES-NI
Intel E5-2690 2.90 2 12,493 Oracle Linux 6.3, IPP/AES-NI
AES-192-CCM
SPARC T5 3.60 2 33,715 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2697 v2 2.70 2 22,634 Oracle Linux 6.3, IPP/AES-NI
Intel E5-2690 2.90 2 14,507 Oracle Linux 6.3, IPP/AES-NI
AES-128-CCM
SPARC T5 3.60 2 39,188 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2697 v2 2.70 2 26,951 Oracle Linux 6.3, IPP/AES-NI
Intel E5-2690 2.90 2 17,256 Oracle Linux 6.3, IPP/AES-NI

Encryption Performance – AES-CBC

Performance is presented for in-cache AES-CBC mode encryption. Multiple key sizes of 256-bit, 192-bit and 128-bit are presented. The encryption was performance on 32 KB of pseudo-random data (same data for each run).

AES-CBC
Microbenchmark Performance (MB/sec)
Processor GHz Chips Performance Software Environment
AES-256-CBC
SPARC T5 3.60 2 56,933 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2697 v2 2.70 2 19,962 Oracle Linux 6.3, IPP/AES-NI
Intel E5-2690 2.90 2 12,822 Oracle Linux 6.3, IPP/AES-NI
AES-192-CBC
SPARC T5 3.60 2 63,767 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2697 v2 2.70 2 23,224 Oracle Linux 6.3, IPP/AES-NI
Intel E5-2690 2.90 2 14,915 Oracle Linux 6.3, IPP/AES-NI
AES-128-CBC
SPARC T5 3.60 2 72,508 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2697 v2 2.70 2 27,733 Oracle Linux 6.3, IPP/AES-NI
Intel E5-2690 2.90 2 17,823 Oracle Linux 6.3, IPP/AES-NI

Configuration Summary

SPARC T5-2 server
2 x SPARC T5 processor, 3.6 GHz
512 GB memory
Oracle Solaris 11.1 SRU 4.2

Sun Server X4-2L server
2 x E5-2697 v2 processors, 2.70 GHz
256 GB memory
Oracle Linux 6.3

Sun Server X3-2 server
2 x E5-2690 processors, 2.90 GHz
128 GB memory
Oracle Linux 6.3

Benchmark Description

The benchmark measures cryptographic capabilities in terms of general low-level encryption, in-cache (32 KB encryptions) and on-chip using various ciphers, including AES-128-CFB, AES-192-CFB, AES-256-CFB, AES-128-CBC, AES-192-CBC, AES-256-CBC, AES-128-CCM, AES-192-CCM, AES-256-CCM, AES-128-GCM, AES-192-GCM and AES-256-GCM.

The benchmark results were obtained using tests created by Oracle which use various application interfaces to perform the various ciphers. They were run using optimized libraries for each platform to obtain the best possible performance.

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 9/23/2013.

Tuesday Sep 10, 2013

Oracle ZFS Storage ZS3-4 Delivers World Record SPC-2 Performance

The Oracle Storage ZS3-4 storage system delivered a world record performance result for the SPC-2 benchmark along with excellent price-performance.

  • The Oracle Storage ZS3-4 storage system delivered an overall score of 17,244.22 SPC-2 MBPS™ and a SPC-2 price-performance of $22.53 on the SPC-2 benchmark.

  • This is over a 1.6X generational improvement in performance and over a 1.5X generational improvement in price-performance than over Oracle's Sun ZFS Storage 7420 SPC-2 Benchmark results.

  • The Oracle ZFS Storage ZS3-4 storage system has 6.8X better overall throughput and nearly 1.2X better price-performance than the IBM DS3524 Express turbo, which is IBM's best overall price-performance score on the SPC-2 benchmark.

  • The Oracle ZFS Storage ZS3-4 storage system has over 1.1X overall throughput and 5.8X better price-performance than the IBM DS8870, which is IBM's best overall performance score on the SPC-2 benchmark.

  • The Oracle ZFS Storage ZS3-4 storage system has over 1.3X overall throughput and 3.9X better price-performance than the HP StorageWorks P9500XP Disk Array on the SPC-2 benchmark.

Performance Landscape

SPC-2 Performance Chart (in decreasing performance order)

System SPC-2
MB/s
$/SPC-2
MB/s
ASU
Capacity
(GB)
TSC Price Data
Protection
Level
Date Results
Identifier
Oracle ZFS Storage ZS3-4 17,244.22 $22.53 31,611 $388,472 Mirroring 09/10/13 B00067
Fujitsu DX8700 S2 16,039 $79.51 71,404 $1,275,163 Mirroring 12/03/12 B00063
IBM DS8870 15,424 $131.21 30,924 $2,023,742 RAID-5 10/03/12 B00062
IBM SAN VC v6.4 14,581 $129.14 74,492 $1,883,037 RAID-5 08/01/12 B00061
NEC Storage M700 14,409 $25.13 53,550 $361,613 Mirroring 08/19/12 B00066
Hitachi VSP 13,148 $95.38 129,112 $1,254,093 RAID-5 07/27/12 B00060
HP StorageWorks P9500 13,148 $88.34 129,112 $1,161,504 RAID-5 03/07/12 B00056
Sun ZFS Storage 7420 10,704 $35.24 31,884 $377,225 Mirroring 04/12/12 B00058
IBM DS8800 9,706 $270.38 71,537 $2,624,257 RAID-5 12/01/10 B00051
HP XP24000 8,725 $187.45 18,401 $1,635,434 Mirroring 09/08/08 B00035

SPC-2 MB/s = the Performance Metric
$/SPC-2 MB/s = the Price-Performance Metric
ASU Capacity = the Capacity Metric
Data Protection = Data Protection Metric
TSC Price = Total Cost of Ownership Metric
Results Identifier = A unique identification of the result Metric

SPC-2 Price-Performance Chart (in increasing price-performance order)

System SPC-2
MB/s
$/SPC-2
MB/s
ASU
Capacity
(GB)
TSC Price Data
Protection
Level
Date Results
Identifier
SGI InfiniteStorage 5600 8,855.70 $15.97 28,748 $141,393 RAID6 03/06/13 B00065
Oracle ZFS Storage ZS3-4 17,244.22 $22.53 31,611 $388,472 Mirroring 09/10/13 B00067
Sun Storage J4200 548.80 $22.92 11,995 $12,580 Unprotected 07/10/08 B00033
NEC Storage M700 14,409 $25.13 53,550 $361,613 Mirroring 08/19/12 B00066
Sun Storage J4400 887.44 $25.63 23,965 $22,742 Unprotected 08/15/08 B00034
Sun StorageTek 2530 672.05 $26.15 1,451 $17,572 RAID5 08/16/07 B00026
Sun StorageTek 2530 663.51 $26.48 854 $17,572 Mirroring 08/16/07 B00025
Fujitsu ETERNUS DX80 1,357.55 $26.70 4,681 $36,247 Mirroring 03/15/10 B00050
IBM DS3524 Express Turbo 2,510 $26.76 14,374 $67,185 RAID-5 12/31/10 B00053
Fujitsu ETERNUS DX80 S2 2,685.50 $28.48 17,231 $76,475 Mirroring 08/19/11 B00055

SPC-2 MB/s = the Performance Metric
$/SPC-2 MB/s = the Price-Performance Metric
ASU Capacity = the Capacity Metric
Data Protection = Data Protection Metric
TSC Price = Total Cost of Ownership Metric
Results Identifier = A unique identification of the result Metric

Complete SPC-2 benchmark results may be found at http://www.storageperformance.org/results/benchmark_results_spc2.

Configuration Summary

Storage Configuration:

Oracle ZFS Storage ZS3-4 storage system in clustered configuration
2 x Oracle ZFS Storage ZS3-4 controllers, each with
4 x 2.4 GHz 10-core Intel Xeon processors
1024 GB memory
16 x Sun Disk shelves, each with
24 x 300 GB 15K RPM SAS-2 drives

Benchmark Description

SPC Benchmark-2 (SPC-2): Consists of three distinct workloads designed to demonstrate the performance of a storage subsystem during the execution of business critical applications that require the large-scale, sequential movement of data. Those applications are characterized predominately by large I/Os organized into one or more concurrent sequential patterns. A description of each of the three SPC-2 workloads is listed below as well as examples of applications characterized by each workload.

  • Large File Processing: Applications in a wide range of fields, which require simple sequential process of one or more large files such as scientific computing and large-scale financial processing.
  • Large Database Queries: Applications that involve scans or joins of large relational tables, such as those performed for data mining or business intelligence.
  • Video on Demand: Applications that provide individualized video entertainment to a community of subscribers by drawing from a digital film library.

SPC-2 is built to:

  • Provide a level playing field for test sponsors.
  • Produce results that are powerful and yet simple to use.
  • Provide value for engineers as well as IT consumers and solution integrators.
  • Is easy to run, easy to audit/verify, and easy to use to report official results.

See Also

Disclosure Statement

SPC-2 and SPC-2 MBPS are registered trademarks of Storage Performance Council (SPC). Results as of September 10, 2013, for more information see www.storageperformance.org. Oracle ZFS Storage ZS3-4 B00067, Fujitsu ET 8700 S2 B00063, IBM DS8870 B00062, IBM S.V.C 6.4 B00061, NEC Storage M700 B00066, Hitachi VSP B00060, HP P9500 XP Disk Array B00056, IBM DS8800 B00051.

Oracle ZFS Storage ZS3-4 Produces Best 2-Node Performance on SPECsfs2008 NFSv3

The Oracle ZFS Storage ZS3-4 storage system delivered world record two-node performance on the SPECsfs2008 NFSv3 benchmark, beating results published on NetApp's dual-controller and four-node high-end FAS6240 storage systems.

  • The Oracle ZFS Storage ZS3-4 storage system delivered a world record two-node result of 450,702 SPECsfs2008_nfs.v3 Ops/sec with an Overall Response Time (ORT) of 0.70 msec on the SPECsfs2008 NFSv3 benchmark.

  • The Oracle ZFS Storage ZS3-4 storage system delivered 2.4x higher throughput than the dual-controller NetApp FAS6240 and 4.5x higher throughput than the dual-controller NetApp FAS3270 on the SPECsfs2008_nfs.v3 benchmark at less than half the list price of either result.

  • The Oracle ZFS Storage ZS3-4 storage system had 42 percent higher throughput than the four-node NetApp FAS6240 on the SPECsfs2008 NFSv3 benchmark.

  • The Oracle ZFS Storage ZS3-4 storage aystem has 54 percent better Overall Response Time than the 4-node NetApp FAS6240 on the SPECsfs2008 NFSv3 benchmark.

Performance Landscape

Two node results for SPECsfs2008_nfs.v3 presented (in decreasing SPECsfs2008_nfs.v3 Ops/sec order) along with other select results.

Sponsor System Nodes Disks Throughput
(Ops/sec)
Overall Response
Time (msec)
Oracle ZS3-4 2 464 450,702 0.70
IBM SONAS 1.2 2 1975 403,326 3.23
NetApp FAS6240 4 288 260,388 1.53
NetApp FAS6240 2 288 190,675 1.17
EMC VG8 312 135,521 1.92
Oracle 7320 2 136 134,140 1.51
EMC NS-G8 100 110,621 2.32
NetApp FAS3270 2 360 101,183 1.66

Throughput SPECsfs2008_nfs.v3 Ops/sec — the Performance Metric
Overall Response Time — the corresponding Response Time Metric
Nodes — Nodes and Controllers are being used interchangeably

Complete SPECsfs2008 benchmark results may be found at http://www.spec.org/sfs2008/results/sfs2008.html.

Configuration Summary

Storage Configuration:

Oracle ZFS Storage ZS3-4 storage system in clustered configuration
2 x Oracle ZFS Storage ZS3-4 controllers, each with
8 x 2.4 GHz Intel Xeon E7-4870 processors
2 TB memory
2 x 10GbE NICs
20 x Sun Disk shelves
18 x shelves with 24 x 300 GB 15K RPM SAS-2 drives
2 x shelves with 20 x 300 GB 15K RPM SAS-2 drives and 8 x 73 GB SAS-2 flash-enabled write-cache

Benchmark Description

SPECsfs2008 is the latest version of the Standard Performance Evaluation Corporation (SPEC) benchmark suite measuring file server throughput and response time, providing a standardized method for comparing performance across different vendor platforms. SPECsfs2008 results summarize the server's capabilities with respect to the number of operations that can be handled per second, as well as the overall latency of the operations. The suite is a follow-on to the SFS97_R1 benchmark, adding a CIFS workload, an updated NFSv3 workload, support for additional client platforms, and a new test harness and reporting/submission framework.

See Also

Disclosure Statement

SPEC and SPECsfs are registered trademarks of Standard Performance Evaluation Corporation (SPEC). Results as of September 10, 2013, for more information see www.spec.org. Oracle ZFS Storage ZS3-4 Appliance 450,702 SPECsfs2008_nfs.v3 Ops/sec, 0.70 msec ORT, NetApp Data ONTAP 8.1 Cluster-Mode (4-node FAS6240) 260,388 SPECsfs2008_nfs.v3 Ops/Sec, 1.53 msec ORT, NetApp FAS6240 190,675 SPECsfs2008_nfs.v3 Ops/Sec, 1.17 msec ORT. NetApp FAS3270 101,183 SPECsfs2008_nfs.v3 Ops/Sec, 1.66 msec ORT.

Nodes refer to the item in the SPECsfs2008 disclosed Configuration Bill of Materials that have the Processing Elements that perform the NFS Processing Function. These are the first item listed in each of disclosed Configuration Bill of Materials except for EMC where it is both the first and third items listed, and HP, where it is the second item listed as Blade Servers. The number of nodes is from the QTY disclosed in the Configuration Bill of Materials as described above. Configuration Bill of Materials list price for Oracle result of US$ 423,644. Configuration Bill of Materials list price for NetApp FAS3270 result of US$ 1,215,290. Configuration Bill of Materials list price for NetApp FAS6240 result of US$ 1,028,118. Oracle pricing from https://shop.oracle.com/pls/ostore/f?p=dstore:home:0, traverse to "Storage and Tape" and then to "NAS Storage". NetApp's pricing from http://www.netapp.com/us/media/na-list-usd-netapp-custom-state-new-discounts.html.

Oracle ZFS Storage ZS3-2 Beats Comparable NetApp on SPECsfs2008 NFSv3

Oracle ZFS Storage ZS3-2 storage system delivered outstanding performance on the SPECsfs2008 NFSv3 benchmark, beating results published on NetApp's fastest midrange platform, the NetApp FAS3270, the NetApp FAS6240 and the EMC Gateway NS-G8 Server Failover Cluster.

  • The Oracle ZFS Storage ZS3-2 storage system delivered 210,535 SPECsfs2008_nfs.v3 Ops/sec with an Overall Response Time (ORT) of 1.12 msec on the SPECsfs2008 NFSv3 benchmark.

  • The Oracle ZFS Storage ZS3-2 storage system delivered 10% higher throughput than the NetApp FAS6240 on the SPECsfs2008 NFSv3 benchmark.

  • The Oracle ZFS Storage ZS3-2 storage system has 52% higher throughput than the NetApp FAS3270 on the SPECsfs2008 NFSv3 benchmark.

  • The Oracle ZFS Storage ZS3-2 storage system has 5% better Overall Response Time than the NetApp FAS6240 on the SPECsfs2008 NFSv3 benchmark.

  • The Oracle ZFS Storage ZS3-2 storage system has 33% better Overall Response Time than the NetApp FAS3270 on the SPECsfs2008 NFSv3 benchmark.

Performance Landscape

Results for SPECsfs2008 NFSv3 (in decreasing SPECsfs2008_nfs.v3 Ops/sec order) for competitive systems.

Sponsor System Throughput
(Ops/sec)
Overall Response
Time (msec)
Oracle ZS3-2 210,535 1.12
NetApp FAS6240 190,675 1.17
EMC VG8 135,521 1.92
EMC NS-G8 110,621 2.32
NetApp FAS3270 101,183 1.66
NetApp FAS3250 100,922 1.76

Throughput SPECsfs2008_nfs.v3 Ops/sec = the Performance Metric
Overall Response Time = the corresponding Response Time Metric

Complete SPECsfs2008 benchmark results may be found at http://www.spec.org/sfs2008/results/sfs2008.html.

Configuration Summary

Storage Configuration:

Oracle ZFS Storage ZS3-2 storage system in clustered configuration
2 x Oracle ZFS Storage ZS3-2 controllers, each with
4 x 2.1 GHz Intel Xeon E5-2658 processors
512 GB memory
8 x Sun Disk shelves
3 x shelves with 24 x 900 GB 10K RPM SAS-2 drives
3 x shelves with 20 x 900 GB 10K RPM SAS-2 drives
2 x shelves with 20 x 900 GB 10K RPM SAS-2 drives and 4 x 73 GB SAS-2 flash-enabled write-cache

Benchmark Description

SPECsfs2008 is the latest version of the Standard Performance Evaluation Corporation (SPEC) benchmark suite measuring file server throughput and response time, providing a standardized method for comparing performance across different vendor platforms. SPECsfs2008 results summarize the server's capabilities with respect to the number of operations that can be handled per second, as well as the overall latency of the operations. The suite is a follow-on to the SFS97_R1 benchmark, adding a CIFS workload, an updated NFSv3 workload, support for additional client platforms, and a new test harness and reporting/submission framework.

 

See Also

Disclosure Statement

SPEC and SPECsfs are registered trademarks of Standard Performance Evaluation Corporation (SPEC). Results as of September 10, 2013, for more information see www.spec.org. Oracle ZFS Storage ZS3-2 Appliance 210,535 SPECsfs2008_nfs.v3 Ops/sec, 1.12 msec ORT, NetApp FAS6240 190,675 SPECsfs2008_nfs.v3 Ops/Sec, 1.17 msec ORT, EMC Celerra VG8 Server Failover Cluster, 2 Data Movers (1 stdby) / Symmetrix VMAX 135,521 SPECsfs2008_nfs.v3 Ops/Sec, 1.92 msec ORT, EMC Celerra Gateway NS-G8 Server Failover Cluster, 3 Datamovers (1 stdby) / Symmetrix V-Max 110,621 SPECsfs2008_nfs.v3 Ops/Sec, 2.32 msec ORT. NetApp FAS3270 101,183 SPECsfs2008_nfs.v3 Ops/Sec, 1.66 msec ORT. NetApp FAS3250 100,922 SPECsfs2008_nfs.v3 Ops/Sec, 1.76 msec ORT.

Wednesday Jun 12, 2013

SPARC T5-4 Produces World Record Single Server TPC-H @3000GB Benchmark Result

Oracle's SPARC T5-4 server delivered world record single server performance of 409,721 QphH@3000GB with price/performance of $3.94/QphH@3000GB on the TPC-H @3000GB benchmark. This result shows that the 4-chip SPARC T5-4 server is significantly faster than the 8-chip server results from IBM (POWER7 based) and HP (Intel x86 based).

This result demonstrates a complete data warehouse solution that shows the performance both of individual and concurrent query processing streams, faster loading, and refresh of the data during business operations. The SPARC T5-4 server delivers superior performance and cost efficiency when compared to the IBM POWER7 result.

  • The SPARC T5-4 server with four SPARC T5 processors is 2.1 times faster than the IBM Power 780 server with eight POWER7 processors and 2.5 times faster than the HP ProLiant DL980 G7 server with eight x86 processors on the TPC-H @3000GB benchmark. The SPARC T5-4 server also delivered better performance per core than these eight processor systems from IBM and HP.

  • The SPARC T5-4 server with four SPARC T5 processors is 2.1 times faster than the IBM Power 780 server with eight POWER7 processors on the TPC-H @3000GB benchmark.

  • The SPARC T5-4 server costs 38% less per $/QphH@3000GB compared to the IBM Power 780 server with the TPC-H @3000GB benchmark.

  • The SPARC T5-4 server took 2 hours, 6 minutes, 4 seconds for data loading while the IBM Power 780 server took 2.8 times longer.

  • The SPARC T5-4 server executed the first refresh function (RF1) in 19.4 seconds, the IBM Power 780 server took 7.6 times longer.

  • The SPARC T5-4 server with four SPARC T5 processors is 2.5 times faster than the HP ProLiant DL980 G7 server with the same number of cores on the TPC-H @3000GB benchmark.

  • The SPARC T5-4 server took 2 hours, 6 minutes, 4 seconds for data loading while the HP ProLiant DL980 G7 server took 4.1 times longer.

  • The SPARC T5-4 server executed the first refresh function (RF1) in 19.4 seconds, the HP ProLiant DL980 G7 server took 8.9 times longer.

  • The SPARC T5-4 server delivered 6% better performance than the SPARC Enterprise M9000-64 server and 2.1 times better than the SPARC Enterprise M9000-32 server on the TPC-H @3000GB benchmark.

Performance Landscape

The table lists the leading TPC-H @3000GB results for non-clustered systems.

TPC-H @3000GB, Non-Clustered Systems
System
Processor
P/C/T – Memory
Composite
(QphH)
$/perf
($/QphH)
Power
(QppH)
Throughput
(QthH)
Database Available
SPARC T5-4
3.6 GHz SPARC T5
4/64/512 – 2048 GB
409,721.8 $3.94 345,762.7 485,512.1 Oracle 11g R2 09/24/13
SPARC Enterprise M9000
3.0 GHz SPARC64 VII+
64/256/256 – 1024 GB
386,478.3 $18.19 316,835.8 471,428.6 Oracle 11g R2 09/22/11
SPARC T4-4
3.0 GHz SPARC T4
4/32/256 – 1024 GB
205,792.0 $4.10 190,325.1 222,515.9 Oracle 11g R2 05/31/12
SPARC Enterprise M9000
2.88 GHz SPARC64 VII
32/128/256 – 512 GB
198,907.5 $15.27 182,350.7 216,967.7 Oracle 11g R2 12/09/10
IBM Power 780
4.1 GHz POWER7
8/32/128 – 1024 GB
192,001.1 $6.37 210,368.4 175,237.4 Sybase 15.4 11/30/11
HP ProLiant DL980 G7
2.27 GHz Intel Xeon X7560
8/64/128 – 512 GB
162,601.7 $2.68 185,297.7 142,685.6 SQL Server 2008 10/13/10

P/C/T = Processors, Cores, Threads
QphH = the Composite Metric (bigger is better)
$/QphH = the Price/Performance metric in USD (smaller is better)
QppH = the Power Numerical Quantity
QthH = the Throughput Numerical Quantity

The following table lists data load times and refresh function times during the power run.

TPC-H @3000GB, Non-Clustered Systems
Database Load & Database Refresh
System
Processor
Data Loading
(h:m:s)
T5
Advan
RF1
(sec)
T5
Advan
RF2
(sec)
T5
Advan
SPARC T5-4
3.6 GHz SPARC T5
02:06:04 1.0x 19.4 1.0x 22.4 1.0x
IBM Power 780
4.1 GHz POWER7
05:51:50 2.8x 147.3 7.6x 133.2 5.9x
HP ProLiant DL980 G7
2.27 GHz Intel Xeon X7560
08:35:17 4.1x 173.0 8.9x 126.3 5.6x

Data Loading = database load time
RF1 = power test first refresh transaction
RF2 = power test second refresh transaction
T5 Advan = the ratio of time to T5 time

Complete benchmark results found at the TPC benchmark website http://www.tpc.org.

Configuration Summary and Results

Hardware Configuration:

SPARC T5-4 server
4 x SPARC T5 processors (3.6 GHz total of 64 cores, 512 threads)
2 TB memory
2 x internal SAS (2 x 300 GB) disk drives

External Storage:

12 x Sun Storage 2540-M2 array with Sun Storage 2501-M2 expansion trays, each with
24 x 15K RPM 300 GB drives, 2 controllers, 2 GB cache
2 x Brocade 6510 Fibre Channel Switches (48 x 16 Gbs port each)

Software Configuration:

Oracle Solaris 11.1
Oracle Database 11g Release 2 Enterprise Edition

Audited Results:

Database Size: 3000 GB (Scale Factor 3000)
TPC-H Composite: 409,721.8 QphH@3000GB
Price/performance: $3.94/QphH@3000GB
Available: 09/24/2013
Total 3 year Cost: $1,610,564
TPC-H Power: 345,762.7
TPC-H Throughput: 485,512.1
Database Load Time: 2:06:04

Benchmark Description

The TPC-H benchmark is a performance benchmark established by the Transaction Processing Council (TPC) to demonstrate Data Warehousing/Decision Support Systems (DSS). TPC-H measurements are produced for customers to evaluate the performance of various DSS systems. These queries and updates are executed against a standard database under controlled conditions. Performance projections and comparisons between different TPC-H Database sizes (100GB, 300GB, 1000GB, 3000GB, 10000GB, 30000GB and 100000GB) are not allowed by the TPC.

TPC-H is a data warehousing-oriented, non-industry-specific benchmark that consists of a large number of complex queries typical of decision support applications. It also includes some insert and delete activity that is intended to simulate loading and purging data from a warehouse. TPC-H measures the combined performance of a particular database manager on a specific computer system.

The main performance metric reported by TPC-H is called the TPC-H Composite Query-per-Hour Performance Metric (QphH@SF, where SF is the number of GB of raw data, referred to as the scale factor). QphH@SF is intended to summarize the ability of the system to process queries in both single and multiple user modes. The benchmark requires reporting of price/performance, which is the ratio of the total HW/SW cost plus 3 years maintenance to the QphH. A secondary metric is the storage efficiency, which is the ratio of total configured disk space in GB to the scale factor.

Key Points and Best Practices

  • Twelve of Oracle's Sun Storage 2540-M2 arrays with Sun Storage 2501-M2 expansion trays were used for the benchmark. Each contains 24 15K RPM drives and is connected to a single dual port 16Gb FC HBA using 2 ports through a Brocade 6510 Fibre Channel switch.

  • The SPARC T5-4 server achieved a peak IO rate of 33 GB/sec from the Oracle database configured with this storage.

  • Oracle Solaris 11.1 required very little system tuning.

  • Some vendors try to make the point that storage ratios are of customer concern. However, storage ratio size has more to do with disk layout and the increasing capacities of disks – so this is not an important metric when comparing systems.

  • The SPARC T5-4 server and Oracle Solaris efficiently managed the system load of two thousand Oracle Database parallel processes.

  • Six Sun Storage 2540-M2/2501-M2 arrays were mirrored to another six Sun Storage 2540-M2/25001-M2 arrays on which all of the Oracle database files were placed. IO performance was high and balanced across all the arrays.

  • The TPC-H Refresh Function (RF) simulates periodical refresh portion of Data Warehouse by adding new sales and deleting old sales data. Parallel DML (parallel insert and delete in this case) and database log performance are a key for this function and the SPARC T5-4 server outperformed both the IBM POWER7 server and HP ProLiant DL980 G7 server. (See the RF columns above.)

See Also

Disclosure Statement

TPC-H, QphH, $/QphH are trademarks of Transaction Processing Performance Council (TPC). For more information, see www.tpc.org, results as of 6/7/13. Prices are in USD. SPARC T5-4 www.tpc.org/3288; SPARC T4-4 www.tpc.org/3278; SPARC Enterprise M9000 www.tpc.org/3262; SPARC Enterprise M9000 www.tpc.org/3258; IBM Power 780 www.tpc.org/3277; HP ProLiant DL980 www.tpc.org/3285. 

Friday Mar 29, 2013

SPARC T5 System Performance for Encryption Microbenchmark

The cryptography benchmark suite was internally developed by Oracle to measure the maximum throughput of in-memory, on-chip encryption operations that a system can perform. Multiple threads are used to achieve the maximum throughput. Systems powered by Oracle's SPARC T5 processor show outstanding performance on the tested encryption operations, beating Intel processor based systems.

  • A SPARC T5 processor running Oracle Solaris 11.1 runs from 2.4x to 4.4x faster on AES 256-bit key encryption than the Intel E5-2690 processor running in-memory encryption of 32 KB blocks using CFB128, CBC, CCM and GCM modes fully hardware subscribed.

  • AES CFB mode is used by the Oracle Database 11g for Transparent Data Encryption (TDE) which provides security to database storage.

Performance Landscape

Presented below are results for running encryption using the AES cipher with the CFB, CBC, CCM and GCM modes for key sizes of 128, 192 and 256. Decryption performance was similar and is not presented. Results are presented as MB/sec (10**6).

Encryption Performance – AES-CFB

Performance is presented for in-memory AES-CFB128 mode encryption. Multiple key sizes of 256-bit, 192-bit and 128-bit are presented. The encryption was performance on 32 KB of pseudo-random data (same data for each run).

AES-CFB
Microbenchmark Performance (MB/sec)
Processor GHz Chips Performance Software Environment
AES-256-CFB
SPARC T5 3.60 2 54,396 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 12,823 IPP/AES-NI
AES-192-CFB
SPARC T5 3.60 2 61,000 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 14,928 IPP/AES-NI
AES-128-CFB
SPARC T5 3.60 2 68,695 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 17,824 IPP/AES-NI

Encryption Performance – AES-CBC

Performance is presented for in-memory AES-CBC mode encryption. Multiple key sizes of 256-bit, 192-bit and 128-bit are presented. The encryption was performance on 32 KB of pseudo-random data (same data for each run).

AES-CBC
Microbenchmark Performance (MB/sec)
Processor GHz Chips Performance Software Environment
AES-256-CBC
SPARC T5 3.60 2 56,933 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 12,822 IPP/AES-NI
AES-192-CBC
SPARC T5 3.60 2 63,767 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 14,915 IPP/AES-NI
AES-128-CBC
SPARC T5 3.60 2 72,508 Oracle Solaris 11.1, libsoftcrypto + libumem
SPARC T4 2.85 2 31,085 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel X5690 3.47 2 20,721 IPP/AES-NI
Intel E5-2690 2.90 2 17,823 IPP/AES-NI

Encryption Performance – AES-CCM

Performance is presented for in-memory AES-CCM mode encryption with authentication. Multiple key sizes of 256-bit, 192-bit and 128-bit are presented. The encryption/authentication was performance on 32 KB of pseudo-random data (same data for each run).

AES-CCM
Microbenchmark Performance (MB/sec)
Processor GHz Chips Performance Software Environment
AES-256-CCM
SPARC T5 3.60 2 29,431 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 12,493 IPP/AES-NI
AES-192-CCM
SPARC T5 3.60 2 33,715 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 14,507 IPP/AES-NI
AES-128-CCM
SPARC T5 3.60 2 39,188 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 17,256 IPP/AES-NI

Encryption Performance – AES-GCM

Performance is presented for in-memory AES-GCM mode encryption with authentication. Multiple key sizes of 256-bit, 192-bit and 128-bit are presented. The encryption/authentication was performance on 32 KB of pseudo-random data (same data for each run).

AES-GCM
Microbenchmark Performance (MB/sec)
Processor GHz Chips Performance Software Environment
AES-256-GCM
SPARC T5 3.60 2 34,101 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 13,520 IPP/AES-NI
AES-192-GCM
SPARC T5 3.60 2 36,852 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 14,159 IPP/AES-NI
AES-128-GCM
SPARC T5 3.60 2 39,003 Oracle Solaris 11.1, libsoftcrypto + libumem
Intel E5-2690 2.90 2 14,877 IPP/AES-NI

Configuration Summary

SPARC T5-2 server
2 x SPARC T5 processor, 3.6 GHz
512 GB memory
Oracle Solaris 11.1 SRU 4.2

Sun Server X3-2 server
2 x E5-2690 processors, 2.90 GHz
128 GB memory

Benchmark Description

The benchmark measures cryptographic capabilities in terms of general low-level encryption, in-memory and on-chip using various ciphers, including AES-128-CFB, AES-192-CFB, AES-256-CFB, AES-128-CBC, AES-192-CBC, AES-256-CBC, AES-128-CCM, AES-192-CCM, AES-256-CCM, AES-128-GCM, AES-192-GCM and AES-256-GCM.

The benchmark results were obtained using tests created by Oracle which use various application interfaces to perform the various ciphers. They were run using optimized libraries for each platform to obtain the best possible performance.

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 3/26/2013.

Tuesday Mar 26, 2013

SPARC T5-8 Produces TPC-C Benchmark Single-System World Record Performance

Oracle's SPARC T5-8 server equipped with eight 3.6 GHz SPARC T5 processors obtained a result of 8,552,523 tpmC on the TPC-C benchmark. This result is a world record for single servers. Oracle demonstrated this world record database performance running Oracle Database 11g Release 2 Enterprise Edition with Partitioning.

  • The SPARC T5-8 server delivered a single system TPC-C world record of 8,552,523 tpmC with a price performance of $0.55/tpmC using Oracle Database 11g Release 2. This configuration is available 09/25/13.

  • The SPARC T5-8 server has 2.8x times better performance than the 4-processor IBM x3850 X5 system equipped with Intel Xeon processors.

  • The SPARC T5-8 server delivers 1.7x the performance compared to the next best eight processor result.

  • The SPARC T5-8 server delivers 2.4x the performance per chip compared to the IBM Power 780 3-node cluster result.

  • The SPARC T5-8 server delivers 1.8x the performance per chip compared to the IBM Power 780 non-clustered result.

  • The SPARC T5-8 server delivers 1.4x the performance per chip compared to the IBM Flex x240 Xeon result.

  • The SPARC T5-8 server delivers 1.7x the performance per chip compared to the Sun Server X2-8 system equipped with Intel Xeon processors.

  • The SPARC T5-8 server demonstrated over 3.1 Million 4KB IOP/sec with 76% idle, in a separate IO intensive workload, demonstrating its ability process a large IO workload with lots of processing headroom.

  • This result showed Oracle's integrated hardware and software stacks provide industry leading performance.

  • The Oracle solution utilized Oracle Solaris 11.1 with Oracle Database 11g Enterprise Edition with Partitioning and demonstrates stability and performance with this highly secure operating environment to produce the world record TPC-C benchmark performance.

Performance Landscape

Select TPC-C results (sorted by tpmC, bigger is better)

System p/c/t tpmC Price
/tpmC
Avail Database Memory
Size
IBM Power 780 Cluster 24/192/768 10,366,254 1.38 USD 10/13/2010 IBM DB2 9.7 6 TB
SPARC T5-8 8/128/1024 8,552,523 0.55 USD 9/25/2013 Oracle 11g R2 4 TB
IBM Power 595 32/64/128 6,085,166 2.81 USD 12/10/2008 IBM DB2 9.5 4 TB
Sun Server X2-8 8/80/160 5,055,888 0.89 USD 7/10/2012 Oracle 11g R2 4 TB
IBM x3850 X5 4/40/80 3,014,684 0.59 USD 7/11/2011 IBM DB2 9.7 3 TB
IBM Flex x240 2/16/32 1,503,544 0.53 USD 8/16/2012 IBM DB2 9.7 768 GB
IBM Power 780 2/8/32 1,200,011 0.69 USD 10/13/2010 IBM DB2 9.5 512 GB

p/c/t - processors, cores, threads
Avail - availability date

Oracle and IBM TPC-C Response times

System tpmC Response Time (sec)
New Order 90th%
Response Time (sec)
New Order Average
IBM Power 780 Cluster 10,366,254 2.100 1.137
SPARC T5-8 8,552,523 0.410 0.234
IBM Power 595 6,085,166 1.690 1.220
IBM Power 780 1,200,011 0.694 0.403

Oracle uses Response Time New Order Average and Response Time New Order 90th% for comparison between Oracle and IBM.

Graphs of Oracle's and IBM's Response Time New Order Average and Response Time New Order 90th% can be found in the full disclosure reports on TPC's website TPC-C Official Result Page.

Configuration Summary and Results

Hardware Configuration:

Server
SPARC T5-8
8 x 3.6 GHz SPARC T5
4 TB memory
2 x 600 GB 10K RPM SAS2 internal disks
12 x 8 Gbs FC HBA

Data Storage
54 x Sun Server X3-2L systems configured as COMSTAR heads, each with
2 x 2.4 GHz Intel Xeon E5-2609 processors
16 GB memory
4 x Sun Flash Accelerator F40 PCIe Cards (400 GB each)
12 x 3 TB 7.2K RPM 3.5" SAS disks
2 x 600 GB 10K RPM SAS2 disks
2 x Brocade 6510 switches

Redo Storage
2 x Sun Server X3-2L systems configured as COMSTAR heads, each with
2 x 2.4 GHz Intel Xeon E5-2609 processors
16 GB memory
12 x 3 TB 7.2K RPM 3.5" SAS disks
2 x 600 GB 10K RPM SAS2 disks

Clients
16 x Sun Server X3-2 servers, each with
2 x 2.9 GHz Intel Xeon E5-2690 processors
64 GB memory
2 x 600 GB 10K RPM SAS2 disks

Software Configuration:

Oracle Solaris 11.1 SRU 4.5 (for SPARC T5-8)
Oracle Solaris 11.1 (for COMSTAR systems)
Oracle Database 11g Release 2 Enterprise Edition with Partitioning
Oracle iPlanet Web Server 7.0 U5
Oracle Tuxedo CFS-R

Results:

System: SPARC T5-8
tpmC: 8,552,523
Price/tpmC: 0.55 USD
Available: 9/25/2013
Database: Oracle Database 11g
Cluster: no
Response Time New Order Average: 0.234 seconds

Benchmark Description

TPC-C is an OLTP system benchmark. It simulates a complete environment where a population of terminal operators executes transactions against a database. The benchmark is centered around the principal activities (transactions) of an order-entry environment. These transactions include entering and delivering orders, recording payments, checking the status of orders, and monitoring the level of stock at the warehouses.

Key Points and Best Practices

  • Oracle Database 11g Release 2 Enterprise Edition with Partitioning scales easily to this high level of performance.

  • COMSTAR (Common Multiprotocol SCSI Target) is the software framework that enables an Oracle Solaris host to serve as a SCSI Target platform. COMSTAR uses a modular approach to break the huge task of handling all the different pieces in a SCSI target subsystem into independent functional modules which are glued together by the SCSI Target Mode Framework (STMF). The modules implementing functionality at SCSI level (disk, tape, medium changer etc.) are not required to know about the underlying transport. And the modules implementing the transport protocol (FC, iSCSI, etc.) are not aware of the SCSI-level functionality of the packets they are transporting. The framework hides the details of allocation providing execution context and cleanup of SCSI commands and associated resources and simplifies the task of writing the SCSI or transport modules.

  • Oracle iPlanet Web Server middleware is used for the client tier of the benchmark. Each web server instance supports more than a quarter-million users while satisfying the response time requirement from the TPC-C benchmark.

See Also

Disclosure Statement

TPC Benchmark C, tpmC, and TPC-C are trademarks of the Transaction Processing Performance Council (TPC). SPARC T5-8 (8/128/1024) with Oracle Database 11g Release 2 Enterprise Edition with Partitioning, 8,552,523 tpmC, $0.55 USD/tpmC, available 9/25/2013. IBM Power 780 Cluster (24/192/768) with DB2 ESE 9.7, 10,366,254 tpmC, $1.38 USD/tpmC, available 10/13/2010. IBM x3850 X5 (4/40/80) with DB2 ESE 9.7, 3,014,684 tpmC, $0.59 USD/tpmC, available 7/11/2011. IBM x3850 X5 (4/32/64) with DB2 ESE 9.7, 2,308,099 tpmC, $0.60 USD/tpmC, available 5/20/2011. IBM Flex x240 (2/16/32) with DB2 ESE 9.7, 1,503,544 tpmC, $0.53 USD/tpmC, available 8/16/2012. IBM Power 780 (2/8/32) with IBM DB2 9.5, 1,200,011 tpmC, $0.69 USD/tpmC, available 10/13/2010. Source: http://www.tpc.org/tpcc, results as of 3/26/2013.

SPARC T5-8 Realizes SAP SD Two-Tier Benchmark World Record for 8 Chip Systems

Oracle's SPARC T5-8 server produced a world record result for systems with 8 processors on the two-tier SAP Sales and Distribution (SD) Standard Application Benchmark.

  • The SPARC T5-8 server achieved 40,000 users with running the two-tier SAP Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement package 5 for SAP ERP 6.0.

  • The SPARC T5-8 server is 57% faster than the IBM Power 760 8-chip running SAP Enhancement Package 5 for SAP ERP 6.0.

  • The SPARC T5-8 server delivers 5% more SAP users per chip than the IBM Power 780 12-chip running SAP Enhancement Package 5 for SAP ERP 6.0.

  • The SPARC T5-8 server solution was run with Oracle Solaris 11 and used Oracle Database 11g.

Performance Landscape

SAP-SD 2-Tier Performance Table (in decreasing performance order). SAP ERP 6.0 Enhancement Pack 5 for SAP ERP 6.0 results (New version of the benchmark as of May 2012).

System OS
Database
Users SAPS SAP
ERP/ECC
Release
Date
SPARC T5-8 Server
8x SPARC T5 @3.6 GHz, 2 TB
Solaris 11
Oracle 11g
40,000 220,950 EHP5 for SAP
ERP 6.0
25-Mar-13
IBM Power 760
8xPOWER7+ @3.41 GHz, 1024 GB
AIX 7.1
DB2 10
25,488 139,220 EHP5 for SAP
ERP 6.0
5-Feb-13

SAP ERP 6.0 Enhancement Pack 4 for SAP ERP 6.0 Results
(Old version of the benchmark, obsolete at the end of April, 2012)

System OS
Database
Users SAPS SAP
ERP/ECC
Release
Date
IBM Power 795
32xPOWER7 @4 GHz, 4 TB
AIX 7.1
DB2 9.7
126,063 688,630 EHP4 for SAP
ERP 6.0
15-Nov-10
SPARC Enterprise Server M9000
64xSPARC64 VII @2.88 GHz, 1152 GB
Solaris 10
Oracle 10g
32,000 175,600 EHP4 for SAP
ERP 6.0
18-Nov-09

Complete benchmark results may be found at the SAP benchmark website http://www.sap.com/benchmark.

Configuration Summary and Results

Hardware Configuration:

1 x SPARC T5-8 server with
8 x 3.6 GHz SPARC T5 processors (total of 8 processors / 128 cores / 1024 threads)
2 TB memory
1 x Sun ZFS Storage 7420 appliance with
72 x 600 GB 15K RPM 3.5" SAS-2 disk
32 x 32 GB memory
1 x Sun Fire X4270 M2 server configured as a COMSTAR device with
10 x 2 TB 7.2K 3.5" SAS disk
18 x 8 GB memory

Software Configuration:

Oracle Solaris 11
SAP enhancement package 5 for SAP ERP 6.0
Oracle Database 11g Release 2

Certified Results (published by SAP)

Performance:
40,000 benchmark users
SAP Certification:
2013008

Benchmark Description

The SAP Standard Application SD (Sales and Distribution) Benchmark is a two-tier ERP business test that is indicative of full business workloads of complete order processing and invoice processing, and demonstrates the ability to run both the application and database software on a single system. The SAP Standard Application SD Benchmark represents the critical tasks performed in real-world ERP business environments.

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

Two-tier SAP Sales and Distribution (SD) Standard Application benchmarks SAP Enhancement package 5 for SAP ERP 6.0 as of 3/26/13:

SPARC T5-8 (8 processors, 128 cores, 1024 threads) 40,000 SAP SD users, 8 x 3.6 GHz SPARC T5, 2 TB memory, Oracle Database 11g, Oracle Solaris 11, Cert# 2013008. IBM Power 760 (8 processors, 48 cores, 192 threads) 25,488 SAP SD users, 8 x 3.41 GHz IBM POWER7+, 1024 GB memory, DB2 10, AIX 7.1, Cert#2013004.

Two-tier SAP Sales and Distribution (SD) Standard Application benchmarks SAP Enhancement package 4 for SAP ERP 6.0 as of 4/30/12:

IBM Power 795 (32 processors, 256 cores, 1024 threads) 126,063 SAP SD users, 32 x 4 GHz IBM POWER7, 4 TB memory, DB2 9.7, AIX7.1, Cert#2010046. SPARC Enterprise Server M9000 (64 processors, 256 cores, 512 threads) 32,000 SAP SD users, 64 x 2.88 GHz SPARC64 VII, 1152 GB memory, Oracle Database 10g, Oracle Solaris 10, Cert# 2009046.

SAP, R/3, reg TM of SAP AG in Germany and other countries. More info www.sap.com/benchmark

SPARC M5-32 Produces SAP SD Two-Tier Benchmark World Record for SAP Enhancement Package 5 for SAP ERP 6.0

Oracle's SPARC M5-32 server produced a world record result on the two-tier SAP Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement package 5 for SAP ERP 6.0.

  • The SPARC M5-32 server achieved 85,050 users running the two-tier SAP Sales and Distribution (SD) Standard Application Benchmark using SAP Enhancement package 5 for SAP ERP 6.0.

  • The SPARC M5-32 solution was run with Oracle Solaris 11 and used the Oracle Database 11g.

Performance Landscape

SAP-SD 2-Tier Performance Table (in decreasing performance order). SAP ERP 6.0 Enhancement Pack 5 for SAP ERP 6.0 results (new version of the benchmark as of May, 2012).

System OS
Database
Users SAPS SAP
ERP/ECC
Release
Date
SPARC M5-32 Server
32x SPARC M5 @3.6 GHz, 4 TB
Solaris 11
Oracle 11g
85,050 472,600 EHP5 for SAP
ERP 6.0
25-Mar-13
IBM Power 780
12xPOWER7+ @3.72 GHz, 1536 GB
AIX 7.1
DB2 10
57,024 311,720 EHP5 for SAP
ERP 6.0
3-Oct-12
IBM Power 760
8xPOWER7+ @3.41 GHz, 1024 GB
AIX 7.1
DB2 10
25,488 139,220 EHP5 for SAP
ERP 6.0
5-Feb-13

SAP ERP 6.0 Enhancement Pack 4 for SAP ERP 6.0 Results
(Old version of the benchmark, obsolete at the end of April, 2012)

System OS
Database
Users SAPS SAP
ERP/ECC
Release
Date
IBM Power 795
32xPOWER7 @4 GHz, 4 TB
AIX 7.1
DB2 9.7
126,063 688,630 EHP4 for SAP
ERP 6.0
15-Nov-10
SPARC Enterprise Server M9000
64xSPARC64 VII @2.88 GHz, 1152 GB
Solaris 10
Oracle 10g
32,000 175,600 EHP4 for SAP
ERP 6.0
18-Nov-09

Complete benchmark results may be found at the SAP benchmark website http://www.sap.com/benchmark.

Configuration Summary and Results

Hardware Configuration:

1 x SPARC M5-32 server with
32 x 3.6 GHz SPARC M5 processors (total of 32 processors / 192 cores / 1536 threads)
4 TB memory
1 x Sun Storage 2540-M2 (12 x 300 GB 5K RPM 3.5" SAS-2 disk & 2 GB cache)
Flash Storage

Software Configuration:

Oracle Solaris 11
SAP enhancement package 5 for SAP ERP 6.0
Oracle Database 11g Release 2

Certified Results (published by SAP)

Performance: 85,050 benchmark users
SAP Certification: 2013009

Benchmark Description

The SAP Standard Application SD (Sales and Distribution) Benchmark is a two-tier ERP business test that is indicative of full business workloads of complete order processing and invoice processing, and demonstrates the ability to run both the application and database software on a single system. The SAP Standard Application SD Benchmark represents the critical tasks performed in real-world ERP business environments.

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

Two-tier SAP Sales and Distribution (SD) standard application benchmarks, SAP Enhancement package 5 for SAP ERP 6.0 as of 3/26/13:

SPARC M5-32 (32 processors, 192 cores, 1536 threads) 85,050 SAP SD users, 32 x 3.6 GHz SPARC M5, 4 TB memory, Oracle Database 11g, Oracle Solaris 11, Cert# 2013009. IBM Power 780 (12 processors, 96 cores, 384 threads) 57,024 SAP SD users, 12 x 3.72 GHz IBM POWER7+, 1536 GB memory, DB210, AIX7.1, Cert#2012033. IBM Power 760 (8 processors, 48 cores, 192 threads) 25,488 SAP SD users, 8 x 3.41 GHz IBM POWER7+, 1024 GB memory, DB2 10, AIX 7.1, Cert#2013004.

Two-tier SAP Sales and Distribution (SD) standard application benchmarks, SAP Enhancement package 4 for SAP ERP 6.0 as of 3/26/13:

IBM Power 795 (32 processors, 256 cores, 1024 threads) 126,063 SAP SD users, 32 x 4 GHz IBM POWER7, 4 TB memory, DB2 9.7, AIX7.1, Cert#2010046. SPARC Enterprise Server M9000 (64 processors, 256 cores, 512 threads) 32,000 SAP SD users, 64 x 2.88 GHz SPARC64 VII, 1152 GB memory, Oracle Database 10g, Oracle Solaris 10, Cert# 2009046.

SAP, R/3, reg TM of SAP AG in Germany and other countries. More info www.sap.com/benchmark

SPARC T5 Systems Deliver SPEC CPU2006 Rate Benchmark Multiple World Records

Oracle's SPARC T5 processor based systems delivered world record performance on the SPEC CPU2006 rate benchmarks. This was accomplished with Oracle Solaris 11.1 and Oracle Solaris Studio 12.3 software.

SPARC T5-8

  • The SPARC T5-8 server delivered world record SPEC CPU2006 rate benchmark results for systems with eight processors.

  • The SPARC T5-8 server achieved scores of 3750 SPECint_rate2006, 3490 SPECint_rate_base2006, 3020 SPECfp_rate2006, and 2770 SPECfp_rate_base2006.

  • The SPARC T5-8 server beat the 8 processor IBM Power 760 with POWER7+ processors by 1.7x on the SPECint_rate2006 benchmark and 2.2x on the SPECfp_rate2006 benchmark.

  • The SPARC T5-8 server beat the 8 processor IBM Power 780 with POWER7 processors by 35% on the SPECint_rate2006 benchmark and 14% on the SPECfp_rate2006 benchmark.

  • The SPARC T5-8 server beat the 8 processor HP DL980 G7 with Intel Xeon E7-4870 processors by 1.7x on the SPECint_rate2006 benchmark and 2.1x on the SPECfp_rate2006 benchmark.

SPARC T5-1B

  • The SPARC T5-1B server module delivered world record SPEC CPU2006 rate benchmark results for systems with one processor.

  • The SPARC T5-1B server module achieved scores of 467 SPECint_rate2006, 436 SPECint_rate_base2006, 369 SPECfp_rate2006, and 350 SPECfp_rate_base2006.

  • The SPARC T5-1B server module beat the 1 processor IBM Power 710 Express with a POWER7 processor by 62% on the SPECint_rate2006 benchmark and 49% on the SPECfp_rate2006 benchmark.

  • The SPARC T5-1B server module beat the 1 processor NEC Express5800/R120d-1M with an Intel Xeon E5-2690 processor by 31% on the SPECint_rate2006 benchmark. The SPARC T5-1B server module beat the 1 processor Huawei RH2288 V2 with an Intel Xeon E5-2690 processor by 44% on the SPECfp_rate2006 benchmark.

  • The SPARC T5-1B server module beat the 1 processor Supermicro A+ 1012G-MTF with an AMD Operton 6386 SE processor by 51% on the SPECint_rate2006 benchmark and 65% on the SPECfp_rate2006 benchmark.

Performance Landscape

Complete benchmark results are at the SPEC website, SPEC CPU2006 Results. The tables below provide the new Oracle results, as well as, select results from other vendors.

SPEC CPU2006 Rate Results – Eight Processors
System Processor ch/co/th * Peak Base
SPECint_rate2006
SPARC T5-8 SPARC T5, 3.6 GHz 8/128/1024 3750 3490
IBM Power 780 POWER7, 3.92 GHz 8/64/256 2770 2420
HP DL980 G7 Xeon E7-4870, 2.4 GHz 8/80/160 2180 2070
IBM Power 760 POWER7+, 3.42 GHz 8/48/192 2170 1480
Dell PowerEdge C6145 Opteron 6180 SE, 2.5 GHz 8/96/96 1670 1440
SPECfp_rate2006
SPARC T5-8 SPARC T5, 3.6 GHz 8/128/1024 3020 2770
IBM Power 780 POWER7, 3.92 GHz 8/64/256 2640 2410
HP DL980 G7 Xeon E7-4870, 2.4 GHz 8/80/160 1430 1380
IBM Power 760 POWER7+, 3.42 GHz 8/48/192 1400 1130
Dell PowerEdge C6145 Opteron 6180 SE, 2.5 GHz 8/96/96 1310 1200

* ch/co/th — chips / cores / threads enabled

SPEC CPU2006 Rate Results – One Processor
System Processor ch/co/th * Peak Base
SPECint_rate2006
SPARC T5-1B SPARC T5, 3.6 GHz 1/16/128 467 436
NEC Express5800/R120d-1M Xeon E5-2690, 2.9 GHz 1/8/16 357 343
Supermicro A+ 1012G-MTF Opteron 6386 SE, 2.8 GHz 1/16/16 309 269
IBM Power 710 Express POWER7, 3.556 GHz 1/8/32 289 255
SPECfp_rate2006
SPARC T5-1B SPARC T5, 3.6 GHz 1/16/128 369 350
Huawei RH2288 V2 Xeon E5-2690, 2.9 GHz 1/8/16 257 250
IBM Power 710 Express POWER7, 3.556 GHz 1/8/32 248 229
Supermicro A+ 1012G-MTF Opteron 6386 SE, 2.8 GHz 1/16/16 223 199

* ch/co/th — chips / cores / threads enabled

Configuration Summary

Systems Under Test:

SPARC T5-8
8 x 3.6 GHz SPARC T5 processors
4 TB memory (128 x 32 GB dimms)
2 TB on 8 x 600 GB 10K RPM SAS disks, arranged as 4 x 2-way mirrors
Oracle Solaris 11.1 (SRU 4.6)
Oracle Solaris Studio 12.3 1/13 PSE

SPARC T5-1B
1 x 3.6 GHz SPARC T5 processor
256 GB memory (16 x 16 GB dimms)
157 GB on 2 x 300 GB 10K RPM SAS disks (mirrored)
Oracle Solaris 11.1 (SRU 3.4)
Oracle Solaris Studio 12.3 1/13 PSE

Benchmark Description

SPEC CPU2006 is SPEC's most popular benchmark. It measures:

  • Speed — single copy performance of chip, memory, compiler
  • Rate — multiple copy (throughput)

The benchmark is also divided into integer intensive applications and floating point intensive applications:

  • integer: 12 benchmarks derived from real applications such as perl, gcc, XML processing, and pathfinding
  • floating point: 17 benchmarks derived from real applications, including chemistry, physics, genetics, and weather.

It is also divided depending upon the amount of optimization allowed:

  • base: optimization is consistent per compiled language, all benchmarks must be compiled with the same flags per language.
  • peak: specific compiler optimization is allowed per application.

The overall metrics for the benchmark which are commonly used are:

  • SPECint_rate2006, SPECint_rate_base2006: integer, rate
  • SPECfp_rate2006, SPECfp_rate_base2006: floating point, rate
  • SPECint2006, SPECint_base2006: integer, speed
  • SPECfp2006, SPECfp_base2006: floating point, speed

See Also

Disclosure Statement

SPEC and the benchmark names SPECfp and SPECint are registered trademarks of the Standard Performance Evaluation Corporation. Results as of March 26, 2013 from www.spec.org and this report. SPARC T5-8: 3750 SPECint_rate2006, 3490 SPECint_rate_base2006, 3020 SPECfp_rate2006, 2770 SPECfp_rate_base2006; SPARC T5-1B: 467 SPECint_rate2006, 436 SPECint_rate_base2006, 369 SPECfp_rate2006, 350 SPECfp_rate_base2006.

SPARC T5-2 Achieves JD Edwards EnterpriseOne Benchmark World Records

Oracle produced World Record batch throughput for single system results on Oracle's JD Edwards EnterpriseOne Day-in-the-Life benchmark using Oracle's SPARC T5-2 server running Oracle Solaris Containers and consolidating JD Edwards EnterpriseOne, Oracle WebLogic servers and the Oracle Database 11g Release 2. There are two workloads tested: online plus batch workload and batch-only workload.

Online plus batch workload:

  • The SPARC T5-2 server delivered a result of 12,000 online users at 180 msec average response time while concurrently executing a mix of JD Edwards EnterpriseOne long and short batch processes at 198.5 UBEs/min (Universal Batch Engines per minute).

  • The SPARC T5-2 server online plus batch throughput is 2.7x higher than the IBM Power 770 server, both running 12,000 online users.

  • The SPARC T5-2 server online plus batch throughput is 6x higher per chip than the IBM Power 770 server. The SPARC T5-2 server has 2 chips and the IBM Power 770 has 4 chips, both ran 12,000 online users.

  • The SPARC T5-2 server online plus batch throughput is 3x higher per core than the IBM Power 770 server. Both servers have 32 cores and ran 12,000 online users.

Batch-only workload:

  • The SPARC T5-2 server delivered throughput of 880 UBEs/min while executing the batch-only workload (Long and Short batch processes).

  • The SPARC T5-2 server batch-only throughput is 2.7x faster per chip than the IBM Power 770 server. The SPARC T5-2 server has 2 chips and the IBM Power 770 has 4 chips.

  • The SPARC T5-2 server batch-only throughput is 1.4x higher per core than the IBM Power 770 server. Both servers have 32 cores.

  • The SPARC T5-2 server batch-only throughput is 61% faster than the Cisco multiple system solution.

  • The SPARC T5-2 server batch-only throughput is 5x faster per chip than the Cisco UCS B200/B250 M2 servers. The SPARC T5-2 server has 2 chips and the Cisco 3 server solution has 6 chips.

  • The SPARC T5-2 server batch-only throughput is 18x higher per core than the Cisco UCS B200/B250 M2 servers. The SPARC T5-2 server has 32 cores while the Cisco solution utilized 36 cores.

Both workloads:

  • The SPARC T5-2 server offers a 5.4x cost savings for the application server when compared to the IBM Power 770 application server.

  • The SPARC T5-2 server running Oracle Solaris Containers and consolidating JD Edwards EnterpriseOne, Oracle WebLogic servers and the Oracle Database 11g Release 2 utilized a maximum 65% of the available CPU power, leaving headroom for additional processing.

  • The database server in a shared-server configuration allows for optimized CPU resource utilization and significant memory savings on the SPARC T5-2 server without sacrificing performance.

Performance Landscape

JD Edwards EnterpriseOne Day in the Life (DIL) Benchmark
Consolidated Online with Batch Workload
System Rack
Units (U)
Batch
Rate
(UBEs/min)
Online
Users
Users/
U
UBEs/
Core
UBEs/
Chip
Version
SPARC T5-2 (2 x SPARC T5, 3.6 GHz) 3 198.5 12000 4000 6.2 99 9.0.2
IBM Power 770 (4 x POWER7, 3.3 GHz) 8 65 12000 1500 2.0 16 9.0.2

Batch Rate (UBEs/min) — Batch transaction rate in UBEs per minute.

JD Edwards EnterpriseOne Batch Only Benchmark
System Rack
Units (U)
Batch
Rate
(UBEs/min)
UBEs/
U
UBEs/
Core
UBEs/
Chip
Version
SPARC T5-2 (2 x SPARC T5, 3.6 GHz) 3 880 267 25 440 9.0.2
IBM Power 770 (4 x POWER7, 3.3 GHz) 8 643 81 20 161 9.0.2
2 x Cisco B200 M2 (2 x X5690, 3.46 GHz)
1 x Cisco B250 M2 (2 x X5680, 3.33 GHz)
3 546 182 15 91 9.0.2

Configuration Summary

Hardware Configuration:

1 x SPARC T5-2 server with
2 x SPARC T5 processors, 3.6 GHz
512 GB memory
4 x 300 GB 10K RPM SAS internal disk
2 x 300 GB internal SSD
4 x Sun Flash Accelerator F40 PCIe Card (4 x 93 GB)

Software Configuration:

Oracle Solaris 10 1/13
Oracle Solaris Containers
JD Edwards EnterpriseOne 9.0.2
JD Edwards EnterpriseOne Tools (8.98.4.2)
Oracle WebLogic Server 11g (10.3.4)
Oracle HTTP Server 11g
Oracle Database 11g Release 2 (11.2.0.3)

Benchmark Description

JD Edwards EnterpriseOne is an integrated applications suite of Enterprise Resource Planning (ERP) software. Oracle offers 70 JD Edwards EnterpriseOne application modules to support a diverse set of business operations.

Oracle's Day in the Life (DIL) kit is a suite of scripts that exercises most common transactions of JD Edwards EnterpriseOne applications, including business processes such as payroll, sales order, purchase order, work order, and manufacturing processes, such as ship confirmation. These are labeled by industry acronyms such as SCM, CRM, HCM, SRM and FMS. The kit's scripts execute transactions typical of a mid-sized manufacturing company.

  • The workload consists of online transactions and the UBE – Universal Business Engine workload of 61 short and 4 long UBEs.

  • LoadRunner runs the DIL workload, collects the user’s transactions response times and reports the key metric of Combined Weighted Average Transaction Response time.

  • The UBE processes workload runs from the JD Enterprise Application server.

    • Oracle's UBE processes come as three flavors:
      • Short UBEs < 1 minute engage in Business Report and Summary Analysis,
      • Mid UBEs > 1 minute create a large report of Account, Balance, and Full Address,
      • Long UBEs > 2 minutes simulate Payroll, Sales Order, night only jobs.
    • The UBE workload generates large numbers of PDF files reports and log files.
    • The UBE Queues are categorized as the QBATCHD, a single threaded queue for large and medium UBEs, and the QPROCESS queue for short UBEs run concurrently.

Oracle's UBE process performance metric is Number of Maximum Concurrent UBE processes at transaction rate, UBEs/minute.

Key Points and Best Practices

Four Oracle Solaris processors sets were used with Oracle Solaris Containers assigned to the processor sets as follows:

  • one JD Edwards EnterpriseOne Application server, two Oracle WebLogic Servers 11g Release 1 each coupled with an Oracle Web Tier HTTP server instances (online workload), each in an Oracle Solaris Container (three total),

  • one JD Edwards EnterpriseOne Application server (for batch only workload) in an Oracle Solaris Container,

  • Oracle Database 11g Release 2.0.3 database in an Oracle Solaris Container,

  • the Oracle database log writer.

Other items of note:

  • Each Oracle WebLogic vertical cluster, with twelve managed instances, was configured in a dedicated webserver container in order to load balance users' requests and to provide the infrastructure to support high number of users with ease of deployment and high availability.

  • The database redo logs were configured on the raw disk partitions.

  • The mixed batch workload of 44 short UBEs and 8 long UBEs was executed concurrently with the 12,000 online application users, producing a sustained rate of 198.5 UBE/min.

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 03/26/2013

SPARC T5-2 (SPARC T5-2 Server base package, 2xSPARC T5 16-core processors, 32x16GB-1066 DIMMS, 4x600GB 10K RPM 2.5. SAS-2 HDD,2x300GB SSDs, 4x Sun Flash Accelerator F40 PCIe Cards, 2x Power Cables) List Price $98,190. IBM Power 770 (IBM Power 770:9917 Model MMC, 2x3.3GHz 16-core, 32x one processor activation, 2xCEC Enclosure with IBM Bezel, I/O Backplane and System Midplane,2x Service Processor, 16x 0/64GB DDR3 Memory (4x16GB) DIMMS-1066MHz Power7 CoD Memory, 24x Activation of 1 GB DDR3 Power7 Memory, 10x Activation of 100GB DDR3 Power7 Memory, 2x Disk/Media Backplane. 2x 300GB SAS 15K RPM 2.5. HDD (AIX/Linux only), 1x SATA slimline DVD-RAM drive, 4x AC Power Supply 1925W) List Price $532,143. Source: ibm.com, collected 03/18/2013.

SPARC T5-2 Scores Siebel CRM Benchmark World Record

Oracle set a new world record for the Siebel Platform Sizing and Performance Program (PSPP) benchmark using Oracle's SPARC T5-2 servers for the application server with Oracle's Siebel CRM 8.1.1.4 Industry Applications and Oracle Database 11g Release 2 running on Oracle Solaris.

  • The SPARC T5-2 servers running the application tier achieved 40,000 users with sub-second response time and with throughput of 333,339 business transactions per hour on the Siebel PSPP benchmark.

  • The SPARC T5-2 servers in the application tier delivered 2 times better performance on a per chip basis compared to earlier published SPARC T4 numbers.

  • The Siebel 8.1.1.4 PSPP workload includes Siebel Call Center and Order Management System.

  • The SPARC T5-2 server used Oracle Solaris Zones which provide flexible, scalable and manageable virtualization to scale the application within and across multiple nodes.

Performance Landscape

Application Server Transactions/
hour
Users Users/
Core
Call
Center
Order
Mgmt
Response Times (sec)
2 x SPARC T5-2 (2 x SPARC T5 3.6 GHz) 333,339 40,000 625 0.110 0.608
3 x SPARC T4-2 (2 x SPARC T4 2.85 GHz) 239,748 29,000 604 0.165 0.925
2 x IBM Power 750 (POWER7 3.55 GHz, 16 active cores) 176,185 21,000 656 0.052 0.250

Oracle:
Call Center + Order Management
Transactions: 273,786 + 59,553
Users: 28,000 + 12,000

IBM:
Call Center + Order Management
Transactions: 144,457 + 31,728
Users: 14,700 + 6,300

Configuration Summary

Application Server Configuration:

2 x SPARC T5-2 servers, each with
2 x SPARC T5 processors, 3.6 GHz
512 GB memory
6 x 300 GB SAS internal disks
Oracle Solaris 10 8/11
Siebel CRM 8.1.1.4 SIA

Web Server Configuration:

1 x SPARC T4-1 server
1 x SPARC T4 processor, 2.85 GHz
128 GB memory
Oracle Solaris 10 8/11
iPlanet Web Server 7

Database Server Configuration:

1 x SPARC T4-2 server
2 x SPARC T4 processors, 2.85 GHz
256 GB memory
Flash Storage
Oracle Solaris 10 8/11
Oracle Database 11g Release 2 (11.2.0.2)

Benchmark Description

Siebel PSPP benchmark includes Call Center and Order Management:

  • Siebel Financial Services Call Center – Provides the most complete solution for sales and service, allowing customer service and telesales representatives to provide superior customer support, improve customer loyalty, and increase revenues through cross-selling and up-selling.

    High-level description of the use cases tested: Incoming Call Creates Opportunity, Quote and Order and Incoming Call Creates Service Request. Three complex business transactions are executed simultaneously for specific number of concurrent users. The ratios of these 3 scenarios were 30%, 40%, 30% respectively, which together were totaling 70% of all transactions simulated in this benchmark. Between each user operation and the next one, the think time averaged approximately 10, 13, and 35 seconds respectively.

  • Siebel Order Management – Oracle's Siebel Order Management allows employees such as salespeople and call center agents to create and manage quotes and orders through their entire life cycle. Siebel Order Management can be tightly integrated with back-office applications allowing users to perform tasks such as checking credit, confirming availability, and monitoring the fulfillment process.

    High-level description of the use cases tested: Order & Order Items Creation and Order Updates. Two complex Order Management transactions were executed simultaneously for specific number of concurrent users concurrently with aforementioned three Call Center scenarios above. The ratio of these 2 scenarios was 50% each, which together were totaling 30% of all transactions simulated in this benchmark. Between each user operation and the next one, the think time averaged approximately 20 and 67 seconds respectively.

Key Points and Best Practices

  • No processor cores or cache were activated or deactivated on the SPARC T-Series systems to achieve special benchmark effects.

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 26 March 2013.

SPARC T5 Systems Produce Oracle TimesTen Benchmark World Record

The Oracle TimesTen In-Memory Database is optimized to run on Oracle's SPARC T5 processor platforms running Oracle Solaris 11. In this series of tests, systems with the new SPARC T5 processor were significantly faster than systems based on other processors. Two tests were run to explore TimesTen performance: a Mobile Call Processing test (based on customer workload) and Oracle's TimesTen Performance Throughput Benchmark (TPTBM). TimesTen version 11.2.2.4 was used for all tests.

  • On the TimesTen Performance Throughput Benchmark (TPTBM), SPARC T5-8 server produced a world record 59.9 million read transactions per second.

  • On the Mobile Call Processing test, the SPARC T5 processor achieves 2.4 times more throughput than the Intel Xeon E7-4870 processor. The two-chip SPARC T5-2 server is 22% faster than an x86 server with four Intel E7-4870 2.4 GHz processors.

  • On the TimesTen Performance Throughput Benchmark (TPTBM) read-only workload, the SPARC T5 processor achieves 2.2 times higher throughput than the Intel Xeon E7-4870 processor. On the same workload, the two-chip SPARC T5-2 server produces 10% more throughput than an x86 server with four Intel E7-4870 processors and has almost twice the performance of a 2-chip Intel E5-2680 system.

  • With the TPTBM read-only workload, the SPARC T5-8 server delivers 3.8x more throughput than a SPARC T5-2 Server, showing excellent scalability.

  • The SPARC T5 processor delivers over twice the performace of the previous generation SPARC T4 processor and over 4x the performace of the SPARC T3 processor, all in the same amount of space.

  • The SPARC T5-2 server delivers 2.4x the performace of the SPARC T4-2 server in the same 3U space. This is better performance than that of the SPARC T4-4 server which occupies 5U.

Performance Landscape

Mobile Call Processing Test Performance

Processor Tps
SPARC T5, 3.6 GHz 367,600
Intel Xeon E7-4870, 2.4 GHz 302,000
SPARC T4, 2.85 GHz 230,500

All systems measured using Oracle Solaris 11 and Oracle TimesTen In-Memory Database 11.2.2.4.1

TimesTen Performance Throughput Benchmark (TPTBM) Read-Only

System Processor Chips Tps Tps/
Chip
SPARC T5-8 SPARC T5, 3.6 GHz 8 59.9M 7.5M
SPARC T5-2 SPARC T5, 3.6 GHz 2 15.9M 7.9M
x86 Intel Xeon E7-4870, 2.4 GHz 4 14.5M 3.6M
SPARC T4-4 SPARC T4, 3.0 GHz 4 14.2M 3.6M
x86* Intel Xeon E5-2680, 2.7 GHz 2 8.5M 4.3
SPARC T4-2 SPARC T4, 2.85 GHz 2 6.5M 3.3M
SPARC T3-4 SPARC T3, 1.65 GHz 4 7.9M 1.9M
T5440 SPARC T2+, 1.4 GHz 4 3.1M 0.8M

All systems measured using Oracle Solaris 11 and Oracle TimesTen In-Memory Database 11.2.2.4.1

*Intel E5-2680 using Oracle Linux and Oracle TimesTen In-Memory Database 11.2.2.4.1

TimesTen Performance Throughput Benchmark (TPTBM) Update-Only

Processor Tps
SPARC T5, 3.6 GHz 1,031.7K
Intel Xeon E7-4870, 2.4 GHz 988.1K
Intel Xeon E5-2680, 2.7 GHz * 944.3K
SPARC T4, 3.0 GHz 678.0K

All systems measured using Oracle Solaris 11 and Oracle TimesTen In-Memory Database 11.2.2.4.1

*Intel E5-2680 using Oracle Linux and Oracle TimesTen In-Memory Database 11.2.2.4.1

Configuration Summary

Hardware Configurations:

SPARC T5-8 server
8 x SPARC T5 processors, 3.6 GHz
2 TB memory
1 x 8 Gbs FC Qlogic HBA
1 x 6 Gbs SAS HBA
2 x 300 GB internal disks
Oracle Solaris 11
TimesTen 11.2.2.4.1
1 x Sun Fire X4275 server configured as COMSTAR redo head (log)

SPARC T5-2 server
2 x SPARC T5 processors, 3.6 GHz
512 GB memory
1 x 8 Gbs FC Qlogic HBA
1 x 6 Gbs SAS HBA
2 x 300 GB internal disks
Oracle Solaris 11
TimesTen 11.2.2.4.1
1 x Sun Fire X4275 server configured as COMSTAR redo head (log)

SPARC T4-4 server
4 x SPARC T4 processors, 3.0 GHz
1 TB memory
1 x 8 Gbs FC Qlogic HBA
1 x 6 Gbs SAS HBA
6 x 300 GB internal disks
Oracle Solaris 11
TimesTen 11.2.2.4.1
Sun Storage F5100 Flash Array (80 x 24 GB flash modules)
1 x Sun Fire X4275 server configured as COMSTAR redo head (log)

SPARC T4-2 server
2 x SPARC T4 processors, 2.85 GHz
256 GB memory
1 x 8 Gbs FC Qlogic HBA
1 x 6 Gbs SAS HBA
4 x 300 GB internal disks
Oracle Solaris 11
TimesTen 11.2.2.4.1
Sun Storage F5100 Flash Array (40 x 24 GB flash modules)
1 x Sun Fire X4275 server configured as COMSTAR head

SPARC T3-4 server
4 x SPARC T3 processors, 1.6 GHz
512 GB memory
1 x 8 Gbs FC Qlogic HBA
8 x 146 GB internal disks
Oracle Solaris 11
TimesTen 11.2.2.4.1
1 x Sun Fire X4275 server configured as COMSTAR head

Intel Server x86_64
2 x Intel Xeon E5-2680 processors, 2.7 GHz
256 GB memory
4 x SSD SAS disks (log)
1 x 600 GB internal disks
Oracle Linux
TimesTen 11.2.2.4.1

Sun Server X2-4
4 x Intel Xeon E7-4870 processors, 2.4 GHz
512 GB memory
1 x 8 Gbs FC Qlogic HBA
6 x 146 GB internal disks
Oracle Solaris 11
TimesTen 11.2.2.4.1
1 x Sun Fire X4275 server configured as COMSTAR redo head (log)

Benchmark Descriptions

TimesTen Performance Throughput BenchMark (TPTBM) is shipped with TimesTen and measures the total throughput of the system. The benchmark workloads can be reads, inserts, updates, and delete operations, or a mix of them as required.

Mobile Call Processing is a customer-based workload for processing calls made by mobile phone subscribers. The workload has a mixture of read-only, update, and insert-only transactions. The peak throughput performance is measured from multiple concurrent processes executing the transactions until a peak performance is reached via saturation of the available resources.

Key Points and Best Practices

The Mobile Call Processing test utilized Oracle Solaris processor sets in all environments for optimum performance. This features isolates running processes from other processes in the system. Combined with parameters to limit memory pages to the lgroup within the processor set and isolating the processor set to a single processor within the system.

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 26 March 2013.

SPARC T5-8 Delivers Oracle OLAP World Record Performance

Oracle's SPARC T5-8 server delivered world record query performance with near real-time analytic capability using the Oracle OLAP Perf Version 3 workload running Oracle Database 11g Release 2 on Oracle Solaris 11.

  • The maximum query throughput on the SPARC T5-8 server is 1.6x higher than that of the 8-chip Intel Xeon E7-8870 server. Both systems had sub-second response time.

  • The SPARC T5-8 server with the Oracle Database demonstrated the ability to support at least 600 concurrent users querying OLAP cubes (with no think time), processing 2.93 million analytic queries per hour with an average response time of 0.66 seconds per query. This performance was enabled by keeping the entire cube in-memory utilizing the 4 TB of memory on the SPARC T5-8 server.

  • Assuming a 60 second think time between query requests, the SPARC T5-8 server can support approximately 49,450 concurrent users with the same 0.66 sec response time.

  • The SPARC T5-8 server delivered 4.3x times the maximum query throughput of a SPARC T4-4 server.

  • The workload uses a set of realistic BI queries that run against an OLAP cube based on a 4 billion row fact table of sales data. The 4 billion rows are partitioned by month spanning 10 years.

  • The combination of the Oracle Database with the Oracle OLAP option running on a SPARC T5-8 server supports live data updates occurring concurrently with minimally impacted user query executions.

Performance Landscape

Oracle OLAP Perf Version 3 Benchmark
Oracle cube base on 4 billion fact table rows
10 years of data partitioned by month
System Queries/
hour
Users* Average Response
Time (sec)
0 sec think time 60 sec think time
SPARC T5-8 2,934,000 600 49,450 0.66
8-chip Intel Xeon E7-8870 1,823,000 120 30,500 0.19
SPARC T4-4 686,500 150 11,580 0.71

Configuration Summary and Results

SPARC T5-8 Hardware Configuration:

1 x SPARC T5-8 server with
8 x SPARC T5 processors, 3.6 GHz
4 TB memory
Data Storage and Redo Storage
1 x Sun Storage F5100 Flash Array (with 80 FMODs)
Oracle Solaris 11.1
Oracle Database 11g Release 2 (11.2.0.3) with Oracle OLAP option

Sun Server X2-8 Hardware Configuration:

1 x Sun Server X2-8 with
8 x Intel Xeon E7-8870 processors, 2.4 GHz
512 GB memory
Data Storage and Redo Storage
3 x StorageTek 2540/2501 array pairs
Oracle Solaris 10 10/12
Oracle Database 11g Release 2 (11.2.0.2) with Oracle OLAP option

SPARC T4-4 Hardware Configuration:

1 x SPARC T4-4 server with
4 x SPARC T4 processors, 3.0 GHz
1 TB memory
Data Storage
1 x Sun Fire X4275 (using COMSTAR)
2 x Sun Storage F5100 Flash Array (each with 80 FMODs)
Redo Storage
1 x Sun Fire X4275 (using COMSTAR with 8 HDD)
Oracle Solaris 11 11/11
Oracle Database 11g Release 2 (11.2.0.3) with Oracle OLAP option

Benchmark Description

The Oracle OLAP Perf Version 3 benchmark is a workload designed to demonstrate and stress the ability of the OLAP Option to deliver fast query, near real-time updates and rich calculations using a multi-dimensional model in the context of the Oracle data warehousing.

The bulk of the benchmark entails running a number of concurrent users, each issuing typical multidimensional queries against an Oracle cube. The cube has four dimensions: time, product, customer, and channel. Each query user issues approximately 150 different queries. One query chain may ask for total sales in a particular region (e.g South America) for a particular time period (e.g. Q4 of 2010) followed by additional queries which drill down into sales for individual countries (e.g. Chile, Peru, etc.) with further queries drilling down into individual stores, etc. Another query chain may ask for yearly comparisons of total sales for some product category (e.g. major household appliances) and then issue further queries drilling down into particular products (e.g. refrigerators, stoves. etc.), particular regions, particular customers, etc.

While the core of every OLAP Perf benchmark is real world query performance, the benchmark itself offers numerous execution options such as varying data set sizes, number of users, numbers of queries for any given user and cube update frequency. Version 3 of the benchmark is executed with a much larger number of query streams than previous versions and used a cube designed for near real-time updates. The results produced by version 3 of the benchmark are not directly comparable to results produced by previous versions of the benchmark.

The near real-time update capability is implemented along the following lines. A large Oracle cube, H, is built from a 4 billion row star schema, containing data up until the end of last business day. A second small cube, D, is then created which will contain all of today's new data coming in from outside the world. It will be updated every L minutes with the data coming in within the last L minutes. A third cube, R, joins cubes H and D for reporting purposes much like a view might join data from two tables. Calculations are installed into cube R. The use of a reporting cube which draws data from different storage cubes is a common practice.

Query users are never locked out of query operations while new data is added to the update cube. The point of the demonstration is to show that an Oracle OLAP system can be designed which results in data being no more than L minutes out of date, where L may be as low as just a few minutes. This is what is meant by near real-time analytics.

Key Points and Best Practices

  • Update performance of the D cube was optimized by running update processes in the FX class with a priority greater than 0. The maximum lag time between updates to the source fact table and data availability to query users (what was referred to as L in the benchmark description) was less than 3 minutes for the benchmark environment on the SPARC T5-8 server.

  • Building and querying cubes with the Oracle OLAP option requires a large temporary tablespace. Normally temporary tablespaces would reside on disk storage. However, because the SPARC T5-8 server used in this benchmark had 4 TB of main memory, it was possible to use main memory for the OLAP temporary tablespace. This was done by using files in /tmp for the temporary tablespace datafiles.

  • Since typical BI users are often likely to issue similar queries, either with the same, or different, constants in the where clauses, setting the init.ora parameter "cursor_sharing" to "force" provides for additional query throughput and a larger number of potential users.

  • Assuming the normal Oracle initialization parameters (e.g. SGA, PGA, processes etc.) are appropriately set, out of the box performance for the OLAP Perf workload should be close to what is reported here. Additional performance resulted from (a)using memory for the OLAP temporary tablespace (b)setting "cursor_sharing" to force.

  • For a given number of query users with zero think time, the main measured metrics are the average query response time and the query throughput. A derived metric is the maximum number of users the system can support, with the same response time, assuming some non-zero think time. The calculation of this maximum is from the well-known "response-time law"

      N = (rt + tt) * tp

    where rt is the average response time, tt is the think time and tp is the measured throughput.

    Setting tt to 60 seconds, rt to 0.66 seconds and tp to 815 queries/sec (2,934,000 queries/hour), the above formula shows that the SPARC T5-8 server will support 49,450 concurrent users with a think time of 60 seconds and an average response time of 0.66 seconds.

    For more information about the "response-time law" see chapter 3 from the book "Quantitative System Performance" cited below.

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 03/26/2013.

SPARC T5-2 Achieves ZFS File System Encryption Benchmark World Record

Oracle continues to lead in enterprise security. Oracle's SPARC T5 processors combined with the Oracle Solaris ZFS file system demonstrate faster file system encryption than equivalent x86 systems using the Intel Xeon Processor E5-2600 Sequence chips which have AES-NI security instructions.

Encryption is the process where data is encoded for privacy and a key is needed by the data owner to access the encoded data.

  • The SPARC T5-2 server is 3.4x faster than a 2 processor Intel Xeon E5-2690 server running Oracle Solaris 11.1 that uses the AES-NI GCM security instructions for creating encrypted files.

  • The SPARC T5-2 server is 2.2x faster than a 2 processor Intel Xeon E5-2690 server running Oracle Solaris 11.1 that uses the AES-NI CCM security instructions for creating encrypted files.

  • The SPARC T5-2 server consumes a significantly less percentage of system resources as compared to a 2 processor Intel Xeon E5-2690 server.

Performance Landscape

Below are results running two different ciphers for ZFS encryption. Results are presented for runs without any cipher, labeled clear, and a variety of different key lengths. The results represent the maximum delivered values measured for 3 concurrent sequential write operations using 1M blocks. Performance is measured in MB/sec (bigger is better). System utilization is reported as %CPU as measured by iostat (smaller is better).

The results for the x86 server were obtained using Oracle Solaris 11.1 with performance bug fixes.

Encryption Using AES-GCM Ciphers

System GCM Encryption: 3 Concurrent Sequential Writes
Clear AES-256-GCM AES-192-GCM AES-128-GCM
MB/sec %CPU MB/sec %CPU MB/sec %CPU MB/sec %CPU
SPARC T5-2 server 3,918 7 3,653 14 3,676 15 3,628 14
SPARC T4-2 server 2,912 11 2,662 31 2,663 30 2,779 31
2-Socket Intel Xeon E5-2690 3,969 42 1,062 58 1,067 58 1,076 57
SPARC T5-2 vs x86 server 1.0x 3.4x 3.4x 3.4x

Encryption Using AES-CCM Ciphers

System CCM Encryption: 3 Concurrent Sequential Writes
Clear AES-256-CCM AES-192-CCM AES-128-CCM
MB/sec %CPU MB/sec %CPU MB/sec %CPU MB/sec %CPU
SPARC T5-2 server 3,862 7 3,665 15 3,622 14 3,707 12
SPARC T4-2 server 2,945 11 2,471 26 2,801 26 2,442 25
2-Socket Intel Xeon E5-2690 3,868 42 1,566 64 1,632 63 1,689 66
SPARC T5-2 vs x86 server 1.0x 2.3x 2.2x 2.2x

Configuration Summary

Storage Configuration:

Sun Storage 6780 array
4 CSM2 trays, each with 16 83GB 15K RPM drives
8x 8 GB/sec Fiber Channel ports per host
R0 Write cache enabled, controller mirroring off for peak write bandwidth
8 Drive R0 512K stripe pools mirrored via ZFS to storage

Sun Storage 6580 array
9 CSM2 trays, each with 16 136GB 15K RPM drives
8x 4 GB/sec Fiber Channel ports per host
R0 Write cache enabled, controller mirroring off for peak write bandwidth
4 Drive R0 512K stripe pools mirrored via ZFS to storage

Server Configuration:

SPARC T5-2 server
2 x SPARC T5 3.6 GHz processors
512 GB memory
Oracle Solaris 11.1

SPARC T4-2 server
2 x SPARC T4 2.85 GHz processors
256 GB memory
Oracle Solaris 11.1

Sun Server X3-2L server
2 x Intel Xeon E5-2690, 2.90 GHz processors
128 GB memory
Oracle Solaris 11.1

Switch Configuration:

Brocade 5300 FC switch

Benchmark Description

This benchmark evaluates secure file system performance by measuring the rate at which encrypted data can be written. The Vdbench tool was used to generate the IO load. The test performed 3 concurrent sequential write operations using 1M blocks to 3 separate files.

Key Points and Best Practices

  • ZFS encryption is integrated with the ZFS command set. Like other ZFS operations, encryption operations such as key changes and re-key are performed online.

  • Data is encrypted using AES (Advanced Encryption Standard) with key lengths of 256, 192, and 128 in the CCM and GCM operation modes.

  • The flexibility of encrypting specific file systems is a key feature.

  • ZFS encryption is inheritable to descendent file systems. Key management can be delegated through ZFS delegated administration.

  • ZFS encryption uses the Oracle Solaris Cryptographic Framework which gives it access to SPARC T5 and Intel Xeon E5-2690 processor hardware acceleration or to optimized software implementations of the encryption algorithms automatically.

  • On modern computers with multiple threads per core, simple statistics like %utilization measured in tools like iostat and vmstat are not "hard" indications of the resources that might be available for other processing. For example, 90% idle may not mean that 10 times the work can be done. So drawing numerical conclusions must be done carefully.

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of March 26, 2013.

SPARC T5-2 Obtains Oracle Internet Directory Benchmark World Record Performance

Oracle's SPARC T5-2 server running Oracle Internet Directory (OID, Oracle's LDAP Directory Server) on Oracle Solaris 11 achieved a record result for LDAP searches/second with 1000 clients.

  • The SPARC T5-2 server running Oracle Internet Directory on Oracle Solaris 11 achieved a result of 944,624 LDAP searches/sec with an average latency of 1.05 ms with 1000 clients.

  • The SPARC T5-2 server running Oracle Internet Directory demonstrated 2.7x better throughput and 39% better latency improvement over similarly configured OID and SPARC T4 benchmark environment.

  • The SPARC T5-2 server running Oracle Internet Directory demonstrates 39% better throughput and latency for LDAP searches on core-to-core comparison over an x86 system configured with two Intel Xeon X5675 processors.

  • Oracle Internet Directory achieved near linear scaling on the SPARC T5-2 server with 68,399 LDAP searches/sec with 2 cores to 944,624 LDAP searches/sec with 32 cores.

  • Oracle Internet Directory and the SPARC T5-2 server achieved up to 12,453 LDAP modifys/sec with an average latency of 3.9 msec for 50 clients.

Performance Landscape

Oracle Internet Directory Tests
System c/c/th Search Modify Add
ops/sec lat (msec) ops/sec lat (msec) ops/sec lat (msec)
SPARC T5-2 2/32/256 944,624 1.05 12,453 3.9 888 17.9
SPARC T4-4 4/32/256 682,000 1.46 12,000 4.0 835 19.0

In order to compare the SPARC T5-2 to a 12-core x86 system, only 1 processor and 12 cores was used in the SPARC T5-2.

Oracle Internet Directory Tests – Comparing Against x86
System c/c/th Search Compare Authentication
ops/sec lat (msec) ops/sec lat (msec) ops/sec lat (msec)
SPARC T5-2 1/12/96 417,000 1.19 274,185 1.82 149,623 3.30
x86 2 x Intel X5675 2/12/24 299,000 1.66 202,433 2.46 119,198 4.19

Scaling runs were also made on the SPARC T5-2 server.

Scaling of Search Tests – SPARC T5-2
Cores Clients ops/sec Latency (msec)
32 1000 944,624 1.05
24 1000 823,741 1.21
16 500 560,709 0.88
8 500 270,601 1.84
4 100 145,879 0.68
2 100 68,399 1.46

Configuration Summary

System Under Test:

SPARC T5-2
2 x SPARC T5 processors, 3.6 GHz
512 GB memory
4 x 300 GB internal disks
Flash Storage (used for database and log files)
1 x Sun Storage 2540-M2 (used for redo logs)
Oracle Solaris 11.1
Oracle Internet Directory 11g Release 1 PS6 (11.1.1.7.0)
Oracle Database 11g Enterprise Edition 11.2.0.3 (64-bit)

Benchmark Description

Oracle Internet Directory (OID) is Oracle's LDAPv3 Directory Server. The throughput for five key operations are measured — Search, Compare, Modify, Mix and Add.

LDAP Search Operations Test

This test scenario involved concurrent clients binding once to OID and then performing repeated LDAP Search operations. The salient characteristics of this test scenario is as follows:

  • SLAMD SearchRate job was used.
  • BaseDN of the search is root of the DIT, the scope is SUBTREE, the search filter is of the form UID=, DN and UID are the required attribute.
  • Each LDAP search operation matches a single entry.
  • The total number concurrent clients was 1000 and were distributed amongst two client nodes.
  • Each client binds to OID once and performs repeated LDAP Search operations, each search operation resulting in the lookup of a unique entry in such a way that no client looks up the same entry twice and no two clients lookup the same entry and all entries are searched randomly.
  • In one run of the test, random entries from the 50 Million entries are looked up in as many LDAP Search operations.
  • Test job was run for 60 minutes.

LDAP Compare Operations Test

This test scenario involved concurrent clients binding once to OID and then performing repeated LDAP Compare operations on userpassword attribute. The salient characteristics of this test scenario is as follows:

  • SLAMD CompareRate job was used.
  • Each LDAP compare operation matches user password of user.
  • The total number concurrent clients was 1000 and were distributed amongst two client nodes.
  • Each client binds to OID once and performs repeated LDAP compare operations.
  • In one run of the test, random entries from the 50 Million entries are compared in as many LDAP compare operations.
  • Test job was run for 60 minutes.

LDAP Modify Operations Test

This test scenario consisted of concurrent clients binding once to OID and then performing repeated LDAP Modify operations. The salient characteristics of this test scenario is as follows:

  • SLAMD LDAP modrate job was used.
  • A total of 50 concurrent LDAP clients were used.
  • Each client updates a unique entry each time and a total of 50 Million entries are updated.
  • Test job was run for 60 minutes.
  • Value length was set to 11.
  • Attribute that is being modified is not indexed.

LDAP Mixed Load Test

The test scenario involved both the LDAP search and LDAP modify clients enumerated above.

  • The ratio involved 60% LDAP search clients, 30% LDAP bind and 10% LDAP modify clients.
  • A total of 1000 concurrent LDAP clients were used and were distributed on 2 client nodes.
  • Test job was run for 60 minutes.

LDAP Add Load Test

The test scenario involved concurrent clients adding new entries as follows.

  • Slamd standard add rate job is used.
  • A total of 500,000 entries were added.
  • A total of 16 concurrent LDAP clients were used.
  • Slamd add's inetorgperson objectclass entry with 21 attributes (includes operational attributes).

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 26 March 2013.

SPARC T5-2 Scores Oracle FLEXCUBE Universal Banking Benchmark World Record Performance

Oracle's SPARC T5-2 server running Oracle FLEXCUBE Universal Banking Release 12 along with Oracle Database 11g Release 2 on Oracle Solaris 11 produced record results.

  • A SPARC T5-2 server running Oracle FLEXCUBE Universal Banking Release 12 and Oracle Real Application Clusters (RAC) Database 11g Release 2 processed 25 million accounts in 150 minutes for the End of Month workloads with an average utilization of 55% and 196 minutes utilizing 20 cores with an average cpu utilization of 85%.

  • A SPARC T5-2 server running Oracle FLEXCUBE Universal Banking Release 12 and Oracle Real Application Clusters (RAC) Database 11g Release 2 processed 25 million accounts in 56 minutes for the End of Day workload utilizing just 20 cores.

  • A SPARC T5-2 server running Oracle FLEXCUBE Universal Banking Release 12 achieved twice the throughput compared to a SPARC T4-4 server (which has twice the number of processors) for End of Month batch processing.

  • A SPARC T5-2 server running Oracle FLEXCUBE Universal Banking Release 12 achieved a record result processing 10.14 million accounts in 28 minutes for the End of Day workload with an average cpu utilization of 72% on a single server.

  • These results demonstrate how SPARC T5 processor systems along with Oracle Solaris 11 can benefit global, private and corporate financial institutions who are running Oracle FLEXCUBE Universal Banking. The uniquely co-engineered Oracle software and SPARC T5 processor based system unlock unique agile capabilities demanded by modern business environments.

  • The SPARC T5-2 system along with Oracle Solaris is able to provide a combination of uniquely essential characteristics that resonate with core values for a modern financial services institution.

  • The SPARC T5 processor based systems are capable of delivering higher performance and lower total cost of ownership (TCO) than older SPARC infrastructure, without introducing the unseen tax and risk of migrating applications away from older SPARC systems.

Performance Landscape

Oracle FLEXCUBE Universal Banking Release 12
End of Month Batch Processing
System Customer
Accounts
Time in Minutes Notes
SPARC T5-2 25M 150.66 RAC (two systems)
SPARC T5-2 10.14M 101.92 single instance
SPARC T4-4 10.14M 108.77 single instance
SPARC T4-4 5M 106.18 single instance, two chips

Oracle FLEXCUBE Universal Banking Release 12
End of Day Batch Processing
System Customer
Accounts
Time in Minutes Notes
SPARC T5-2 25M 56.05 RAC (two systems)
SPARC T5-2 10.14M 27.87 single instance

Configuration Summary

SPARC T5 Configuration:

1 x SPARC T5-2 with
2 x SPARC T5 processors, 3.6 GHz
512 GB memory
1 x SPARC T5-2 with
2 x SPARC T5 processors, 3.6 GHz
256 GB memory
Oracle Solaris 11 11/11
Oracle Database 11g Release 2 (RAC/ASM 11.2.0.3.0)
Oracle FLEXCUBE Universal Banking Release 12.0.1

SPARC T4 Configuration:

2 x SPARC T4-4, each with
4 x SPARC T4 processors, 3.0 GHz
512 GB memory
Oracle Solaris 11 11/11
Oracle Database 11g Release 2 (RAC/ASM 11.2.0.3.0)
Oracle FLEXCUBE Universal Banking Release 12.0.1

Storage Configuration:

3 x Sun Storage 6180 Array with
16 x 300 GB disks, 15K RPM (total of 48)
4 x Sun Storage CSM200 Expansion Trays, each with
16 x 73 GB disks, 15K RPM (total of 64)
Configured as RAID0, ASM external redundancy
Tests run with single instance DB (single node) and with ASM two nodes
ASM configuration identical on both 2 machines
Oracle Database 11g Release 2 ASM 11.2.0.3.0 64bit (19 TB)

Benchmark Description

The Oracle FLEXCUBE Universal Banking Release 12 benchmark models an actual customer bank with End of Cycle transaction batch jobs which typically execute during non-banking hours. This benchmark includes end of day accrual for savings and term deposit accounts, interest capitalization for saving accounts, and interest pay out for term deposit accounts. The results of the benchmark are certified by Oracle and a white paper is published.

End of cycle batch tests are conducted to measure the throughput capabilities of the system. It helps banks to decide the end of cycle processing window required to do the back office processing. The End of Day (EOD) batch test includes the following:

  • Mark End of Transaction Input
  • Value Dated Balance update
  • Interest and Charges (IC) Batch
  • Mark End of Financial Input
  • Mark End of Day
  • Date Change
  • Mark Transaction Input
The End of Month (EOM) batch test includes additional tests. These batches typically execute during non-banking hours. The goal is to ensure that the system is able to complete the batch operations for the planned volumes End of Day (EOD) within 60 minutes and End of Month (EOM) including interest and charges liquidation within 180 minutes.

 

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 26 March 2013.

Thursday Nov 08, 2012

SPARC T4-4 Delivers World Record Performance on Oracle OLAP Perf Version 2 Benchmark

Oracle's SPARC T4-4 server delivered world record performance with subsecond response time on the Oracle OLAP Perf Version 2 benchmark using Oracle Database 11g Release 2 running on Oracle Solaris 11.

  • The SPARC T4-4 server achieved throughput of 430,000 cube-queries/hour with an average response time of 0.85 seconds and the median response time of 0.43 seconds. This was achieved by using only 60% of the available CPU resources leaving plenty of headroom for future growth.

Performance Landscape

Oracle OLAP Perf Version 2 Benchmark
4 Billion Fact Table Rows
System Queries/
hour
Users* Response Time (sec)
Average Median
SPARC T4-4 430,000 7,300 0.85 0.43

* Users - the supported number of users with a given think time of 60 seconds

Configuration Summary and Results

Hardware Configuration:

SPARC T4-4 server with
4 x SPARC T4 processors, 3.0 GHz
1 TB memory
Data Storage
1 x Sun Fire X4275 (using COMSTAR)
2 x Sun Storage F5100 Flash Array (each with 80 FMODs)
Redo Storage
1 x Sun Fire X4275 (using COMSTAR with 8 HDD)

Software Configuration:

Oracle Solaris 11 11/11
Oracle Database 11g Release 2 (11.2.0.3) with Oracle OLAP option

Benchmark Description

The Oracle OLAP Perf Version 2 benchmark is a workload designed to demonstrate and stress the Oracle OLAP product's core features of fast query, fast update, and rich calculations on a multi-dimensional model to support enhanced Data Warehousing.

The bulk of the benchmark entails running a number of concurrent users, each issuing typical multidimensional queries against an Oracle OLAP cube. The cube has four dimensions: time, product, customer, and channel. Each query user issues approximately 150 different queries. One query chain may ask for total sales in a particular region (e.g South America) for a particular time period (e.g. Q4 of 2010) followed by additional queries which drill down into sales for individual countries (e.g. Chile, Peru, etc.) with further queries drilling down into individual stores, etc. Another query chain may ask for yearly comparisons of total sales for some product category (e.g. major household appliances) and then issue further queries drilling down into particular products (e.g. refrigerators, stoves. etc.), particular regions, particular customers, etc.

Results from version 2 of the benchmark are not comparable with version 1. The primary difference is the type of queries along with the query mix.

Key Points and Best Practices

  • Since typical BI users are often likely to issue similar queries, with different constants in the where clauses, setting the init.ora prameter "cursor_sharing" to "force" will provide for additional query throughput and a larger number of potential users. Except for this setting, together with making full use of available memory, out of the box performance for the OLAP Perf workload should provide results similar to what is reported here.

  • For a given number of query users with zero think time, the main measured metrics are the average query response time, the median query response time, and the query throughput. A derived metric is the maximum number of users the system can support achieving the measured response time assuming some non-zero think time. The calculation of the maximum number of users follows from the well-known response-time law

      N = (rt + tt) * tp

    where rt is the average response time, tt is the think time and tp is the measured throughput.

    Setting tt to 60 seconds, rt to 0.85 seconds and tp to 119.44 queries/sec (430,000 queries/hour), the above formula shows that the T4-4 server will support 7,300 concurrent users with a think time of 60 seconds and an average response time of 0.85 seconds.

    For more information see chapter 3 from the book "Quantitative System Performance" cited below.

See Also

Disclosure Statement

Copyright 2012, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 11/2/2012.

About

BestPerf is the source of Oracle performance expertise. In this blog, Oracle's Strategic Applications Engineering group explores Oracle's performance results and shares best practices learned from working on Enterprise-wide Applications.

Index Pages
Search

Archives
« July 2016
SunMonTueWedThuFriSat
     
1
2
3
4
5
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
      
Today