Tuesday Mar 26, 2013

SPARC T5-2 Scores Siebel CRM Benchmark World Record

Oracle set a new world record for the Siebel Platform Sizing and Performance Program (PSPP) benchmark using Oracle's SPARC T5-2 servers for the application server with Oracle's Siebel CRM 8.1.1.4 Industry Applications and Oracle Database 11g Release 2 running on Oracle Solaris.

  • The SPARC T5-2 servers running the application tier achieved 40,000 users with sub-second response time and with throughput of 333,339 business transactions per hour on the Siebel PSPP benchmark.

  • The SPARC T5-2 servers in the application tier delivered 2 times better performance on a per chip basis compared to earlier published SPARC T4 numbers.

  • The Siebel 8.1.1.4 PSPP workload includes Siebel Call Center and Order Management System.

  • The SPARC T5-2 server used Oracle Solaris Zones which provide flexible, scalable and manageable virtualization to scale the application within and across multiple nodes.

Performance Landscape

Application Server Transactions/
hour
Users Users/
Core
Call
Center
Order
Mgmt
Response Times (sec)
2 x SPARC T5-2 (2 x SPARC T5 3.6 GHz) 333,339 40,000 625 0.110 0.608
3 x SPARC T4-2 (2 x SPARC T4 2.85 GHz) 239,748 29,000 604 0.165 0.925
2 x IBM Power 750 (POWER7 3.55 GHz, 16 active cores) 176,185 21,000 656 0.052 0.250

Oracle:
Call Center + Order Management
Transactions: 273,786 + 59,553
Users: 28,000 + 12,000

IBM:
Call Center + Order Management
Transactions: 144,457 + 31,728
Users: 14,700 + 6,300

Configuration Summary

Application Server Configuration:

2 x SPARC T5-2 servers, each with
2 x SPARC T5 processors, 3.6 GHz
512 GB memory
6 x 300 GB SAS internal disks
Oracle Solaris 10 8/11
Siebel CRM 8.1.1.4 SIA

Web Server Configuration:

1 x SPARC T4-1 server
1 x SPARC T4 processor, 2.85 GHz
128 GB memory
Oracle Solaris 10 8/11
iPlanet Web Server 7

Database Server Configuration:

1 x SPARC T4-2 server
2 x SPARC T4 processors, 2.85 GHz
256 GB memory
Flash Storage
Oracle Solaris 10 8/11
Oracle Database 11g Release 2 (11.2.0.2)

Benchmark Description

Siebel PSPP benchmark includes Call Center and Order Management:

  • Siebel Financial Services Call Center – Provides the most complete solution for sales and service, allowing customer service and telesales representatives to provide superior customer support, improve customer loyalty, and increase revenues through cross-selling and up-selling.

    High-level description of the use cases tested: Incoming Call Creates Opportunity, Quote and Order and Incoming Call Creates Service Request. Three complex business transactions are executed simultaneously for specific number of concurrent users. The ratios of these 3 scenarios were 30%, 40%, 30% respectively, which together were totaling 70% of all transactions simulated in this benchmark. Between each user operation and the next one, the think time averaged approximately 10, 13, and 35 seconds respectively.

  • Siebel Order Management – Oracle's Siebel Order Management allows employees such as salespeople and call center agents to create and manage quotes and orders through their entire life cycle. Siebel Order Management can be tightly integrated with back-office applications allowing users to perform tasks such as checking credit, confirming availability, and monitoring the fulfillment process.

    High-level description of the use cases tested: Order & Order Items Creation and Order Updates. Two complex Order Management transactions were executed simultaneously for specific number of concurrent users concurrently with aforementioned three Call Center scenarios above. The ratio of these 2 scenarios was 50% each, which together were totaling 30% of all transactions simulated in this benchmark. Between each user operation and the next one, the think time averaged approximately 20 and 67 seconds respectively.

Key Points and Best Practices

  • No processor cores or cache were activated or deactivated on the SPARC T-Series systems to achieve special benchmark effects.

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 26 March 2013.

SPARC T5-2 Scores Oracle FLEXCUBE Universal Banking Benchmark World Record Performance

Oracle's SPARC T5-2 server running Oracle FLEXCUBE Universal Banking Release 12 along with Oracle Database 11g Release 2 on Oracle Solaris 11 produced record results.

  • A SPARC T5-2 server running Oracle FLEXCUBE Universal Banking Release 12 and Oracle Real Application Clusters (RAC) Database 11g Release 2 processed 25 million accounts in 150 minutes for the End of Month workloads with an average utilization of 55% and 196 minutes utilizing 20 cores with an average cpu utilization of 85%.

  • A SPARC T5-2 server running Oracle FLEXCUBE Universal Banking Release 12 and Oracle Real Application Clusters (RAC) Database 11g Release 2 processed 25 million accounts in 56 minutes for the End of Day workload utilizing just 20 cores.

  • A SPARC T5-2 server running Oracle FLEXCUBE Universal Banking Release 12 achieved twice the throughput compared to a SPARC T4-4 server (which has twice the number of processors) for End of Month batch processing.

  • A SPARC T5-2 server running Oracle FLEXCUBE Universal Banking Release 12 achieved a record result processing 10.14 million accounts in 28 minutes for the End of Day workload with an average cpu utilization of 72% on a single server.

  • These results demonstrate how SPARC T5 processor systems along with Oracle Solaris 11 can benefit global, private and corporate financial institutions who are running Oracle FLEXCUBE Universal Banking. The uniquely co-engineered Oracle software and SPARC T5 processor based system unlock unique agile capabilities demanded by modern business environments.

  • The SPARC T5-2 system along with Oracle Solaris is able to provide a combination of uniquely essential characteristics that resonate with core values for a modern financial services institution.

  • The SPARC T5 processor based systems are capable of delivering higher performance and lower total cost of ownership (TCO) than older SPARC infrastructure, without introducing the unseen tax and risk of migrating applications away from older SPARC systems.

Performance Landscape

Oracle FLEXCUBE Universal Banking Release 12
End of Month Batch Processing
System Customer
Accounts
Time in Minutes Notes
SPARC T5-2 25M 150.66 RAC (two systems)
SPARC T5-2 10.14M 101.92 single instance
SPARC T4-4 10.14M 108.77 single instance
SPARC T4-4 5M 106.18 single instance, two chips

Oracle FLEXCUBE Universal Banking Release 12
End of Day Batch Processing
System Customer
Accounts
Time in Minutes Notes
SPARC T5-2 25M 56.05 RAC (two systems)
SPARC T5-2 10.14M 27.87 single instance

Configuration Summary

SPARC T5 Configuration:

1 x SPARC T5-2 with
2 x SPARC T5 processors, 3.6 GHz
512 GB memory
1 x SPARC T5-2 with
2 x SPARC T5 processors, 3.6 GHz
256 GB memory
Oracle Solaris 11 11/11
Oracle Database 11g Release 2 (RAC/ASM 11.2.0.3.0)
Oracle FLEXCUBE Universal Banking Release 12.0.1

SPARC T4 Configuration:

2 x SPARC T4-4, each with
4 x SPARC T4 processors, 3.0 GHz
512 GB memory
Oracle Solaris 11 11/11
Oracle Database 11g Release 2 (RAC/ASM 11.2.0.3.0)
Oracle FLEXCUBE Universal Banking Release 12.0.1

Storage Configuration:

3 x Sun Storage 6180 Array with
16 x 300 GB disks, 15K RPM (total of 48)
4 x Sun Storage CSM200 Expansion Trays, each with
16 x 73 GB disks, 15K RPM (total of 64)
Configured as RAID0, ASM external redundancy
Tests run with single instance DB (single node) and with ASM two nodes
ASM configuration identical on both 2 machines
Oracle Database 11g Release 2 ASM 11.2.0.3.0 64bit (19 TB)

Benchmark Description

The Oracle FLEXCUBE Universal Banking Release 12 benchmark models an actual customer bank with End of Cycle transaction batch jobs which typically execute during non-banking hours. This benchmark includes end of day accrual for savings and term deposit accounts, interest capitalization for saving accounts, and interest pay out for term deposit accounts. The results of the benchmark are certified by Oracle and a white paper is published.

End of cycle batch tests are conducted to measure the throughput capabilities of the system. It helps banks to decide the end of cycle processing window required to do the back office processing. The End of Day (EOD) batch test includes the following:

  • Mark End of Transaction Input
  • Value Dated Balance update
  • Interest and Charges (IC) Batch
  • Mark End of Financial Input
  • Mark End of Day
  • Date Change
  • Mark Transaction Input
The End of Month (EOM) batch test includes additional tests. These batches typically execute during non-banking hours. The goal is to ensure that the system is able to complete the batch operations for the planned volumes End of Day (EOD) within 60 minutes and End of Month (EOM) including interest and charges liquidation within 180 minutes.

 

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 26 March 2013.

SPARC T5-1B Performance Running Oracle Communications ASAP

Oracle's SPARC T5-1B server module delivered outstanding results on Oracle Communications ASAP. The SPARC T5-1B server module ran Oracle Solaris 11 with Oracle Database 11g Release 2, Oracle WebLogic Server 11g and Oracle Communications ASAP version 7.2.

  • Running Oracle Communications ASAP, the SPARC T5-1B server module achieved 1,722 ASDLs (atomic network activation actions) per second, the highest throughput that has been achieved in the 12NEP test for a single Oracle Communications ASAP instance across any SPARC architecture.

  • The SPARC T5-1B server module running a single instance of the Oracle Communications ASAP application, with both the application and database tiers consolidated onto a single machine, easily supported the service activation volumes of 1,722 ASDLs/sec which is representative of a typical mobile operator with more than 100 million subscribers.

  • Oracle Communications ASAP v7.2 delivered 48% higher throughput on a the SPARC T5-1B server module when compared to the SPARC T4-2 server.

  • The SPARC T5 processor delivered over 2 times the throughput compared to the previous generation SPARC T4 processor.

Performance Landscape

ASAP 7.2.0 12NEP Test Results
System ASDLs/sec CPU Usage
SPARC T5-1B 1,722.2 44.8%
SPARC T4-2 1,114.3 42.7%

Configuration Summary

Hardware Configuration:

SPARC T5-1B server module
1 x SPARC T5 processor at 3.6 GHz
256 GB memory

SPARC T4-2 server
2 x SPARC T4 processors at 2.85 GHz
256 GB memory

Storage Configuration:

Pillar Axiom

Software Configuration:

Oracle Solaris 11.1
Oracle Database 11g Release 2 (11.2.0.3.0)
Oracle WebLogic Server 10.3.6.0
Oracle Communications ASAP 7.2.0 (SR2B23)
Oracle JDK 7 update 7

Benchmark Description

Oracle Communications ASAP is used to activate a variety of services including data, video, voice and content services across mobile, fixed and satellite networks. Typical activities performed include activating new subscribers and services, moving / adding / changing / deleting services of existing subscribers and deleting existing subscribers and services.

The throughput of ASAP is measured in atomic actions per second (or ASDLs/sec). An atomic action is a single command or operation that can be executed on a network element. Atomic actions are grouped together to form a common service action, where each service action typically relates to an orderable item, such as "GSM voice" or "voice mail" or "GSM data". One or more service actions are invoked by an order management system via an activation work order request.

The workload resembles a typical mobile order to activate a GSM subscriber. A single service action to add a subscriber consists of seven atomic actions where each atomic action executes a command on a network element. Each network element was serviced by a dedicated Network Element Processor (NEP). The ASAP benchmark can vary the number of NEPs, which correlate to the complexity of a Telco operator's environment.

See Also

Disclosure Statement

Copyright 2013, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 26 March 2013.

Tuesday Oct 02, 2012

World Record Oracle E-Business Consolidated Workload on SPARC T4-2

Oracle set a World Record for the Oracle E-Business Suite Standard Medium multiple-online module benchmark using Oracle's SPARC T4-2 and SPARC T4-4 servers which ran the application and database.

  • Oracle's SPARC T4 servers demonstrate performance leadership and world-record results on Oracle E-Business Suite Applications R12 OLTP benchmark by publishing the first result using multiple concurrent online application modules with Oracle Database 11g Release 2 running Solaris.

  •  

  • This results shows that a multi-tier configuration of SPARC T4 servers running the Oracle E-Business Suite R12.1.2 application and Oracle Database 11g Release 2 is capable of supporting 4,100 online users with outstanding response-times, executing a mix of complex transactions consolidating 4 Oracle E-Business modules (iProcurement, Order Management, Customer Service and HR Self-Service).

  •  

  • The SPARC T4-2 server in the application tier utilized about 65% and the SPARC T4-4 server in the database tier utilized about 30%, providing significant headroom for additional Oracle E-Business Suite R12.1.2 processing modules, more online users, and future growth.

  •  

  • Oracle E-Business Suite Applications were run in Oracle Solaris Containers on SPARC T4 servers and provides a consolidation platform for multiple E-Business instances.

  •  

Performance Landscape

Multiple Online Modules (Self-Service, Order-Management, iProcurement, Customer-Service)
Medium Configuration
System Users Average
Response Time
90th Percentile
Response Time
SPARC T4-2 4,100 2.08 sec 2.52 sec

Configuration Summary

Application Tier Configuration:

1 x SPARC T4-2 server
2 x SPARC T4 processors, 2.85 GHz
256 GB memory
3 x 300 GB internal disks
Oracle Solaris 10
Oracle E-Business Suite 12.1.2

Database Tier Configuration:

1 x SPARC T4-4 server
4 x SPARC T4 processors, 3.0 GHz
256 GB memory
2 x 300 GB internal disks
Oracle Solaris 10
Oracle Solaris Containers
Oracle Database 11g Release 2

Storage Configuration:

1 x Sun Storage F5100 Flash Array (80 x 24 GB flash modules)

Benchmark Description

The Oracle R12 E-Business Suite Standard Benchmark combines online transaction execution by simulated users with multiple online concurrent modules to model a typical scenario for a global enterprise. The online component exercises the common UI flows which are most frequently used by a majority of our customers. This benchmark utilized four concurrent flows of OLTP transactions, for Order to Cash, iProcurement, Customer Service and HR Self-Service and measured the response times. The selected flows model simultaneous business activities inclusive of managing customers, services, products and employees.

See Also

Disclosure Statement

Oracle E-Business Suite R12 medium multiple-online module benchmark, SPARC T4-2, SPARC T4, 2.85 GHz, 2 chips, 16 cores, 128 threads, 256 GB memory, SPARC T4-4, SPARC T4, 3.0 GHz, 4 chips, 32 cores, 256 threads, 256 GB memory, average response time 2.08 sec, 90th percentile response time 2.52 sec, Oracle Solaris 10, Oracle Solaris Containers, Oracle E-Business Suite 12.1.2, Oracle Database 11g Release 2, Results as of 9/30/2012.

Monday Oct 01, 2012

World Record Batch Rate on Oracle JD Edwards Consolidated Workload with SPARC T4-2

Oracle produced a World Record batch throughput for single system results on Oracle's JD Edwards EnterpriseOne Day-in-the-Life benchmark using Oracle's SPARC T4-2 server running Oracle Solaris Containers and consolidating JD Edwards EnterpriseOne, Oracle WebLogic servers and the Oracle Database 11g Release 2. The workload includes both online and batch workload.

  • The SPARC T4-2 server delivered a result of 8,000 online users while concurrently executing a mix of JD Edwards EnterpriseOne Long and Short batch processes at 95.5 UBEs/min (Universal Batch Engines per minute).

  • In order to obtain this record benchmark result, the JD Edwards EnterpriseOne, Oracle WebLogic and Oracle Database 11g Release 2 servers were executed each in separate Oracle Solaris Containers which enabled optimal system resources distribution and performance together with scalable and manageable virtualization.

  • One SPARC T4-2 server running Oracle Solaris Containers and consolidating JD Edwards EnterpriseOne, Oracle WebLogic servers and the Oracle Database 11g Release 2 utilized only 55% of the available CPU power.

  • The Oracle DB server in a Shared Server configuration allows for optimized CPU resource utilization and significant memory savings on the SPARC T4-2 server without sacrificing performance.

  • This configuration with SPARC T4-2 server has achieved 33% more Users/core, 47% more UBEs/min and 78% more Users/rack unit than the IBM Power 770 server.

  • The SPARC T4-2 server with 2 processors ran the JD Edwards "Day-in-the-Life" benchmark and supported 8,000 concurrent online users while concurrently executing mixed batch workloads at 95.5 UBEs per minute. The IBM Power 770 server with twice as many processors supported only 12,000 concurrent online users while concurrently executing mixed batch workloads at only 65 UBEs per minute.

  • This benchmark demonstrates more than 2x cost savings by consolidating the complete solution in a single SPARC T4-2 server compared to earlier published results of 10,000 users and 67 UBEs per minute on two SPARC T4-2 and SPARC T4-1.

  • The Oracle DB server used mirrored (RAID 1) volumes for the database providing high availability for the data without impacting performance.

Performance Landscape

JD Edwards EnterpriseOne Day in the Life (DIL) Benchmark
Consolidated Online with Batch Workload

System Rack
Units
(U)
Batch
Rate
(UBEs/m)
Online
Users
Users
/ U
Users
/ Core
Version
SPARC T4-2 (2 x SPARC T4, 2.85 GHz) 3 95.5 8,000 2,667 500 9.0.2
IBM Power 770 (4 x POWER7, 3.3 GHz, 32 cores) 8 65 12,000 1,500 375 9.0.2

Batch Rate (UBEs/m) — Batch transaction rate in UBEs per minute

Configuration Summary

Hardware Configuration:

1 x SPARC T4-2 server with
2 x SPARC T4 processors, 2.85 GHz
256 GB memory
4 x 300 GB 10K RPM SAS internal disk
2 x 300 GB internal SSD
2 x Sun Storage F5100 Flash Arrays

Software Configuration:

Oracle Solaris 10
Oracle Solaris Containers
JD Edwards EnterpriseOne 9.0.2
JD Edwards EnterpriseOne Tools (8.98.4.2)
Oracle WebLogic Server 11g (10.3.4)
Oracle HTTP Server 11g
Oracle Database 11g Release 2 (11.2.0.1)

Benchmark Description

JD Edwards EnterpriseOne is an integrated applications suite of Enterprise Resource Planning (ERP) software. Oracle offers 70 JD Edwards EnterpriseOne application modules to support a diverse set of business operations.

Oracle's Day in the Life (DIL) kit is a suite of scripts that exercises most common transactions of JD Edwards EnterpriseOne applications, including business processes such as payroll, sales order, purchase order, work order, and manufacturing processes, such as ship confirmation. These are labeled by industry acronyms such as SCM, CRM, HCM, SRM and FMS. The kit's scripts execute transactions typical of a mid-sized manufacturing company.

  • The workload consists of online transactions and the UBE – Universal Business Engine workload of 61 short and 4 long UBEs.

  • LoadRunner runs the DIL workload, collects the user’s transactions response times and reports the key metric of Combined Weighted Average Transaction Response time.

  • The UBE processes workload runs from the JD Enterprise Application server.

    • Oracle's UBE processes come as three flavors:

      • Short UBEs < 1 minute engage in Business Report and Summary Analysis,

      • Mid UBEs > 1 minute create a large report of Account, Balance, and Full Address,

      • Long UBEs > 2 minutes simulate Payroll, Sales Order, night only jobs.

    • The UBE workload generates large numbers of PDF files reports and log files.

    • The UBE Queues are categorized as the QBATCHD, a single threaded queue for large and medium UBEs, and the QPROCESS queue for short UBEs run concurrently.

Oracle's UBE process performance metric is Number of Maximum Concurrent UBE processes at transaction rate, UBEs/minute.

Key Points and Best Practices

Two JD Edwards EnterpriseOne Application Servers, two Oracle WebLogic Servers 11g Release 1 coupled with two Oracle Web Tier HTTP server instances and one Oracle Database 11g Release 2 database on a single SPARC T4-2 server were hosted in separate Oracle Solaris Containers bound to four processor sets to demonstrate consolidation of multiple applications, web servers and the database with best resource utilizations.

  • Interrupt fencing was configured on all Oracle Solaris Containers to channel the interrupts to processors other than the processor sets used for the JD Edwards Application server, Oracle WebLogic servers and the database server.

  • A Oracle WebLogic vertical cluster was configured on each WebServer Container with twelve managed instances each to load balance users' requests and to provide the infrastructure that enables scaling to high number of users with ease of deployment and high availability.

  • The database log writer was run in the real time RT class and bound to a processor set.

  • The database redo logs were configured on the raw disk partitions.

  • The Oracle Solaris Container running the Enterprise Application server completed 61 Short UBEs, 4 Long UBEs concurrently as the mixed size batch workload.

  • The mixed size UBEs ran concurrently from the Enterprise Application server with the 8,000 online users driven by the LoadRunner.

See Also

Disclosure Statement

Copyright 2012, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 09/30/2012.

Tuesday Aug 28, 2012

SPARC T4-2 Produces World Record Oracle Essbase Aggregate Storage Benchmark Result

Significance of Results

Oracle's SPARC T4-2 server configured with a Sun Storage F5100 Flash Array and running Oracle Solaris 10 with Oracle Database 11g has achieved exceptional performance for the Oracle Essbase Aggregate Storage Option benchmark. The benchmark has upwards of 1 billion records, 15 dimensions and millions of members. Oracle Essbase is a multi-dimensional online analytical processing (OLAP) server and is well-suited to work well with SPARC T4 servers.

  • The SPARC T4-2 server (2 cpus) running Oracle Essbase 11.1.2.2.100 outperformed the previous published results on Oracle's SPARC Enterprise M5000 server (4 cpus) with Oracle Essbase 11.1.1.3 on Oracle Solaris 10 by 80%, 32% and 2x performance improvement on Data Loading, Default Aggregation and Usage Based Aggregation, respectively.

  • The SPARC T4-2 server with Sun Storage F5100 Flash Array and Oracle Essbase running on Oracle Solaris 10 achieves sub-second query response times for 20,000 users in a 15 dimension database.

  • The SPARC T4-2 server configured with Oracle Essbase was able to aggregate and store values in the database for a 15 dimension cube in 398 minutes with 16 threads and in 484 minutes with 8 threads.

  • The Sun Storage F5100 Flash Array provides more than a 20% improvement out-of-the-box compared to a mid-size fiber channel disk array for default aggregation and user-based aggregation.

  • The Sun Storage F5100 Flash Array with Oracle Essbase provides the best combination for large Oracle Essbase databases leveraging Oracle Solaris ZFS and taking advantage of high bandwidth for faster load and aggregation.

  • Oracle Fusion Middleware provides a family of complete, integrated, hot pluggable and best-of-breed products known for enabling enterprise customers to create and run agile and intelligent business applications. Oracle Essbase's performance demonstrates why so many customers rely on Oracle Fusion Middleware as their foundation for innovation.

Performance Landscape

System Data Size
(millions of items)
Database
Load
(minutes)
Default
Aggregation
(minutes)
Usage Based
Aggregation
(minutes)
SPARC T4-2, 2 x SPARC T4 2.85 GHz 1000 149 398* 55
Sun M5000, 4 x SPARC64 VII 2.53 GHz 1000 269 526 115
Sun M5000, 4 x SPARC64 VII 2.4 GHz 400 120 448 18

* – 398 mins with CALCPARALLEL set to 16; 484 mins with CALCPARALLEL threads set to 8

Configuration Summary

Hardware Configuration:

1 x SPARC T4-2
2 x 2.85 GHz SPARC T4 processors
128 GB memory
2 x 300 GB 10000 RPM SAS internal disks

Storage Configuration:

1 x Sun Storage F5100 Flash Array
40 x 24 GB flash modules
SAS HBA with 2 SAS channels
Data Storage Scheme Striped - RAID 0
Oracle Solaris ZFS

Software Configuration:

Oracle Solaris 10 8/11
Installer V 11.1.2.2.100
Oracle Essbase Client v 11.1.2.2.100
Oracle Essbase v 11.1.2.2.100
Oracle Essbase Administration services 64-bit
Oracle Database 11g Release 2 (11.2.0.3)
HP's Mercury Interactive QuickTest Professional 9.5.0

Benchmark Description

The objective of the Oracle Essbase Aggregate Storage Option benchmark is to showcase the ability of Oracle Essbase to scale in terms of user population and data volume for large enterprise deployments. Typical administrative and end-user operations for OLAP applications were simulated to produce benchmark results.

The benchmark test results include:

  • Database Load: Time elapsed to build a database including outline and data load.
  • Default Aggregation: Time elapsed to build aggregation.
  • User Based Aggregation: Time elapsed of the aggregate views proposed as a result of tracked retrieval queries.

Summary of the data used for this benchmark:

  • 40 flat files, each of size 1.2 GB, 49.4 GB in total
  • 10 million rows per file, 1 billion rows total
  • 28 columns of data per row
  • Database outline has 15 dimensions (five of them are attribute dimensions)
  • Customer dimension has 13.3 million members
  • 3 rule files

Key Points and Best Practices

  • The Sun Storage F5100 Flash Array has been used to accelerate the application performance.

  • Setting data load threads (DLTHREADSPREPARE) to 64 and Load Buffer to 6 improved dataloading by about 9%.

  • Factors influencing aggregation materialization performance are "Aggregate Storage Cache" and "Number of Threads" (CALCPARALLEL) for parallel view materialization. The optimal values for this workload on the SPARC T4-2 server were:

      Aggregate Storage Cache: 32 GB
      CALCPARALLEL: 16

     

See Also

Disclosure Statement

Copyright 2012, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 28 August 2012.

Thursday Apr 19, 2012

Sun ZFS Storage 7420 Appliance Delivers 2-Node World Record SPECsfs2008 NFS Benchmark

Oracle's Sun ZFS Storage 7420 appliance delivered world record two-node performance on the SPECsfs2008 NFS benchmark, beating results published on NetApp's dual-controller and 4-node high-end FAS6240 storage systems.

  • The Sun ZFS Storage 7420 appliance delivered a world record two-node result of 267,928 SPECsfs2008_nfs.v3 Ops/sec with an Overall Response Time (ORT) of 1.31 msec on the SPECsfs2008 NFS benchmark.

  • The Sun ZFS Storage 7420 appliance delivered 1.4x higher throughput than the dual-controller NetApp FAS6240 and 2.6x higher throughput than the dual-controller NetApp FAS3270 on the SPECsfs2008_nfs.v3 benchmark at less than half the list price of either result.

  • The Sun ZFS Storage 7420 appliance required 10 percent less rack space than the dual-controller NetApp FAS6240.

  • The Sun ZFS Storage 7420 appliance had 3 percent higher throughput than the 4-node NetApp FAS6240 on the SPECsfs2008_nfs.v3 benchmark.

  • The Sun ZFS Storage 7420 appliance required 25 percent less rack space than the 4-node NetApp FAS6240.

  • The Sun ZFS Storage 7420 appliance has 14 percent better Overall Response Time than the 4-node NetApp FAS6240 on the SPECsfs2008_nfs.v3 benchmark.

Performance Landscape

SPECsfs2008_nfs.v3 Performance Chart (in decreasing SPECsfs2008_nfs.v3 Ops/sec order)

Sponsor System Throughput
(Ops/sec)
Overall Response
Time (msec)
Nodes Memory (GB)
Including Flash
Disks Rack Units –
Controllers
+Disks
Oracle 7420 267,928 1.31 2 6,728 280 54
NetApp FAS6240 260,388 1.53 4 2,256 288 72
NetApp FAS6240 190,675 1.17 2 1,128 288 60
EMC VG8 135,521 1.92 280 312
Oracle 7320 134,140 1.51 2 4,968 136 26
EMC NS-G8 110,621 2.32 264 100
NetApp FAS3270 101,183 1.66 2 40 360 66

Throughput SPECsfs2008_nfs.v3 Ops/sec — the Performance Metric
Overall Response Time — the corresponding Response Time Metric
Nodes — Nodes and Controllers are being used interchangeably

Complete SPECsfs2008 benchmark results may be found at http://www.spec.org/sfs2008/results/sfs2008.html.

Configuration Summary

Storage Configuration:

Sun ZFS Storage 7420 appliance in clustered configuration
2 x Sun ZFS Storage 7420 controllers, each with
4 x 2.4 GHz Intel Xeon E7-4870 processors
1 TB memory
4 x 512 GB SSD flash-enabled read-cache
2 x 10GbE NICs
12 x Sun Disk shelves
10 x shelves with 24 x 300 GB 15K RPM SAS-2 drives
2 x shelves with 20 x 300 GB 15K RPM SAS-2 drives and 4 x 73 GB SAS-2 flash-enabled write-cache

Server Configuration:

4 x Sun Fire X4270 M2 servers, each with
2 x 3.3 GHz Intel Xeon E5680 processors
144 GB memory
1 x 10 GbE NIC
Oracle Solaris 10 9/10

Switches:

1 x 24-port 10Gb Ethernet Switch

Benchmark Description

SPECsfs2008 is the latest version of the Standard Performance Evaluation Corporation (SPEC) benchmark suite measuring file server throughput and response time, providing a standardized method for comparing performance across different vendor platforms. SPECsfs2008 results summarize the server's capabilities with respect to the number of operations that can be handled per second, as well as the overall latency of the operations. The suite is a follow-on to the SFS97_R1 benchmark, adding a CIFS workload, an updated NFSv3 workload, support for additional client platforms, and a new test harness and reporting/submission framework.

See Also

Disclosure Statement

SPEC and SPECsfs are registered trademarks of Standard Performance Evaluation Corporation (SPEC). Results as of April 18, 2012, for more information see www.spec.org. Sun ZFS Storage 7420 Appliance 267,928 SPECsfs2008_nfs.v3 Ops/sec, 1.31 msec ORT, NetApp Data ONTAP 8.1 Cluster-Mode (4-node FAS6240) 260,388 SPECsfs2008_nfs.v3 Ops/Sec, 1.53 msec ORT, NetApp FAS6240 190,675 SPECsfs2008_nfs.v3 Ops/Sec, 1.17 msec ORT. NetApp FAS3270 101,183 SPECsfs2008_nfs.v3 Ops/Sec, 1.66 msec ORT.

Nodes refer to the item in the SPECsfs2008 disclosed Configuration Bill of Materials that have the Processing Elements that perform the NFS Processing Function. These are the first item listed in each of disclosed Configuration Bill of Materials except for EMC where it is both the first and third items listed, and HP, where it is the second item listed as Blade Servers. The number of nodes is from the QTY disclosed in the Configuration Bill of Materials as described above. Configuration Bill of Materials list price for Oracle result of US$ 423,644. Configuration Bill of Materials list price for NetApp FAS3270 result of US$ 1,215,290. Configuration Bill of Materials list price for NetApp FAS6240 result of US$ 1,028,118. Oracle pricing from https://shop.oracle.com/pls/ostore/f?p=dstore:home:0, traverse to "Storage and Tape" and then to "NAS Storage". NetApp's pricing from http://www.netapp.com/us/media/na-list-usd-netapp-custom-state-new-discounts.html.

Sunday Apr 15, 2012

Sun ZFS Storage 7420 Appliance Delivers Top High-End Price/Performance Result for SPC-2 Benchmark

Oracle's Sun ZFS Storage 7420 appliance delivered leading high-end price/performance on the SPC Benchmark 2 (SPC-2).

  • The Sun ZFS Storage 7420 appliance delivered a result of 10,704 SPC-2 MB/s at $35.24 $/SPC-2 MB/s on the SPC-2 benchmark.

  • The Sun ZFS Storage 7420 appliance beats the IBM DS8800 result by over 10% on SPC-2 MB/s and has 7.7x better $/SPC-2 MB/s.

  • The Sun ZFS Storage 7420 appliance achieved the best price/performance for the top 18 posted unique performance results on the SPC-2 benchmark.

Performance Landscape

SPC-2 Performance Chart (in decreasing performance order)

System SPC-2
MB/s
$/SPC-2
MB/s
ASU
Capacity
(GB)
TSC Price Data
Protection
Level
Date Results
Identifier
HP StorageWorks P9500 13,148 $88.34 129,112 $1,161,504 RAID-5 03/07/12 B00056
Sun ZFS Storage 7420 10,704 $35.24 31,884 $377,225 Mirroring 04/12/12 B00058
IBM DS8800 9,706 $270.38 71,537 $2,624,257 RAID-5 12/01/10 B00051
HP XP24000 8,725 $187.45 18,401 $1,635,434 Mirroring 09/08/08 B00035
Hitachi Storage Platform V 8,725 $187.49 18,401 $1,635,770 Mirroring 09/08/08 B00036
TMS RamSan-630 8,323 $49.37 8,117 $410,927 RAID-5 05/10/11 B00054
IBM XIV 7,468 $152.34 154,619 $1,137,641 RAID-1 10/19/11 BE00001
IBM DS8700 7,247 $277.22 32,642 $2,009,007 RAID-5 11/30/09 B00049
IBM SAN Vol Ctlr 4.2 7,084 $463.66 101,155 $3,284,767 RAID-5 07/12/07 B00024
Fujitsu ETERNUS DX440 S2 5,768 $66.50 42,133 $383,576 Mirroring 04/12/12 B00057
IBM DS5300 5,634 $74.13 16,383 $417,648 RAID-5 10/21/09 B00045
Sun Storage 6780 5,634 $47.03 16,383 $264,999 RAID-5 10/28/09 B00047
IBM DS5300 5,544 $75.33 14,043 $417,648 RAID-6 10/21/09 B00046
Sun Storage 6780 5,544 $47.80 14,043 $264,999 RAID-6 10/28/09 B00048
IBM DS5300 4,818 $93.80 16,383 $451,986 RAID-5 09/25/08 B00037
Sun Storage 6780 4,818 $53.61 16,383 $258,329 RAID-5 02/02/09 B00039
IBM DS5300 4,676 $96.67 14,043 $451,986 RAID-6 09/25/08 B00038
Sun Storage 6780 4,676 $55.25 14,043 $258,329 RAID-6 02/03/09 B00040
IBM SAN Vol Ctlr 4.1 4,544 $400.78 51,265 $1,821,301 RAID-5 09/12/06 B00011
IBM SAN Vol Ctlr 3.1 3,518 $563.93 20,616 $1,983,785 Mirroring 12/14/05 B00001
Fujitsu ETERNUS8000 1100 3,481 $238.93 4,570 $831,649 Mirroring 03/08/07 B00019
IBM DS8300 3,218 $539.38 15,393 $1,735,473 Mirroring 12/14/05 B00006
IBM Storwize V7000 3,133 $71.32 29,914 $223,422 RAID-5 12/13/10 B00052

SPC-2 MB/s = the Performance Metric
$/SPC-2 MB/s = the Price/Performance Metric
ASU Capacity = the Capacity Metric
Data Protection = Data Protection Metric
TSC Price = Total Cost of Ownership Metric
Results Identifier = A unique identification of the result Metric

Complete SPC-2 benchmark results may be found at http://www.storageperformance.org.

Configuration Summary

Storage Configuration:

Sun ZFS Storage 7420 appliance in clustered configuration
2 x Sun ZFS Storage 7420 controllers, each with
4 x 2.0 GHz Intel Xeon X7550 processors
512 GB memory, 64 x 8 GB 1066 MHz DDR3 DIMMs
16 x Sun Disk shelves, each with
24 x 300 GB 15K RPM SAS-2 drives

Server Configuration:

1 x Sun Fire X4470 server, with
4 x 2.4 GHz Intel Xeon E7-4870 processors
512 GB memory
8 x 8 Gb FC connections to the Sun ZFS Storage 7420 appliance
Oracle Solaris 11 11/11

2 x Sun Fire X4470 servers, each with
4 x 2.4 GHz Intel Xeon E7-4870 processors
256 GB memory
8 x 8 Gb FC connections to the Sun ZFS Storage 7420 appliance
Oracle Solaris 11 11/11

Benchmark Description

SPC Benchmark-2 (SPC-2): Consists of three distinct workloads designed to demonstrate the performance of a storage subsystem during the execution of business critical applications that require the large-scale, sequential movement of data. Those applications are characterized predominately by large I/Os organized into one or more concurrent sequential patterns. A description of each of the three SPC-2 workloads is listed below as well as examples of applications characterized by each workload.

  • Large File Processing: Applications in a wide range of fields, which require simple sequential process of one or more large files such as scientific computing and large-scale financial processing.
  • Large Database Queries: Applications that involve scans or joins of large relational tables, such as those performed for data mining or business intelligence.
  • Video on Demand: Applications that provide individualized video entertainment to a community of subscribers by drawing from a digital film library.

SPC-2 is built to:

  • Provide a level playing field for test sponsors.
  • Produce results that are powerful and yet simple to use.
  • Provide value for engineers as well as IT consumers and solution integrators.
  • Is easy to run, easy to audit/verify, and easy to use to report official results.

See Also

Disclosure Statement

SPC-2, SPC-2 MB/s, $/SPC-2 MB/s are registered trademarks of Storage Performance Council (SPC). Results as of April 12, 2012, for more information see www.storageperformance.org. Sun ZFS Storage 7420 Appliance http://www.storageperformance.org/results/benchmark_results_spc2#b00058; IBM DS8800 http://www.storageperformance.org/results/benchmark_results_spc2#b00051.

Thursday Apr 12, 2012

Sun Fire X4270 M3 SAP Enhancement Package 4 for SAP ERP 6.0 (Unicode) Two-Tier Standard Sales and Distribution (SD) Benchmark

Oracle's Sun Fire X4270 M3 server (now known as Sun Server X3-2L) achieved 8,320 SAP SD Benchmark users running SAP enhancement package 4 for SAP ERP 6.0 with unicode software using Oracle Database 11g and Oracle Solaris 10.

  • The Sun Fire X4270 M3 server using Oracle Database 11g and Oracle Solaris 10 beat both IBM Flex System x240 and IBM System x3650 M4 server running DB2 9.7 and Windows Server 2008 R2 Enterprise Edition.

  • The Sun Fire X4270 M3 server running Oracle Database 11g and Oracle Solaris 10 beat the HP ProLiant BL460c Gen8 server using SQL Server 2008 and Windows Server 2008 R2 Enterprise Edition by 6%.

  • The Sun Fire X4270 M3 server using Oracle Database 11g and Oracle Solaris 10 beat Cisco UCS C240 M3 server running SQL Server 2008 and Windows Server 2008 R2 Datacenter Edition by 9%.

  • The Sun Fire X4270 M3 server running Oracle Database 11g and Oracle Solaris 10 beat the Fujitsu PRIMERGY RX300 S7 server using SQL Server 2008 and Windows Server 2008 R2 Enterprise Edition by 10%.

Performance Landscape

SAP-SD 2-Tier Performance Table (in decreasing performance order).

SAP ERP 6.0 Enhancement Pack 4 (Unicode) Results
(benchmark version from January 2009 to April 2012)

System OS
Database
Users SAP
ERP/ECC
Release
SAPS SAPS/
Proc
Date
Sun Fire X4270 M3
2xIntel Xeon E5-2690 @2.90GHz
128 GB
Oracle Solaris 10
Oracle Database 11g
8,320 2009
6.0 EP4
(Unicode)
45,570 22,785 10-Apr-12
IBM Flex System x240
2xIntel Xeon E5-2690 @2.90GHz
128 GB
Windows Server 2008 R2 EE
DB2 9.7
7,960 2009
6.0 EP4
(Unicode)
43,520 21,760 11-Apr-12
HP ProLiant BL460c Gen8
2xIntel Xeon E5-2690 @2.90GHz
128 GB
Windows Server 2008 R2 EE
SQL Server 2008
7,865 2009
6.0 EP4
(Unicode)
42,920 21,460 29-Mar-12
IBM System x3650 M4
2xIntel Xeon E5-2690 @2.90GHz
128 GB
Windows Server 2008 R2 EE
DB2 9.7
7,855 2009
6.0 EP4
(Unicode)
42,880 21,440 06-Mar-12
Cisco UCS C240 M3
2xIntel Xeon E5-2690 @2.90GHz
128 GB
Windows Server 2008 R2 DE
SQL Server 2008
7,635 2009
6.0 EP4
(Unicode)
41,800 20,900 06-Mar-12
Fujitsu PRIMERGY RX300 S7
2xIntel Xeon E5-2690 @2.90GHz
128 GB
Windows Server 2008 R2 EE
SQL Server 2008
7,570 2009
6.0 EP4
(Unicode)
41,320 20,660 06-Mar-12

Complete benchmark results may be found at the SAP benchmark website http://www.sap.com/benchmark.

Configuration and Results Summary

Hardware Configuration:

Sun Fire X4270 M3
2 x 2.90 GHz Intel Xeon E5-2690 processors
128 GB memory
Sun StorageTek 6540 with 4 * 16 * 300GB 15Krpm 4Gb FC-AL

Software Configuration:

Oracle Solaris 10
Oracle Database 11g
SAP enhancement package 4 for SAP ERP 6.0 (Unicode)

Certified Results (published by SAP):

Number of benchmark users:
8,320
Average dialog response time:
0.95 seconds
Throughput:

Fully processed order line:
911,330

Dialog steps/hour:
2,734,000

SAPS:
45,570
SAP Certification:
2012014

Benchmark Description

The SAP Standard Application SD (Sales and Distribution) Benchmark is a two-tier ERP business test that is indicative of full business workloads of complete order processing and invoice processing, and demonstrates the ability to run both the application and database software on a single system. The SAP Standard Application SD Benchmark represents the critical tasks performed in real-world ERP business environments.

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

Two-tier SAP Sales and Distribution (SD) standard SAP SD benchmark based on SAP enhancement package 4 for SAP ERP 6.0 (Unicode) application benchmark as of 04/11/12: Sun Fire X4270 M3 (2 processors, 16 cores, 32 threads) 8,320 SAP SD Users, 2 x 2.90 GHz Intel Xeon E5-2690, 128 GB memory, Oracle 11g, Solaris 10, Cert# 2012014. IBM Flex System x240 (2 processors, 16 cores, 32 threads) 7,960 SAP SD Users, 2 x 2.90 GHz Intel Xeon E5-2690, 128 GB memory, DB2 9.7, Windows Server 2008 R2 EE, Cert# 2012016. IBM System x3650 M4 (2 processors, 16 cores, 32 threads) 7,855 SAP SD Users, 2 x 2.90 GHz Intel Xeon E5-2690, 128 GB memory, DB2 9.7, Windows Server 2008 R2 EE, Cert# 2012010. Cisco UCS C240 M3 (2 processors, 16 cores, 32 threads) 7,635 SAP SD Users, 2 x 2.90 GHz Intel Xeon E5-2690, 128 GB memory, SQL Server 2008, Windows Server 2008 R2 DE, Cert# 2012011. Fujitsu PRIMERGY RX300 S7 (2 processors, 16 cores, 32 threads) 7,570 SAP SD Users, 2 x 2.90 GHz Intel Xeon E5-2690, 128 GB memory, SQL Server 2008, Windows Server 2008 R2 EE, Cert# 2012008. HP ProLiant DL380p Gen8 (2 processors, 16 cores, 32 threads) 7,865 SAP SD Users, 2 x 2.90 GHz Intel Xeon E5-2690, 128 GB memory, SQL Server 2008, Windows Server 2008 R2 EE, Cert# 2012012.

SAP, R/3, reg TM of SAP AG in Germany and other countries. More info www.sap.com/benchmark

Tuesday Apr 10, 2012

SPEC CPU2006 Results on Oracle's Sun x86 Servers

Oracle's new Sun x86 servers delivered world records on the benchmarks SPECfp2006 and SPECint_rate2006 for two processor servers. This was accomplished with Oracle Solaris 11 and Oracle Solaris Studio 12.3 software.

  • The Sun Fire X4170 M3 (now known as Sun Server X3-2) server achieved a world record result in for SPECfp2006 benchmark with a score of 96.8.

  • The Sun Blade X6270 M3 server module (now known as Sun Blade X3-2B) produced best integer throughput performance for all 2-socket servers with a SPECint_rate2006 score of 705.

  • The Sun x86 servers with Intel Xeon E5-2690 2.9 GHz processors produced a cross-generational performance improvement up to 1.8x over the previous generation, Sun x86 M2 servers.

Performance Landscape

Complete benchmark results are at the SPEC website, SPEC CPU2006 Results. The tables below provide the new Oracle results, as well as, select results from other vendors.

SPECint2006
System Processor c/c/t * Peak Base O/S Compiler
Fujitsu PRIMERGY BX924 S3 Intel E5-2690, 2.9 GHz 2/16/16 60.8 56.0 RHEL 6.2 Intel 12.1.2.273
Sun Fire X4170 M3 Intel E5-2690, 2.9 GHz 2/16/32 58.5 54.3 Oracle Linux 6.1 Intel 12.1.0.225
Sun Fire X4270 M2 Intel X5690, 3.47 GHz 2/12/12 46.2 43.9 Oracle Linux 5.5 Intel 12.0.1.116

SPECfp2006
System Processor c/c/t * Peak Base O/S Compiler
Sun Fire X4170 M3 Intel E5-2690, 2.9 GHz 2/16/32 96.8 86.4 Oracle Solaris 11 Studio 12.3
Sun Blade X6270 M3 Intel E5-2690, 2.9 GHz 2/16/32 96.0 85.2 Oracle Solaris 11 Studio 12.3
Sun Fire X4270 M3 Intel E5-2690, 2.9 GHz 2/16/32 95.9 85.1 Oracle Solaris 11 Studio 12.3
Fujitsu CELSIUS R920 Intel E5-2687, 2.9 GHz 2/16/16 93.8 87.6 RHEL 6.1 Intel 12.1.2.273
Sun Fire X4270 M2 Intel X5690, 3.47 GHz 2/12/24 64.2 59.2 Oracle Solaris 10 Studio 12.2

Only 2-chip server systems listed below, excludes workstations.

SPECint_rate2006
System Processor Base
Copies
c/c/t * Peak Base O/S Compiler
Sun Blade X6270 M3 Intel E5-2690, 2.9 GHz 32 2/16/32 705 632 Oracle Solaris 11 Studio 12.3
Sun Fire X4270 M3 Intel E5-2690, 2.9 GHz 32 2/16/32 705 630 Oracle Solaris 11 Studio 12.3
Sun Fire X4170 M3 Intel E5-2690, 2.9 GHz 32 2/16/32 702 628 Oracle Solaris 11 Studio 12.3
Cisco UCS C220 M3 Intel E5-2690, 2.9 GHz 32 2/16/32 697 671 RHEL 6.2 Intel 12.1.0.225
Sun Blade X6270 M2 Intel X5690, 3.47 GHz 24 2/12/24 410 386 Oracle Linux 5.5 Intel 12.0.1.116

SPECfp_rate2006
System Processor Base
Copies
c/c/t * Peak Base O/S Compiler
Cisco UCS C240 M3 Intel E5-2690, 2.9 GHz 32 2/16/32 510 496 RHEL 6.2 Intel 12.1.2.273
Sun Fire X4270 M3 Intel E5-2690, 2.9 GHz 64 2/16/32 497 461 Oracle Solaris 11 Studio 12.3
Sun Blade X6270 M3 Intel E5-2690, 2.9 GHz 32 2/16/32 497 460 Oracle Solaris 11 Studio 12.3
Sun Fire X4170 M3 Intel E5-2690, 2.9 GHz 64 2/16/32 495 464 Oracle Solaris 11 Studio 12.3
Sun Fire X4270 M2 Intel E5690, 3.47 GHz 24 2/12/24 273 265 Oracle Linux 5.5 Intel 12.0.1.116

* c/c/t — chips / cores / threads enabled

Configuration Summary and Results

Hardware Configuration:

Sun Fire X4170 M3 server
2 x 2.90 GHz Intel Xeon E5-2690 processors
128 GB memory (16 x 8 GB 2Rx4 PC3-12800R-11, ECC)

Sun Fire X4270 M3 server
2 x 2.90 GHz Intel Xeon E5-2690 processors
128 GB memory (16 x 8 GB 2Rx4 PC3-12800R-11, ECC)

Sun Blade X6270 M3 server module
2 x 2.90 GHz Intel Xeon E5-2690 processors
128 GB memory (16 x 8 GB 2Rx4 PC3-12800R-11, ECC)

Software Configuration:

Oracle Solaris 11 11/11 (SRU2)
Oracle Solaris Studio 12.3 (patch update 1 nightly build 120313) Oracle Linux Server Release 6.1
Intel C++ Studio XE 12.1.0.225
SPEC CPU2006 V1.2

Benchmark Description

SPEC CPU2006 is SPEC's most popular benchmark. It measures:

  • Speed — single copy performance of chip, memory, compiler
  • Rate — multiple copy (throughput)

The benchmark is also divided into integer intensive applications and floating point intensive applications:

  • integer: 12 benchmarks derived from real applications such as perl, gcc, XML processing, and pathfinding
  • floating point: 17 benchmarks derived from real applications, including chemistry, physics, genetics, and weather.

It is also divided depending upon the amount of optimization allowed:

  • base: optimization is consistent per compiled language, all benchmarks must be compiled with the same flags per language.
  • peak: specific compiler optimization is allowed per application.

The overall metrics for the benchmark which are commonly used are:

  • SPECint_rate2006, SPECint_rate_base2006: integer, rate
  • SPECfp_rate2006, SPECfp_rate_base2006: floating point, rate
  • SPECint2006, SPECint_base2006: integer, speed
  • SPECfp2006, SPECfp_base2006: floating point, speed

See here for additional information.

See Also

Disclosure Statement

SPEC and the benchmark names SPECfp and SPECint are registered trademarks of the Standard Performance Evaluation Corporation. Results as of 10 April 2012 from www.spec.org and this report.

Thursday Mar 29, 2012

Sun Server X2-8 (formerly Sun Fire X4800 M2) Posts World Record x86 SPECjEnterprise2010 Result

Oracle's Sun Server X2-8 (formerly Sun Fire X4800 M2) using the Intel Xeon E7-8870 processor and Sun Server X2-4 using the Intel Xeon E7-4870 processor, produced a world record single application server SPECjEnterprise2010 benchmark result of 27,150.05 SPECjEnterprise2010 EjOPS. The Sun Server X2-8 ran the application tier and the Sun Server X2-4 was used for the database tier.

  • The Sun Server X2-8 demonstrated 63% better performance compared to IBM P780 server result of 16,646.34 SPECjEnterprise2010 EjOPS.

  • The Sun Server X2-8 demonstrated 4% better performance than the Cisco UCS B440 M2 result, both results used the same number of processors.

  • This result used Oracle WebLogic Server 12c, Java HotSpot(TM) 64-Bit Server 1.7.0_02, and Oracle Database 11g.

  • This result was produced using Oracle Linux.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results. The table below compares against the best results from IBM and Cisco.

SPECjEnterprise2010 Performance Chart
as of 7/11/2012
Submitter EjOPS* Application Server Database Server
Oracle 27,150.05 1x Sun Server X2-8
8x 2.4 GHz Intel Xeon E7-8870
Oracle WebLogic 12c
1x Sun Server X2-4
4x 2.4 GHz Intel Xeon E7-4870
Oracle Database 11g (11.2.0.2)
Cisco 26,118.67 2x UCS B440 M2 Blade Server
4x 2.4 GHz Intel Xeon E7-4870
Oracle WebLogic 11g (10.3.5)
1x UCS C460 M2 Blade Server
4x 2.4 GHz Intel Xeon E7-4870
Oracle Database 11g (11.2.0.2)
IBM 16,646.34 1x IBM Power 780
8x 3.86 GHz POWER 7
WebSphere Application Server V7
1x IBM Power 750 Express
4x 3.55 GHz POWER 7
IBM DB2 9.7 Workgroup Server Edition FP3a

* SPECjEnterprise2010 EjOPS, bigger is better.

Configuration Summary

Application Server:

1 x Sun Server X2-8

8 x 2.4 GHz Intel Xeon processor E7-8870
256 GB memory
4 x 10 GbE NIC
2 x FC HBA
Oracle Linux 5 Update 6
Oracle WebLogic Server Standard Edition Release 12.1.1
Java HotSpot(TM) 64-Bit Server VM on Linux, version 1.7.0_02 (Java SE 7 Update 2)

Database Server:

1 x Sun Server X2-4
4 x 2.4 GHz Intel Xeon E7-4870
512 GB memory
4 x 10 GbE NIC
2 x FC HBA
2 x Sun StorageTek 2540 M2
4 x Sun Fire X4270 M2
4 x Sun Storage F5100 Flash Array
Oracle Linux 5 Update 6
Oracle Database 11g Enterprise Edition Release 11.2.0.2

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The SPECjEnterprise2010 benchmark has been designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems.

The workload consists of an end to end web based order processing domain, an RMI and Web Services driven manufacturing domain and a supply chain model utilizing document based Web Services. The application is a collection of Java classes, Java Servlets, Java Server Pages, Enterprise Java Beans, Java Persistence Entities (pojo's) and Message Driven Beans.

The SPECjEnterprise2010 benchmark heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second ("SPECjEnterprise2010 EjOPS"). This metric is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is no price/performance metric in this benchmark.

Key Points and Best Practices

  • Sixteen Oracle WebLogic server instances were started using numactl, binding 2 instances per chip.
  • Eight Oracle database listener processes were started, binding 2 instances per chip using taskset.
  • Additional tuning information is in the report at http://spec.org.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Sun Server X2-8, 27,150.05 SPECjEnterprise2010 EjOPS; IBM Power 780, 16,646.34 SPECjEnterprise2010 EjOPS; Cisco UCS B440 M2, 26,118.67 SPECjEnterprise2010 EjOPS. Results from www.spec.org as of 7/11/2012.

Monday Feb 27, 2012

Sun ZFS Storage 7320 Appliance 33% Faster Than NetApp FAS3270 on SPECsfs2008

Oracle's Sun ZFS Storage 7320 appliance delivered outstanding performance on the SPECsfs2008 NFS benchmark, beating results published on NetApp's fastest midrange platform, the NetApp FAS3270, and the EMC Gateway NS-G8 Server Failover Cluster.

  • The Sun ZFS Storage 7320 appliance delivered 134,140 SPECsfs2008_nfs.v3 Ops/sec with an Overall Response Time (ORT) of 1.51 msec on the SPECsfs2008 NFS benchmark.

  • The Sun ZFS Storage 7320 appliance has 33% higher throughput than the NetApp FAS3270 on the SPECsfs2008 NFS benchmark.

  • The Sun ZFS Storage 7320 appliance required less than half the rack space of the NetApp FAS3270.

  • The Sun ZFS Storage 7320 appliance has 9% better Overall Response Time than the NetApp FAS3270 on the SPECsfs2008 NFS benchmark.

Performance Landscape

SPECsfs2008_nfs.v3 Performance Chart (in decreasing SPECsfs2008_nfs.v3 Ops/sec order)

Sponsor System Throughput
(Ops/sec)
Overall Response
Time (msec)
Memory
(GB)
Disks Exported
Capacity (TB)
Rack Units
Controllers+Disks
EMC VG8 135,521 1.92 280 312 19.2
Oracle 7320 134,140 1.51 288 136 37.0 26
EMC NS-G8 110,621 2.32 264 100 17.6
NetApp FAS3270 101,183 1.66 40 360 110.1 66

Throughput SPECsfs2008_nfs.v3 Ops/sec = the Performance Metric
Overall Response Time = the corresponding Response Time Metric

Complete SPECsfs2008 benchmark results may be found at http://www.spec.org/sfs2008/results/sfs2008.html.

Configuration Summary

Storage Configuration:

Sun ZFS Storage 7320 appliance in clustered configuration
2 x Sun ZFS Storage 7320 controllers, each with
2 x 2.4 GHz Intel Xeon E5620 processors
144 GB memory
4 x 512 GB SSD flash-enabled read-cache
6 x Sun Disk shelves
4 x shelves with 24 x 300 GB 15K RPM SAS-2 drives
2 x shelves with 20 x 300 GB 15K RPM SAS-2 drives and 4 x 73 GB SAS-2 flash-enabled write-cache

Server Configuration:

3 x Sun Fire X4270 M2 servers, each with
2 x 2.4 GHz Intel Xeon E5620 processors
12 GB memory
1 x 10 GbE connection to the Sun ZFS Storage 7320 appliance
Oracle Solaris 10 8/11

Benchmark Description

SPECsfs2008 is the latest version of the Standard Performance Evaluation Corporation (SPEC) benchmark suite measuring file server throughput and response time, providing a standardized method for comparing performance across different vendor platforms. SPECsfs2008 results summarize the server's capabilities with respect to the number of operations that can be handled per second, as well as the overall latency of the operations. The suite is a follow-on to the SFS97_R1 benchmark, adding a CIFS workload, an updated NFSv3 workload, support for additional client platforms, and a new test harness and reporting/submission framework.

See Also

Disclosure Statement

SPEC and SPECsfs are registered trademarks of Standard Performance Evaluation Corporation (SPEC). Results as of February 22, 2012, for more information see www.spec.org. Sun ZFS Storage 7320 Appliance 134,140 SPECsfs2008_nfs.v3 Ops/sec, 1.51 msec ORT, NetApp FAS3270 101,183 SPECsfs2008_nfs.v3 Ops/Sec, 1.66 msec ORT, EMC Celerra Gateway NS-G8 Server Failover Cluster, 3 Datamovers (1 stdby) / Symmetrix V-Max 110,621 SPECsfs2008_nfs.v3 Ops/Sec, 2.32 msec ORT.

Monday Oct 03, 2011

SPARC T4-4 Servers Set World Record on SPECjEnterprise2010, Beats IBM POWER7, Cisco x86

Oracle produced a world record SPECjEnterprise2010 benchmark result of 40,104.86 SPECjEnterprise2010 EjOPS using four of Oracle's SPARC T4-4 servers in the application tier and two more SPARC T4-4 servers for the database server.

  • The four SPARC T4-4 server configuration (sixteen SPARC T4 processors total, 3.0 GHz) demonstrated 2.4x better performance compared to the IBM Power 780 server (eight POWER7 processors, 3.86 THz) result of 16,646.34 SPECjEnterprise2010 EjOPS.

  • In the database tier, two SPARC T4-4 servers with a total of eight SPARC T4 processors at 3.0 GHz, processed 2.4x more transactions compared to the IBM result of 16,646.34 SPECjEnterprise2010 EjOPS which used four POWER7 processors at 3.55 GHz.

  • The four SPARC T4-4 server configuration demonstrated 1.5x better performance compared to the Cisco UCS B440 M2 Blade Server result of 26,118.67 SPECjEnterprise2010 EjOPS.

  • The four SPARC T4-4 server configuration demonstrated 2.3x better performance compared to the Cisco UCS B440 M1 Blade Server result of 17,301.86 SPECjEnterprise2010 EjOPS.

  • This result demonstrated less than 1 second average response times for all SPECjEnterprise2010 transactions and 90% of all transaction times took less than 1 second.

  • This result demonstrated a sustained Java EE 5 transaction load generated by approximately 320,000 users.

  • This result using 16 Oracle WebLogic 10.3.5 server instances demonstrated 4.8x better performance per application server instance when compared to the IBM result which used 32 WebSphere instances.

  • The SPARC T4-4 servers delivered a 6.7x price/performance advantage over the IBM Power 780 for the servers used in the application tier (see disclosure statement below for details). This price/performance advantage in the application tier was accomplished with a SPARC T4-4 server configuration with 2 TB of total memory compared to the IBM solution with 0.5 TB of memory.

  • The SPARC T4-4 servers had a 1.9x advantage over IBM in performance per space for the application tier (see disclosure statement below for details) even though the Oracle solution had four servers.

  • The four SPARC T4-4 servers used for the application tier used Oracle Solaris Containers to consolidate four Oracle WebLogic application server instances on each server to achieve this result.

  • The two SPARC T4-4 servers used for the database tier hosted Oracle Database 11g Release 2 and Oracle RAC cluster software using Oracle Automatic Storage Management (ASM).

  • Oracle Fusion Middleware provides a family of complete, integrated, hot pluggable and best-of-breed products known for enabling enterprise customers to create and run agile and intelligent business applications. Oracle WebLogic Server's on-going, record-setting Java application server performance demonstrates why so many customers rely on Oracle Fusion Middleware as their foundation for innovation.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results.

SPECjEnterprise2010 Performance Chart
as of 10/11/2011
Submitter EjOPS* Java EE Server DB Server
Oracle 40,104.86 4 x SPARC T4-4
4 chips, 32 cores, 3.0 GHz SPARC T4
Oracle WebLogic 11g (10.3.5)
2 x SPARC T4-4
4 chips, 32 cores, 3.0 GHz SPARC T4
Oracle 11g DB 11.2.0.2
Cisco 26,118.67 2 x Cisco UCS B440 M2
4 chips, 40 cores, 2.4 GHz Xeon E7-4870
Oracle WebLogic 11g (10.3.5)
1 x Cisco UCS C460 M2
4 chips, 40 cores, 2.4 GHz Xeon E7-4870
Oracle 11g DB 11.2.0.2
Cisco 17,301.86 2 x Cisco UCS B440 M1
4 chips, 32 cores, 2.26 GHz Xeon X7560
Oracle WebLogic 10.3.4
1 x Cisco UCS C460 M1
4 chips, 32 cores, 2.26 GHz Xeon X7560
Oracle 11g DB 11.2.0.2
IBM 16,646.34 1 x IBM Power 780
8 chips, 64 cores, 3.86 GHz POWER7
WebSphere Application Server V7.0
1 x IBM Power 750 Express
4 chips, 32 cores, 3.55 GHz POWER7
IBM DB2 Universal Database 9.7

* SPECjEnterprise2010 EjOPS (bigger is better)

Configuration Summary

Application Servers:

4 x SPARC T4-4 servers, each with
4 x 3.0 GHz SPARC T4 processors
512 GB memory
2 x 10GbE NIC
Oracle Solaris 10 8/11
Oracle WebLogic Server 11g Release 1 (10.3.5)
Java HotSpot(TM) 64-Bit Server VM on Solaris, version 1.6.0_26 (Java SE 6 Update 26)

Database Servers:

2 x SPARC T4-4 servers, each with
4 x 3.0 GHz SPARC T4 processors
1024 GB memory
2 x 10GbE NIC
4 x 8Gb FC HBA
Oracle Solaris 10 8/11
Oracle Database 11g Enterprise Edition Release 11.2.0.2
Oracle Real Application Clusters 11g Release 2

Storage Servers:

8 x Sun Fire X4270 M2 (12-Drive)
1 x 3.0 GHz Intel Xeon
8 GB memory
1 x 8Gb FC HBA
Oracle Solaris 11 Express 2010.11
8 x Sun Storage F5100 Flash Arrays

Switch Hardware:

2 x Sun Network 10GbE 72-port Top of Rack (ToR) Switch
1 x Brocade 5300 80-port Fiber Channel Switch

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The new SPECjEnterprise2010 benchmark has been re-designed and developed to cover the Java EE 5 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems,
  • The web container, servlets, and web services
  • The EJB container
  • JPA 1.0 Persistence Model
  • JMS and Message Driven Beans
  • Transaction management
  • Database connectivity
Moreover, SPECjEnterprise2010 also heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second (SPECjEnterprise2010 EjOPS). The primary metric for the SPECjEnterprise2010 benchmark is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is NO price/performance metric in this benchmark.

Key Points and Best Practices

  • Four Oracle WebLogic server instances on each SPARC T4-4 server were hosted in 4 separate Oracle Solaris Containers to demonstrate consolidation of multiple application servers.
  • Each Oracle Solaris Container was bound to a separate processor set, each contained 7 cores (total 56 threads). This was done to improve performance by reducing memory access latency by using the physical memory closest to the processors. The default set was used for network and disk interrupt handling.
  • The Oracle WebLogic application servers were executed in the FX scheduling class to improve performance by reducing the frequency of context switches.
  • The Oracle database processes were run in 2 processor sets using psrset(1M) and executed in the FX scheduling class. This improved performance by reducing memory access latency and reducing context switches.
  • The Oracle log writer process was run in a separate processor set containing 2 threads and run in the RT scheduling class. This insured that the log writer had the most efficient use of CPU resources.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Results from www.spec.org as of 10/11/2011. SPARC T4-4, 40,104.86 SPECjEnterprise2010 EjOPS; Cisco UCS B440 M2, 26,118.67 SPECjEnterprise2010 EjOPS; Cisco UCS B440 M1, 17,301.86 SPECjEnterprise2010 EjOPS; IBM Power 780, 16,646.34 SPECjEnterprise2010 EjOPS.

SPECjEnterprise2010 models contemporary Java-based applications that run on large Java EE (Java Enterprise Edition) servers, backed by network infrastructure and database servers. Focusing on the critical Java EE server hardware & OS, the IBM result includes a Java EE server with a list price of $1.30 million. The Oracle Java EE servers have a list price of $0.47 million. The Java EE server price versus delivered EjOPS is $77.97/EjOPS for IBM versus $11.67/EjOPS for Oracle. Oracle's $/perf advantage is 6.7x better than IBM ($77.97/$11.67).

Pricing details for IBM, IBM p780 512GB based on public pricing at http://tpc.org/results/FDR/TPCH/TPC-H_1TB_IBM780_Sybase-FDR.pdf. Adjusted hardware costs to license all 64 cores. AIX pricing at: http://www-304.ibm.com/easyaccess3/fileserve?contentid=214347 and AIX Standard Edition V7.1 per processor (5765-G98-0017 64*2,600=$166,400). This gives application tier hardware & OS Price/perf: $77.97/EjOPS (1297956/16646.34)

Pricing details for Oracle, four SPARC T4-4 512 GB, HW acquisition price from Oracle's price list: $467,856 http://www.oracle.com. This gives application tier hardware & OS Price/perf: $11.67/EjOPS (467856/40104.86)

The Oracle application tier servers occupy 20U of space, 40,140.86/20=2005 EjOPS/U. The IBM application tier server occupies 16U of space, 16,646.34/16=1040 EjOPS/U. 2005/1040=1.9x

Thursday Sep 29, 2011

SPARC T4-2 Server Beats Intel (Westmere AES-NI) on SSL Network Tests

Oracle's SPARC T4 processor is faster and more efficient than the Intel Xeon X5690 processor (with AES-NI) when running network SSL thoughput tests.

  • The SPARC T4 processor at 2.85 GHz is 20% faster than the 3.46 GHz Intel Xeon X5690 processor on single stream network SSL encryption.

  • The SPARC T4 processor requires fewer streams to attain near-linespeed of a 10 GbE secure network and does this with 5 times less CPU resources compared to the Intel Xeon X5690 processor.

  • Oracle's SPARC T4-2 server using 8 threads achieves line speed over a 10 GbE network with only 9% CPU utilization.

  • Oracle's Sun Fire X4270 M2 with two Intel Xeon X5690 processors achieves line speed with 8 threads, but at 45% CPU utilization.

The SPARC T4 processor has hardware support via Encryption Instruction Accelerators for encryption and decryption for AES and many other ciphers. The Intel Xeon X5690 processor has AES-NI instructions which accelerate only AES ciphers.

Performance Landscape

The following table shows single stream results running encrypted (SSL Read) and unencrypted (Clear Text) messages of 1 MB in size. These tests were run with the uperf benchmark and used the AES-256-CBC cipher. They were run across a 10 GbE connection. Write messages saw similar performance.

Single Stream Network Communication with Uperf
Processor Performance (Mb/sec)
Clear Text SSL Read
SPARC T4, 2.85 GHz 4,194 1,678
Intel Xeon X5690, 3.46 GHz 5,591 1,398

The next table shows how many streams it takes to achieve 90% of the 10 GbE network bandwidth (9000 Mb/sec) for encrypted read messages of 1 MB in size. These tests were run with the uperf benchmark and used the AES-256-CBC cipher. Write messages saw similar performance.

Uperf SSL Read with AES-256-CBC
Processor Number of
Streams for 90%
Network Utilization
CPU Utilization
SPARC T4, 2.85 GHz 8 9%
Intel Xeon X5690, 3.46 GHz 12 45%

Configuration Summary

SPARC T4 Configuration:

2 x SPARC T4-2 servers each with
2 x SPARC T4 processors, 2.85 GHz
128 GB memory
1 x 10-Gigabit Ethernet XAUI Adapter
Oracle Solaris 11
Back-to-back 10 GbE connection

Intel Configuration:

2 x Sun Fire X4270 M2 servers each with
2 x Intel Xeon X5690 processors, 3.46 GHz
48 GB memory
1 x Sun Dual Port 10GbE PCIe 2.0 Networking Card with Intel 82599 10GbE Controller
Oracle Solaris 11
Back-to-back 10 GbE connection

Software Configuration:

OpenSSL 1.0.0.d
uperf 1.0.3
gcc 3.4.3

Benchmark Description

Uperf is an open source benchmark program for simulating and measuring network performance. Uperf is able to measure the performance of various protocols, including TCP, UDP, SCTP and SSL. The uperf benchmark uses an input-defined workload to test network performance. This input workload can be used to model complex situations or to isolate simple tasks. The workload used for these tests was simple network reads and simple network writes.

Key Points and Best Practices

  • The Encryption Instruction Accelerators are accessed through a platform independent API for cryptographic engines.
  • The OpenSSL libraries use the API. The default is to not use the Encryption Instruction Accelerators.
  • Cryptography is compute intensive. Using 8 streams, the SPARC T4 processor was able to match the bandwidth of the 10 GbE network with 8 threads.

See Also

Disclosure Statement

Copyright 2011, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 9/26/2011.

Wednesday Sep 28, 2011

SPARC T4 Servers Set World Record on Oracle E-Business Suite R12 X-Large Order to Cash

With Oracle's SPARC T4-2 server running the application and SPARC T4-4 server running the database, Oracle set a world record result for the Oracle E-Business Suite Standard X-Large Order to Cash (OLTP) benchmark.

  • The combination of a SPARC T4-2 server running the Oracle E-Business Suite R12.1.2 application and a SPARC T4-4 server running the Oracle Database 11g Release 2 database enabled 2400 Order to Cash users of the X-Large Benchmark to simultaneously execute a large volume of medium to heavy transactions with an average response time of 2.4 seconds.

  • The SPARC T4-2 server in the application tier and the SPARC T4-4 server in the database tier are only about half utilized providing significant headroom for additional Oracle E-Business Suite R12.1.2 processing modules and future growth.

Performance Landscape

This is the first published result for the X-large benchmark using Oracle E-Business Order Management module.

OLTP Workload: Order to Cash
X-Large Configuration
System Users Average
Response Time
90th Percentile
Response Time
SPARC T4-2 2400 2.413 sec. 3.114 sec.

Configuration Summary

Application Tier Configuration:

1 x SPARC T4-2 server
2 x SPARC T4 processors, 2.85 GHz
256 GB memory
Oracle Solaris 10 8/11
Oracle E-Business Suite 12.1.2

Database Tier Configuration:

1 x SPARC T4-4 server
4 x SPARC T4 processors, 3.0 GHz
256 GB memory
Oracle Solaris 10 8/11
Oracle Database 11g Release 2

Storage Configuration:

1 x Sun Storage F5100 Flash Array

Benchmark Description

The Oracle R12 E-Business Suite Standard Benchmark combines online transaction execution by simulated users with concurrent batch processing to model a typical scenario for a global enterprise. This benchmark ran one OLTP component, Order to Cash, in the Extra-Large size. The goal is to obtain reference response times.

Results can be published in four sizes and utilize different combination

  • X-large: Maximum online users running all business flows between 10,000 to 20,000; 750,000 order to cash lines per hour and 250,000 payroll checks per hour.
    • Order to Cash Online -- 2400 users
      • The percentage across the 5 transactions in Order Management module is:
        • Insert Manual Invoice -- 16.66%
        • Insert Order -- 32.33%
        • Order Pick Release -- 16.66%
        • Ship Confirm -- 16.66%
        • Order Summary Report -- 16.66%
    • HR Self-Service -- 4000 users
    • Customer Support Flow -- 8000 users
    • Procure to Pay -- 2000 users
  • Large: 10,000 online users; 100,000 order to cash lines per hour and 100,000 payroll checks per hour.
  • Medium: up to 3000 online users; 50,000 order to cash lines per hour and 10,000 payroll checks per hour.
  • Small: up to 1000 online users; 10,000 order to cash lines per hour and 5,000 payroll checks per hour.

See Also

Disclosure Statement

Oracle E-Business X-Large Order to Cash benchmark, SPARC T4-2, SPARC T4, 2.85 GHz, 2 chips, 16 cores, 128 threads, 256 GB memory, SPARC T4-4, SPARC T4, 3.0 GHz, 4 chips, 32 cores, 256 threads, 256 GB memory, average response time 2.413 sec, 90th percentile response time 3.114 sec, Oracle Solaris 10 8/11, Oracle E-Business Suite 12.1.2, Oracle Database 11g Release 2, Results as of 9/26/2011.

SPARC T4-2 Server Beats Intel (Westmere AES-NI) on Oracle Database Tablespace Encryption Queries

Oracle's SPARC T4 processor with Encryption Instruction Accelerators greatly improves performance over software implementations. This will greatly expand the use of TDE for many customers.

  • Oracle's SPARC T4-2 server is over 42% faster than Oracle's Sun Fire X4270 M2 (Intel AES-NI) when running DSS-style queries referencing an encrypted tablespace.

Oracle's Transparent Data Encryption (TDE) feature of the Oracle Database simplifies the encryption of data within datafiles preventing unauthorized access to it from the operating system. Tablespace encryption allows encryption of the entire contents of a tablespace.

TDE tablespace encryption has been certified with Siebel, PeopleSoft, and Oracle E-Business Suite applications

Performance Landscape

Total Query Time (time in seconds)
System GHz AES-128 AES-192 AES-256
SPARC T4-2 server 2.85 588 588 588
Sun Fire X4270 M2 (Intel X5690) 3.46 836 841 842
SPARC T4-2 Advantage
42% 43% 43%

Configuration Summary

SPARC Configuration:

SPARC T4-2 server
2 x SPARC T4 processors, 2.85 GHz
256 GB memory
2 x Sun Storage F5100 Flash Array
Oracle Solaris 11
Oracle Database 11g Release 2

Intel Configuration:

Sun Fire X4270 M2 server
2 x Intel Xeon X5690 processors, 3.46 GHz
48 GB memory
2 x Sun Storage F5100 Flash Array
Oracle Linux 5.7
Oracle Database 11g Release 2

Benchmark Description

To test the performance of TDE, a 1 TB database was created. To demonstrate secure transactions, four 25 GB tables emulating customer private data were created: clear text, encrypted AES-128, encrypted AES-192, and encrypted AES-256. Eight queries of varying complexity that join on the customer table were executed.

The time spent scanning the customer table during each query was measured and query plans analyzed to ensure a fair comparison, e.g. no broken queries. The total query time for all queries is reported.

Key Points and Best Practices

  • Oracle Database 11g Release 2 is required for SPARC T4 processor Encryption Instruction Accelerators support with TDE tablespaces.

  • TDE tablespaces support the SPARC T4 processor Encryption Instruction Accelerators for Advanced Encryption Standard (AES) only.

  • AES-CFB is the mode used in the Oracle database with TDE

  • Prior to using TDE tablespaces you must create a wallet and setup an encryption key. Here is one method to do that:

  • Create a wallet entry in $ORACLE_HOME/network/admin/sqlnet.ora.
    ENCRYPTION_WALLET_LOCATION=
    (SOURCE=(METHOD=FILE)(METHOD_DATA=
    (DIRECTORY=/oracle/app/oracle/product/11.2.0/dbhome_1/encryption_wallet)))
    
    Set an encryption key. This also opens the wallet.
    $ sqlplus / as sysdba
    SQL> ALTER SYSTEM SET ENCRYPTION KEY IDENTIFIED BY "tDeDem0";
    
    On subsequent instance startup open the wallet.
    $ sqlplus / as sysdba
    SQL> STARTUP;
    SQL> ALTER SYSTEM SET ENCRYPTION WALLET OPEN IDENTIFIED BY "tDeDem0";
    
  • TDE tablespace encryption and decryption occur on physical writes and reads of database blocks, respectively.

  • For parallel query using direct path reads decryption overhead varies inversely with the complexity of the query.

    For a simple full table scan query overhead can be reduced and performance improved by reducing the degree of parallelism (DOP) of the query.

See Also

Disclosure Statement

Copyright 2011, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 9/26/2011.

Tuesday Sep 27, 2011

SPARC T4-2 Servers Set World Record on JD Edwards EnterpriseOne Day in the Life Benchmark with Batch, Outperforms IBM POWER7

Using Oracle's SPARC T4-2 server for the application tier and a SPARC T4-1 server for the database tier, a world record result was produced running the Oracle's JD Edwards EnterpriseOne application Day in the Life (DIL) benchmark concurrently with a batch workload.

  • The SPARC T4-2 server running online and batch with JD Edwards EnterpriseOne 9.0.2 is 1.7x faster and has better response time than the IBM Power 750 system which only ran the online component of JD Edwards EnterpriseOne 9.0 Day in the Life test.

  • The combination of SPARC T4 servers delivered a Day in the Life benchmark result of 10,000 online users with 0.35 seconds of average transaction response time running concurrently with 112 Universal Batch Engine (UBE) processes at 67 UBEs/minute.

  • This is the first JD Edwards EnterpriseOne benchmark for 10,000 users and payroll batch on a SPARC T4-2 server for the application tier and the database tier with Oracle Database 11g Release 2. All servers ran with the Oracle Solaris 10 operating system.

  • The single-thread performance of the SPARC T4 processor produced sub-second response for the online components and provided dramatic performance for the batch jobs.

  • The SPARC T4 servers, JD Edwards EnterpriseOne 9.0.2, and Oracle WebLogic Server 11g Release 1 support 17% more users per JAS (Java Application Server) than the SPARC T3-1 server for this benchmark.

  • The SPARC T4-2 server provided a 6.7x better batch processing rate than the previous SPARC T3-1 server record result and had 2.5x faster response time.

  • The SPARC T4-2 server used Oracle Solaris Containers, which provide flexible, scalable and manageable virtualization.

  • JD Edwards EnterpriseOne uses Oracle Fusion Middleware WebLogic Server 11g R1 and Oracle Fusion Middleware Cluster Web Tier Utilities 11g HTTP server.

  • The combination of the SPARC T4-2 server and Oracle JD Edwards EnterpriseOne in the application tier with a SPARC T4-1 server in the database tier measured low CPU utilization providing headroom for growth.

Performance Landscape

JD Edwards EnterpriseOne Day in the Life Benchmark
Online with Batch Workload

System Online
Users
Resp
Time (sec)
Batch
Concur
(# of UBEs)
Batch
Rate
(UBEs/m)
Version
2xSPARC T4-2 (app+web)
SPARC T4-1 (db)
10000 0.35 112 67 9.0.2
SPARC T3-1 (app+web)
SPARC Enterprise M3000 (db)
5000 0.88 19 10 9.0.1

Resp Time (sec) — Response time of online jobs reported in seconds
Batch Concur (# of UBEs) — Batch concurrency presented in the number of UBEs
Batch Rate (UBEs/m) — Batch transaction rate in UBEs per minute

Edwards EnterpriseOne Day in the Life Benchmark
Online Workload Only

System Online
Users
Response
Time (sec)
Version
SPARC T3-1, 1 x SPARC T3 (1.65 GHz), Solaris 10 (app)
M3000, 1 x SPARC64 VII (2.75 GHz), Solaris 10 (db)
5000 0.52 9.0.1
IBM Power 750, POWER7 (3.55 GHz) (app+db) 4000 0.61 9.0

IBM result from http://www-03.ibm.com/systems/i/advantages/oracle/, IBM used WebSphere

Configuration Summary

Application Tier Configuration:

1 x SPARC T4-2 server with
2 x 2.85 GHz SPARC T4 processors
128 GB main memory
6 x 300 GB 10K RPM SAS internal HDD
Oracle Solaris 10 9/10
JD Edwards EnterpriseOne 9.0.2 with Tools 8.98.3.3

Web Tier Configuration:

1 x SPARC T4-2 server with
2 x 2.85 GHz SPARC T4 processors
256 GB main memory
2 x 300 GB SSD
4 x 300 GB 10K RPM SAS internal HDD
Oracle Solaris 10 9/10
Oracle WebLogic Server 11g Release 1

Database Tier Configuration:

1 x SPARC T4-1 server with
1 x 2.85 GHz SPARC T4 processor
128 GB main memory
6 x 300 GB 10K RPM SAS internal HDD
2 x Sun Storage F5100 Flash Array
Oracle Solaris 10 9/10
Oracle Database 11g Release 2

Benchmark Description

JD Edwards EnterpriseOne is an integrated applications suite of Enterprise Resource Planning (ERP) software. Oracle offers 70 JD Edwards EnterpriseOne application modules to support a diverse set of business operations.

Oracle's Day in the Life (DIL) kit is a suite of scripts that exercises most common transactions of JD Edwards EnterpriseOne applications, including business processes such as payroll, sales order, purchase order, work order, and manufacturing processes, such as ship confirmation. These are labeled by industry acronyms such as SCM, CRM, HCM, SRM and FMS. The kit's scripts execute transactions typical of a mid-sized manufacturing company.

  • The workload consists of online transactions and the UBE – Universal Business Engine workload of 42 short, 8 medium and 4 long UBEs.

  • LoadRunner runs the DIL workload, collects the user’s transactions response times and reports the key metric of Combined Weighted Average Transaction Response time.

  • The UBE processes workload runs from the JD Enterprise Application server.

    • Oracle's UBE processes come as three flavors:
      • Short UBEs < 1 minute engage in Business Report and Summary Analysis,
      • Mid UBEs > 1 minute create a large report of Account, Balance, and Full Address,
      • Long UBEs > 2 minutes simulate Payroll, Sales Order, night only jobs.
    • The UBE workload generates large numbers of PDF files reports and log files.
    • The UBE Queues are categorized as the QBATCHD, a single threaded queue for large and medium UBEs, and the QPROCESS queue for short UBEs run concurrently.

Oracle’s UBE process performance metric is Number of Maximum Concurrent UBE processes at transaction rate, UBEs/minute.

Key Points and Best Practices

One JD Edwards EnterpriseOne Application Server and two Oracle WebLogic Servers 11g R1 coupled with two Oracle Fusion Middleware 11g Web Tier HTTP Server instances on the SPARC T4-2 servers were hosted in three separate Oracle Solaris Containers to demonstrate consolidation of multiple application and web servers.

  • Interrupt fencing was configured on all Oracle Solaris Containers to channel the interrupts to processors other than the processor sets used for the JD Edwards Application server and WebLogic servers.

  • Processor 0 was left alone for clock interrupts.

  • The applications were executed in the FX scheduling class to improve performance by reducing the frequency of context switches.

  • A WebLogic vertical cluster was configured on each WebServer Container with twelve managed instances each to load balance users' requests and to provide the infrastructure that enables scaling to high number of users with ease of deployment and high availability.

  • The database server was run in an Oracle Solaris Container hosted on the SPARC T4-2 server.

  • The database log writer was run in the real time RT class and bound to a processor set.

  • The database redo logs were configured on the raw disk partitions.

  • The private network between the SPARC T4-2 servers was configured with a 10 GbE interface.

  • The Oracle Solaris Container on the Enterprise Application server ran 42 Short UBEs, 8 Medium UBEs and 4 Long UBEs concurrently as the mixed size batch workload.

  • The mixed size UBEs ran concurrently from the application server with the 10000 online users driven by the LoadRunner.

See Also

Disclosure Statement

Copyright 2011, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 9/26/2011.

SPARC T4-4 Server Sets World Record on PeopleSoft Payroll (N.A.) 9.1, Outperforms IBM Mainframe, HP Itanium

Oracle's SPARC T4-4 server achieved world record performance on the Unicode version of Oracle's PeopleSoft Enterprise Payroll (N.A) 9.1 extra-large volume model benchmark using Oracle Database 11g Release 2 running on Oracle Solaris 10.

  • The SPARC T4-4 server was able to process 1,460,544 payments/hour using PeopleSoft Payroll N.A 9.1.

  • The SPARC T4-4 server UNICODE result of 30.84 minutes on Payroll 9.1 is 2.8x faster than IBM z10 EC 2097 Payroll 9.0 (UNICODE version) result of 87.4 minutes. The IBM mainframe is rated at 6,512 MIPS.

  • The SPARC T4-4 server UNICODE result of 30.84 minutes on Payroll 9.1 is 3.1x faster than HP rx7640 Itanium2 non-UNICODE result of 96.17 minutes, on Payroll 9.0.

  • The average CPU utilization on the SPARC T4-4 server was only 30%, leaving significant room for business growth.

  • The SPARC T4-4 server processed payroll for 500,000 employees, 750,000 payments, in 30.84 minutes compared to the earlier world record result of 46.76 minutes on Oracle's SPARC Enterprise M5000 server.

  • The SPARC Enterprise M5000 server configured with eight 2.66 GHz SPARC64 VII processors has a result of 46.76 minutes on Payroll 9.1. That is 7% better than the result of 50.11 minutes on the SPARC Enterprise M5000 server configured with eight 2.53 GHz SPARC64 VII processors on Payroll 9.0. The difference in clock speed between the two processors is ~5%. That is close to the difference in the two results, thereby showing that the impact of the Payroll 9.1 benchmark on the overall result is about the same as that of Payroll 9.0.

Performance Landscape

PeopleSoft Payroll (N.A.) 9.1 – 500K Employees (7 Million SQL PayCalc, Unicode)

System OS/Database Payroll Processing
Result (minutes)
Run 1
(minutes)
Num of
Streams
SPARC T4-4, 4 x 3.0 GHz SPARC T4 Solaris/Oracle 11g 30.84 43.76 96
SPARC M5000, 8 x 2.66 GHz SPARC64 VII+ Solaris/Oracle 11g 46.76 66.28 32

PeopleSoft Payroll (N.A.) 9.0 – 500K Employees (3 Million SQL PayCalc, Non-Unicode)

System OS/Database Time in Minutes Num of
Streams
Payroll
Processing
Result
Run 1 Run 2 Run 3
Sun M5000, 8 x 2.53 GHz SPARC64 VII Solaris/Oracle 11g 50.11 73.88 534.20 1267.06 32
IBM z10 EC 2097, 9 x 4.4 GHz Gen1 Z/OS /DB2 58.96 80.5 250.68 462.6 8
IBM z10 EC 2097, 9 x 4.4 GHz Gen1 Z/OS /DB2 87.4 ** 107.6 - - 8
HP rx7640, 8 x 1.6 GHz Itanium2 HP-UX/Oracle 11g 96.17 133.63 712.72 1665.01 32

** This result was run with Unicode. The IBM z10 EC 2097 UNICODE result of 87.4 minutes is 48% slower than IBM z10 EC 2097 non-UNICODE result of 58.96 minutes, both on Payroll 9.0, each configured with nine 4.4GHz Gen1 processors.

Payroll 9.1 Compared to Payroll 9.0

Please note that Payroll 9.1 is Unicode based and Payroll 9.0 had non-Unicode and Unicode versions of the workload. There are 7 million executions of an SQL statement for the PayCalc batch process in Payroll 9.1 and 3 million executions of the same SQL statement for the PayCalc batch process in Payroll 9.0. This gets reflected in the elapsed time (27.33 min for 9.1 and 23.78 min for 9.0). The elapsed times of all other batch processes is lower (better) on 9.1.

Configuration Summary

Hardware Configuration:

SPARC T4-4 server
4 x 3.0 GHz SPARC T4 processors
256 GB memory
Sun Storage F5100 Flash Array
80 x 24 GB FMODs

Software Configuration:

Oracle Solaris 10 8/11
PeopleSoft HRMS and Campus Solutions 9.10.303
PeopleSoft Enterprise (PeopleTools) 8.51.035
Oracle Database 11g Release 2 11.2.0.1 (64-bit)
Micro Focus COBOLServer Express 5.1 (64-bit)

Benchmark Description

The PeopleSoft 9.1 Payroll (North America) benchmark is a performance benchmark established by PeopleSoft to demonstrate system performance for a range of processing volumes in a specific configuration. This information may be used to determine the software, hardware, and network configurations necessary to support processing volumes. This workload represents large batch runs typical of OLTP workloads during a mass update.

To measure five application business process run times for a database representing a large organization. The five processes are:

  • Paysheet Creation: Generates payroll data worksheets consisting of standard payroll information for each employee for a given pay cycle.

  • Payroll Calculation: Looks at paysheets and calculates checks for those employees.

  • Payroll Confirmation: Takes information generated by Payroll Calculation and updates the employees' balances with the calculated amounts.

  • Print Advice forms: The process takes the information generated by Payroll Calculations and Confirmation and produces an Advice for each employee to report Earnings, Taxes, Deduction, etc.

  • Create Direct Deposit File: The process takes information generated by the above processes and produces an electronic transmittal file that is used to transfer payroll funds directly into an employee's bank account.

Key Points and Best Practices

  • The SPARC T4-4 server with the Sun Storage F5100 Flash Array device had an average read throughput of up to 103 MB/sec and an average write throughput of up to 124 MB/sec while consuming 30% CPU on average.

  • The Sun Storage F5100 Flash Array device is a solid-state device that provides a read latency of only 0.5 msec. That is about 10 times faster than the normal disk latencies of 5 msec measured on this benchmark.

See Also

  • Oracle PeopleSoft Benchmark White Papers
    oracle.com
  • PeopleSoft Enterprise Human Capital Management (Payroll)
    oracle.com

  • PeopleSoft Enterprise Payroll 9.1 Using Oracle for Solaris (Unicode) on an Oracle's SPARC T4-4 – White Paper
    oracle.com

  • SPARC T4-4 Server
    oracle.com
  • Oracle Solaris
    oracle.com
  • Oracle Database 11g Release 2 Enterprise Edition
    oracle.com
  • Sun Storage F5100 Flash Array
    oracle.com

Disclosure Statement

Oracle's PeopleSoft Payroll 9.1 benchmark, SPARC T4-4 30.84 min,
http://www.oracle.com/us/solutions/benchmark/apps-benchmark/peoplesoft-167486.html, results 9/26/2011.

Thursday Sep 15, 2011

Sun Fire X4800 M2 Servers (now known as Sun Server X2-8) Produce World Record on SAP SD-Parallel Benchmark

Oracle delivered an SAP enhancement package 4 for SAP ERP 6.0 (Unicode) Sales and Distribution - Parallel (SD Parallel) Benchmark world record result using eight of Oracle's Sun Fire X4800 M2 servers (now known as Sun Server X2-8), Oracle Solaris 10 and Oracle Database 11g Real Application Clusters (RAC) software that achieved 180,000 users as of 10/03/2011.

  • The eight Sun Fire X4800 M2 servers delivered a world record result of 180,000 users on the SAP SD Parallel Benchmark.

  • The eight Sun Fire X4800 M2 server SD Parallel result of 180,000 users delivered 43% more performance compared to the IBM Power 795 server SD two-tier result of 126,063 users.

Performance Landscape

Selected SAP Sales and Distribution (SD) benchmark results are presented in decreasing order of performance. All benchmarks were using SAP enhancement package 4 for SAP ERP 6.0 (Unicode).

System OS
Database
Users SAPS Type Cert #
Eight Sun Fire X4800 M2
8 x Intel Xeon E7-8870 @2.4 GHz
512 GB
Oracle Solaris 10
Oracle 11g RAC
180,000 1,016,380 Parallel 2011037
Six Sun Fire X4800 M2
8 x Intel Xeon E7-8870 @2.4 GHz
512 GB
Oracle Solaris 10
Oracle 11g RAC
137,904 765,470 Parallel 2011038
IBM Power 795
32 x POWER7 @4.0 GHz
4096 GB
AIX 7.1
DB2 9.7
126,063 688,630 Two-Tier 2010046
Four Sun Fire X4800 M2
8 x Intel Xeon E7-8870 @2.4 GHz
512 GB
Oracle Solaris 10
Oracle 11g RAC
94,736 546,050 Parallel 2011039
Two Sun Fire X4800 M2
8 x Intel Xeon E7-8870 @2.4 GHz
512 GB
Oracle Solaris 10
Oracle 11g RAC
49,860 274,080 Parallel 2011040
Four Sun Fire X4470
4 x Intel Xeon X7560 @2.26 GHz
256 GB
Solaris 10
Oracle 11g RAC
40,000 221,020 Parallel 2010039

Complete benchmark results and descriptions can be found at the SAP standard applications benchmark website.
For SD benchmark results website: Two-Tier or Three-Tier. For SD Parallel benchmark results website: SD Parallel.

Configuration and Results Summary

Hardware Configuration:

8 x Sun Fire X4800 M2 servers, each with
8 x Intel Xeon E7-8870 @ 2.4 GHz (8 processors, 80 cores, 160 threads)
512 GB memory

Software Configuration:

SAP enhancement package 4 for SAP ERP 6.0
Oracle Database 11g Real Application Clusters (RAC)
Oracle Solaris 10

Results Summary:

Number of SAP SD benchmark users:
180,000
Average dialog response time:
0.63 seconds
Throughput:

Fully processed order line items per hour:
20,327,670

Dialog steps/hour:
60,983,000

SAPS:
1,016,380
Average database request time (dialog/update):
0.010 sec / 0.055 sec
SAP Certification:
2011037

Benchmark Description

The SAP Standard Application Sales and Distribution - Parallel (SD Parallel) Benchmark is a two-tier ERP business test that is indicative of full business workloads of complete order processing and invoice processing and demonstrates the ability to run both the application and database software on a single system. The SAP Standard Application SD Benchmark represents the critical tasks performed in real-world ERP business environments.

The SD Parallel Benchmark consists of the same transactions and user interaction steps as the two-tier and three-tier SD Benchmark. This means that the SD Parallel Benchmark runs the same business processes as the SD Benchmark. The difference between the benchmarks is the technical data distribution. Additionally, the benchmark requires equal distribution of the benchmark users across all database nodes for the used benchmark clients (round-robin method). Following this rule, all database nodes work on data of all clients. This avoids unrealistic configurations such as having only one client per database node.

The SAP Benchmark Council agreed to give the parallel benchmark a different name so that the difference can be easily recognized by any interested parties - customers, prospects, and analysts. The naming convention is SD Parallel for Sales & Distribution - Parallel.

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

SAP enhancement package 4 for SAP ERP 6.0 (Unicode) Sales and Distribution Benchmark, results as of 10/03/2011.

SD Parallel, 8 x Sun Fire X4800 M2 (each 8 processors, 80 cores, 160 threads) 180,000 SAP SD Users, Oracle Solaris 10, Oracle 11g Real Application Clusters (RAC), Certification Number 2011037.
SD Parallel, 6 x Sun Fire X4800 M2 (each 8 processors, 80 cores, 160 threads) 137,904 SAP SD Users, Oracle Solaris 10, Oracle 11g Real Application Clusters (RAC), Certification Number 2011038.
SD Parallel, 4 x Sun Fire X4470 (each 4 processors, 32 cores, 64 threads) 40,000 SAP SD Users, Oracle Solaris 10, Oracle 11g Real Application Clusters (RAC), Certification Number 2010039.
SD Two-Tier, IBM Power 795 (32 processors, 256 cores, 1024 threads) 126,063 SAP SD Users, AIX 7.1, DB2 9.7, Certification Number 2010046.

SAP, R/3 are registered trademarks of SAP AG in Germany and other countries. More information may be found at www.sap.com/benchmark.

Monday Sep 12, 2011

SPARC Enterprise M9000 Produces World Record SAP ATO Benchmark

Oracle delivered an SAP enhancement package 4 for SAP ERP 6.0 Assemble-to-Order (ATO) benchmark world record result using Oracle's SPARC Enterprise M9000 server running Oracle Solaris 10 and Oracle Database 11g along with SAP Enhancement Package 4 for SAP ERP 6.0 (Unicode). The SAP ATO benchmark integrates process chains across SAP Business Suite components, include Financials, Logistics, Human Resources, Basis and Cross Application.

  • The SPARC Enterprise M9000 server containing 64 SPARC64 VII+ 3.0 GHz processors, running Oracle Solaris 10 and Oracle Database 11g along with SAP Enhancement Package 4 for SAP ERP 6.0 (Unicode) delivered a world record 206,000 fully processed assembly orders per hour on the SAP enhancement package 4 for SAP ERP 6.0 ATO benchmark.

  • The SPARC Enterprise M9000 server result shows it can more than consolidate the work of the three-tier HP solution which used 80 different servers.

  • Oracle produced the first SAP ATO benchmark result using Unicode encoding.

  • The SAP ATO benchmark uses multiple components of the SAP Business Suite. See more detail at the SAP ATO benchmark webpage.

Performance Landscape

SAP ATO 2-Tier Performance Table (select results in decreasing performance order)

System OS
Database
Assembly Orders
per hour(*)
SAP
ERP/ECC
Release
Cert Num
SPARC Enterprise M9000
64 x SPARC64 VII+ @3.0 GHz
2048 GB
Oracle Solaris 10
Oracle 11g
206,360 SAP ERP6.0*
(Unicode)
2011033
Fujitsu Siemens Primepower 2000
128 x SPARC64 @560 MHz
128 GB
Solaris 8
Oracle 8.1.7
34,260 4.6B
(non-Unicode)
2001018
HP 9000 Superdome
64 x PA-RISC 8600 @552 MHz
128 GB
HP-UX 11.11
Oracle 8.16
18,870 4.6B
(non-Unicode)
2001014
Fujitsu Siemens Primepower 900
16 x SPARC64 V @1.35 GHz
64 GB
Solaris 8
Oracle 9i
12,170 4.6C
(non-Unicode)
2003012
HP rx5670
4 x Itanium II @1.0 GHz
24 GB
HP-UX 11i
Oracle 9i
3,090 4.6C
(non-Unicode)
2002069

(*) SAP enhancement package 4 for SAP ERP6.0 (Unicode)

SAP ATO 3-Tier Performance Table (top results in decreasing performance order)

System OS
Database
Assembly Orders
per hour(*)
SAP
ERP/ECC
Release
Cert Num
HP 9000 Superdome Enterprise Server
64 x PA-RISC 8700 @ 750MHz
128 GB
HP-UX 11i
Oracle 9i
144,090 4.6 C
(non-Unicode)
2002003
HP 9000 Superdome Enterprise Server
64 x PA-RISC 8700 @750 MHz
128 GB
HP-UX 11i
Oracle 9i
130,570 4.6 C
(non-Unicode)
2001047

(*) Assembly Order: Request to assemble pre-manufactured parts and assemblies to finished products according to an existing sales order.

Complete benchmark results may be found at the SAP benchmark website: http://www.sap.com/benchmark.

Configuration Summary and Results

Hardware Configuration:

SPARC Enterprise M9000
64 SPARC64 VII+ 3.0 GHz processor
2048 GB memory

Software Configuration:

Oracle Solaris 10
SAP enhancement package 4 for SAP ERP 6.0 (Unicode)
Oracle Database 11g

Certified Result:

Fully business processed Assembly Orders/hour:
206,360
SAP Certification Number:
2011033

Benchmark Description

The SAP ATO benchmark integrates process chains across SAP Business Suite components. The ATO scenario is characterized by high volume sales, short production times (from hours to one day), and individual assembly for such products as PCs, pumps, and cars. In general, each benchmark user has its own master data, such as material, vendor, or customer master data to avoid data locking situations. However, the ATO Benchmark has been designed to handle and overcome data locking situations - the ATO benchmark users access common master data, such as material, vendor, or customer master data. (source: http://www12.sap.com/solutions/benchmark/ato.epx).

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

SAP, R/3 are registered trademarks of SAP AG in Germany and other countries. More information may be found at www.sap.com/benchmark

Two-tier SAP ATO standard SAP ERP 6.0 2005/EP4 (Unicode) application benchmarks as of 09/04/11:
Oracle's SPARC Enterprise M9000 (64 processors, 256 cores, 512 threads) 206,360 Assembly Orders/hour, 64 x 3.0 GHz SPARC VIII, 2048 GB memory, Oracle 11g, Oracle Solaris 10, Certification Number 2011033.

Two-tier SAP ATO standard 4.6 C application benchmarks as of 09/04/11:
Fujitsu Siemens Primepower 900 (16-way SMP) 12,170 Assembly Orders/hour, 16 x 1.35 GHz SPARC64 V, 64 GB memory, Oracle 9i, Solaris 8, Certification Number 2003012.
HP rx5670 (4 processors SMP) 3,090 Assembly Orders/hour, 4 x 1.0 GHz Itanium II, 24 GB memory, Oracle 9i, HP-UX 11i, Certification Number 2002069.

Two-tier SAP ATO standard 4.6 B application benchmarks as of 09/04/11:
HP 9000 Superdome (64-way SMP) 18,8770 Assembly Orders/hour, 64 x 552 MHz PA-RISC 8600, 128 GB memory, Oracle 8.1.6, HP-UX 11.11, Certification Number 2001014.
Fujitsu Siemens Primepower 2000 (128 processors SMP) 34,260 Assembly Orders/hour, 128 x 560 MHz SPARC64, 128 GB memory, Oracle 8.1.7, Solaris 8, Certification Number 2001018.

Three-tier SAP ATO standard 4.6 C application benchmarks as of 09/04/11:
HP 9000 Superdome Enterprise Server (64 processors SMP) 144,090 Assembly Orders/hour, 64 x 750 MHz PA-RISC 8700, 128 GB memory, Oracle 9i, HP-UX 11i, Certification Number 2002003
HP 9000 Superdome Enterprise Server (64 processors SMP) 130,570 Assembly Orders/hour, 64 x 750 MHz PA-RISC 8700, 128 GB memory, Oracle 9i, HP-UX 11i, Certification Number 2001047

Friday Aug 12, 2011

Sun Blade X6270 M2 with Oracle WebLogic World Record 2 Processor SPECjEnterprise 2010 Benchmark

Oracle produced a World Record single application server using 2 chips result for the SPECjEnterprise2010 benchmark of 5,427.42 SPECjEnterprise2010 EjOPS using one of Oracle's Sun Blade X6270 M2 server module for the application tier and one Sun Blade X6270 M2 server module for the database.

  • The Sun Blade X6270 M2 server module equipped with two Intel Xeon X5690 processors running at 3.46 GHz, demonstrated 47% better performance compared to the 2-chip IBM System HS22 server result of 3,694.35 SPECjEnterprise2010 EjOPS using the same model of Intel Xeon X5690 processor.

  • The Sun Blade X6270 M2 server module running the application tier demonstrated 33% better performance compared to the 2-chip IBM Power 730 Express server result of 4,062.38 SPECjEnterprise2010 EjOPS.

  • The Sun Blade X6270 M2 server modules used Oracle WebLogic Server 11g Release 1 (10.3.5) application, Java SE 6 Update 26, and Oracle Database 11g Release 2 to produce this result.

Performance Landscape

Complete benchmark results are at the SPEC website, SPECjEnterprise2010 Results.

SPECjEnterprise2010 Performance Chart
as of 8/11/2011
Submitter EjOPS* Application Server Database Server
Oracle 5,427.42 1x Sun Blade X6270 M2
2x 3.46 GHz Intel Xeon X5690
Oracle WebLogic 11g (10.3.5)
1x Sun Blade X6270 M2
2x 3.46 GHz Intel Xeon X5690
Oracle 11g DB 11.2.0.2
IBM 4,062.38 1x IBM Power 730 Express
2x 3.5 GHz POWER 7
WebSphere Application Server V7
1x IBM BladeCenter PS701
1x 3.0 GHz POWER 7
IBM DB2 9.7 Workgroup Server Edition FP3a
IBM 3,694.35 1x IBM HS22
2x 3.46 GHz Intel Xeon X5690
WebSphere Application Server V8
1x IBM x3850 X5
2x 2.4 GHz Intel Xeon E7-4870
IBM DB2 9.7 FP3a

* SPECjEnterprise2010 EjOPS, bigger is better.

Configuration Summary

Application Server:
    1 x Sun Blade X6270 M2
      2 x 3.46 GHz Intel Xeon X5690
      48 GB memory
      4 x 10 GbE NIC
      Oracle Linux 5 Update 6
      Oracle WebLogic Server 11g Release 1 (10.3.5)
      Java HotSpot(TM) 64-Bit Server VM on Linux, version 1.6.0_26 (Java SE 6 Update 26)

Database Server:

    1 x Sun Blade X6270 M2
      2 x 3.46 GHz Intel Xeon X5690
      144 GB memory
      2 x 10 GbE NIC
      2 x Sun Storage 6180
      Oracle Linux 5 Update 6
      Oracle Database 11g Enterprise Edition Release 11.2.0.2

Benchmark Description

SPECjEnterprise2010 is the third generation of the SPEC organization's J2EE end-to-end industry standard benchmark application. The SPECjEnterprise2010 benchmark has been designed and developed to cover the Java EE 5.0 specification's significantly expanded and simplified programming model, highlighting the major features used by developers in the industry today. This provides a real world workload driving the Application Server's implementation of the Java EE specification to its maximum potential and allowing maximum stressing of the underlying hardware and software systems.

The workload consists of an end to end web based order processing domain, an RMI and Web Services driven manufacturing domain and a supply chain model utilizing document based Web Services. The application is a collection of Java classes, Java Servlets, Java Server Pages , Enterprise Java Beans, Java Persistence Entities (pojo's) and Message Driven Beans.

The SPECjEnterprise2010 benchmark heavily exercises all parts of the underlying infrastructure that make up the application environment, including hardware, JVM software, database software, JDBC drivers, and the system network.

The primary metric of the SPECjEnterprise2010 benchmark is jEnterprise Operations Per Second ("SPECjEnterprise2010 EjOPS"). The primary metric for the SPECjEnterprise2010 benchmark is calculated by adding the metrics of the Dealership Management Application in the Dealer Domain and the Manufacturing Application in the Manufacturing Domain. There is no price/performance metric in this benchmark.

Key Points and Best Practices

  • Two Oracle WebLogic server instances were started using numactl binding 1 instance per chip.
  • Two Oracle database listener processes were started and each was bound to a separate chip.
  • Additional tuning information is in the report at http://spec.org.

See Also

Disclosure Statement

SPEC and the benchmark name SPECjEnterprise are registered trademarks of the Standard Performance Evaluation Corporation. Sun Blade X6270 M2, 5,427.42 SPECjEnterprise2010 EjOPS; IBM Power 730 Express, 4,062.38 SPECjEnterprise2010 EjOPS; IBM System HS22, 3,694.35 SPECjEnterprise2010 EjOPS. Results from www.spec.org as of 8/11/2011.

Friday Jul 01, 2011

SPARC T3-1 Record Results Running JD Edwards EnterpriseOne Day in the Life Benchmark with Added Batch Component

Using Oracle's SPARC T3-1 server for the application tier and Oracle's SPARC Enterprise M3000 server for the database tier, a world record result was produced running the Oracle's JD Edwards EnterpriseOne applications Day in the Life benchmark run concurrently with a batch workload.

  • The SPARC T3-1 server based result has 25% better performance than the IBM Power 750 POWER7 server even though the IBM result did not include running a batch component.

  • The SPARC T3-1 server based result has 25% better space/performance than the IBM Power 750 POWER7 server as measured by the online component.

  • The SPARC T3-1 server based result is 5x faster than the x86-based IBM x3650 M2 server system when executing the online component of the JD Edwards EnterpriseOne 9.0.1 Day in the Life benchmark. The IBM result did not include a batch component.

  • The SPARC T3-1 server based result has 2.5x better space/performance than the x86-based IBM x3650 M2 server as measured by the online component.

  • The combination of SPARC T3-1 and SPARC Enterprise M3000 servers delivered a Day in the Life benchmark result of 5000 online users with 0.875 seconds of average transaction response time running concurrently with 19 Universal Batch Engine (UBE) processes at 10 UBEs/minute. The solution exercises various JD Edwards EnterpriseOne applications while running Oracle WebLogic Server 11g Release 1 and Oracle Web Tier Utilities 11g HTTP server in Oracle Solaris Containers, together with the Oracle Database 11g Release 2.

  • The SPARC T3-1 server showed that it could handle the additional workload of batch processing while maintaining the same number of online users for the JD Edwards EnterpriseOne Day in the Life benchmark. This was accomplished with minimal loss in response time.

  • JD Edwards EnterpriseOne 9.0.1 takes advantage of the large number of compute threads available in the SPARC T3-1 server at the application tier and achieves excellent response times.

  • The SPARC T3-1 server consolidates the application/web tier of the JD Edwards EnterpriseOne 9.0.1 application using Oracle Solaris Containers. Containers provide flexibility, easier maintenance and better CPU utilization of the server leaving processing capacity for additional growth.

  • A number of Oracle advanced technology and features were used to obtain this result: Oracle Solaris 10, Oracle Solaris Containers, Oracle Java Hotspot Server VM, Oracle WebLogic Server 11g Release 1, Oracle Web Tier Utilities 11g, Oracle Database 11g Release 2, the SPARC T3 and SPARC64 VII+ based servers.

  • This is the first published result running both online and batch workload concurrently on the JD Enterprise Application server. No published results are available from IBM running the online component together with a batch workload.

  • The 9.0.1 version of the benchmark saw some minor performance improvements relative to 9.0. When comparing between 9.0.1 and 9.0 results, the reader should take this into account when the difference between results is small.

Performance Landscape

JD Edwards EnterpriseOne Day in the Life Benchmark
Online with Batch Workload

This is the first publication on the Day in the Life benchmark run concurrently with batch jobs. The batch workload was provided by Oracle's Universal Batch Engine.

System Rack
Units
Online
Users
Resp
Time (sec)
Batch
Concur
(# of UBEs)
Batch
Rate
(UBEs/m)
Version
SPARC T3-1, 1xSPARC T3 (1.65 GHz), Solaris 10
M3000, 1xSPARC64 VII+ (2.86 GHz), Solaris 10
4 5000 0.88 19 10 9.0.1

Resp Time (sec) — Response time of online jobs reported in seconds
Batch Concur (# of UBEs) — Batch concurrency presented in the number of UBEs
Batch Rate (UBEs/m) — Batch transaction rate in UBEs/minute.

JD Edwards EnterpriseOne Day in the Life Benchmark
Online Workload Only

These results are for the Day in the Life benchmark. They are run without any batch workload.

System Rack
Units
Online
Users
Response
Time (sec)
Version
SPARC T3-1, 1xSPARC T3 (1.65 GHz), Solaris 10
M3000, 1xSPARC64 VII (2.75 GHz), Solaris 10
4 5000 0.52 9.0.1
IBM Power 750, 1xPOWER7 (3.55 GHz), IBM i7.1 4 4000 0.61 9.0
IBM x3650M2, 2xIntel X5570 (2.93 GHz), OVM 2 1000 0.29 9.0

IBM result from http://www-03.ibm.com/systems/i/advantages/oracle/, IBM used WebSphere

Configuration Summary

Hardware Configuration:

1 x SPARC T3-1 server
1 x 1.65 GHz SPARC T3
128 GB memory
16 x 300 GB 10000 RPM SAS
1 x Sun Flash Accelerator F20 PCIe Card, 96 GB
1 x 10 GbE NIC
1 x SPARC Enterprise M3000 server
1 x 2.86 SPARC64 VII+
64 GB memory
1 x 10 GbE NIC
2 x StorageTek 2540 + 2501

Software Configuration:

JD Edwards EnterpriseOne 9.0.1 with Tools 8.98.3.3
Oracle Database 11g Release 2
Oracle 11g WebLogic server 11g Release 1 version 10.3.2
Oracle Web Tier Utilities 11g
Oracle Solaris 10 9/10
Mercury LoadRunner 9.10 with Oracle Day in the Life kit for JD Edwards EnterpriseOne 9.0.1
Oracle’s Universal Batch Engine - Short UBEs and Long UBEs

Benchmark Description

JD Edwards EnterpriseOne is an integrated applications suite of Enterprise Resource Planning (ERP) software. Oracle offers 70 JD Edwards EnterpriseOne application modules to support a diverse set of business operations.

Oracle's Day in the Life (DIL) kit is a suite of scripts that exercises most common transactions of JD Edwards EnterpriseOne applications, including business processes such as payroll, sales order, purchase order, work order, and other manufacturing processes, such as ship confirmation. These are labeled by industry acronyms such as SCM, CRM, HCM, SRM and FMS. The kit's scripts execute transactions typical of a mid-sized manufacturing company.

  • The workload consists of online transactions and the UBE workload of 15 short and 4 long UBEs.

  • LoadRunner runs the DIL workload, collects the user’s transactions response times and reports the key metric of Combined Weighted Average Transaction Response time.

  • The UBE processes workload runs from the JD Enterprise Application server.

    • Oracle's UBE processes come as three flavors:

      • Short UBEs < 1 minute engage in Business Report and Summary Analysis,
      • Mid UBEs > 1 minute create a large report of Account, Balance, and Full Address,
      • Long UBEs > 2 minutes simulate Payroll, Sales Order, night only jobs.
    • The UBE workload generates large numbers of PDF files reports and log files.

    • The UBE Queues are categorized as the QBATCHD, a single threaded queue for large UBEs, and the QPROCESS queue for short UBEs run concurrently.

  • One of the Oracle Solaris Containers ran 4 Long UBEs, while another Container ran 15 short UBEs concurrently.

  • The mixed size UBEs ran concurrently from the SPARC T3-1 server with the 5000 online users driven by the LoadRunner.

  • Oracle’s UBE process performance metric is Number of Maximum Concurrent UBE processes at transaction rate, UBEs/minute.

Key Points and Best Practices

Two JD Edwards EnterpriseOne Application Servers and two Oracle Fusion Middleware WebLogic Servers 11g R1 coupled with two Oracle Fusion Middleware 11g Web Tier HTTP Server instances on the SPARC T3-1 server were hosted in four separate Oracle Solaris Containers to demonstrate consolidation of multiple application and web servers.

See Also

Disclosure Statement

Copyright 2011, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 6/27/2011.

Friday Jun 10, 2011

SPARC Enterprise M5000 Delivers First PeopleSoft Payroll 9.1 Benchmark

Oracle's M-series server sets a world record on Oracle's PeopleSoft Enterprise Payroll (N.A) 9.1 with extra large volume model benchmark (Unicode). Oracle's SPARC Enterprise M5000 server was able to to run faster than the previous generation system result even though the PeopleSoft Payroll 9.1 benchmark is more computationally demanding.

Oracle's SPARC Enterprise M5000 server configured with eight 2.66 GHz SPARC64 VII+ processors together with Oracle's Sun Storage F5100 Flash Array storage achieved world record performance on the Unicode version of Oracle's PeopleSoft Enterprise Payroll (N.A) 9.1 with extra large volume model benchmark using Oracle Database 11g Release 2 running on Oracle Solaris 10.

  • The SPARC Enterprise M5000 server processed payroll payments for the 500K employees PeopleSoft Payroll 9.1 (Unicode) benchmark in 46.76 minutes compared to a previous result of 50.11 minutes for the PeopleSoft Payroll 9.0 (non-Unicode) benchmark configured with 2.53 GHz SPARC64 VII processors resulting in 7% better performance.

  • Note that the IBM z10 Gen1 mainframe running the PeopleSoft Payroll 9.0 (Unicode) benchmark was 48% slower than the 9.0 non-Unicode version. The IBM z10 mainframe with nine 4.4 GHz Gen1 processors has a list price over $6M and is rated at 6,512 MIPS.

  • The SPARC Enterprise M5000 server with the Sun Storage F5100 Flash Array system processed payroll for 500K employees completing the end-to-end run in 66.28 mins, 11% faster than earlier published result of 73.88 mins with Payroll 9.0 configured with 2.53 GHz SPARC64 VII processors.

  • The Sun Storage F5100 Flash Array device is a high performance, high-density solid-state flash array which provides a read latency of only 0.5 msec which is about 10 times faster than the normal disk latencies of 5 msec measured on this benchmark.

Performance Landscape

PeopleSoft Payroll (N.A.) 9.1 – 500K Employees (7 Million SQL PayCalc, Unicode)

System Processor OS/Database Payroll Processing
Result (minutes)
Run 1
(minutes)
Num of
Streams
SPARC M5000 8x 2.66GHz SPARC64 VII+ Solaris/Oracle 11g 46.76 66.28 32

PeopleSoft Payroll (N.A.) 9.0 – 500K Employees (3 Million SQL PayCalc, Non-Unicode)

System Processor OS/Database Time in Minutes Num of
Streams
Payroll
Processing
Result
Run 1 Run 2 Run 3
Sun M5000 8x 2.53GHz SPARC64 VII Solaris/Oracle 11g 50.11 73.88 534.20 1267.06 32
IBM z10 9x 4.4GHz Gen1 Z/OS /DB2 58.96 80.5 250.68 462.6 8
IBM z10 9x 4.4GHz Gen1 Z/OS /DB2 87.4 ** 107.6 - - 8
HP rx7640 8x 1.6GHz Itanium2 HP-UX/Oracle 11g 96.17 133.63 712.72 1665.01 32

** This result was run with Unicode

Payroll 9.1 Compared to Payroll 9.0

Please note that Payroll 9.1 is Unicode based and Payroll 9.0 is non-Unicode. There are 7 million executions of an SQL statement for the PayCalc batch process in Payroll 9.1 and 3 million executions of the same SQL statement for the PayCalc batch process in Payroll 9.0. This gets reflected in the elapsed time (27.33 min for 9.1 and 23.78 min for 9.0). The elapsed times of all other batch processes is lower (better) on 9.1.

Configuration Summary

Hardware Configuration:

SPARC Enterprise M5000 server
8 x 2.66 GHz SPARC64 VII+ processors
128 GB memory
2 x SAS HBA (SG-XPCIE8SAS-E-Z - PCIe HBA for Rack Servers)
Sun Storage F5100 Flash Array
40 x 24 GB FMODs
1 x StorageTek 2501 array with
12 x 146 GB SAS 15K RPM disks
1 x StorageTek 2540 array with
12 x 146 GB SAS 15K RPM disks

Software Configuration:

Oracle Solaris 10 09/10
PeopleSoft HRMS and Campus Solutions 9.10.303
PeopleSoft Enterprise (PeopleTools) 8.51.035
Oracle Database 11g Release 2 11.2.0.1 (64-bit)
Micro Focus COBOLServer Express 5.1 (64-bit)

Benchmark Description

The PeopleSoft 9.1 Payroll (North America) benchmark is a performance benchmark established by PeopleSoft to demonstrate system performance for a range of processing volumes in a specific configuration. This information may be used to determine the software, hardware, and network configurations necessary to support processing volumes. This workload represents large batch runs typical of OLTP workloads during a mass update.

To measure five application business process run times for a database representing a large organization. The five processes are:

  • Paysheet Creation: Generates payroll data worksheets consisting of standard payroll information for each employee for a given pay cycle.

  • Payroll Calculation: Looks at paysheets and calculates checks for those employees.

  • Payroll Confirmation: Takes information generated by Payroll Calculation and updates the employees' balances with the calculated amounts.

  • Print Advice forms: The process takes the information generated by Payroll Calculations and Confirmation and produces an Advice for each employee to report Earnings, Taxes, Deduction, etc.

  • Create Direct Deposit File: The process takes information generated by the above processes and produces an electronic transmittal file that is used to transfer payroll funds directly into an employee's bank account.

For the benchmark, we collected at least three data points with different numbers of job streams (parallel jobs). This batch benchmark allows a maximum of thirty-two job streams to be configured to run in parallel.

See Also

Disclosure Statement

Oracle's PeopleSoft Payroll 9.1 benchmark, SPARC Enterprise M5000 46.76 min, www.oracle.com/apps_benchmark/html/white-papers-peoplesoft.html, results 6/10/2011.

Wednesday Mar 23, 2011

SPARC T3-1B Doubles Performance on Oracle Fusion Middleware WebLogic Avitek Medical Records Sample Application

The Oracle WebLogic Server 11g software was used to demonstrate the performance of the Avitek Medical Records sample application. A configuration using SPARC T3-1B and SPARC Enterprise M5000 servers from Oracle was used and showed excellent scaling of different configurations as well as doubling previous generation SPARC blade performance.

  • A SPARC T3-1B server, running a typical real-world J2EE application on Oracle WebLogic Server 11g, together with a SPARC Enterprise M5000 server running the Oracle database, had 2.1x times the transactional throughput over the previous generation UltraSPARC T2 processor based Sun Blade T6320 server module.

  • The SPARC T3-1B server shows linear scaling as the number of cores in the SPARC T3 processor used in the SPARC T3-1B system module are doubled.

  • The Avitek Medical Records application instances were deployed in Oracle Solaris zones on the SPARC T3-1B server, allowing for flexible, scalable and lightweight architecture of the application tier.

Performance Landscape

Performance for the application tier is presented. Results are the maximum transactions per second (TPS).

Server Processor Memory Maximum TPS
SPARC T3-1B 1 x SPARC T3, 1.65 GHz, 16 cores 128 GB 28,156
SPARC T3-1B 1 x SPARC T3, 1.65 GHz, 8 cores 128 GB 14,030
Sun Blade T6320 1 x UltraSPARC T2, 1.4 GHz, 8 cores 64 GB 13,386

The same SPARC Enterprise M5000 server from Oracle was used in each case as the database server. Internal disk storage was used.

Configuration Summary

Hardware Configuration:

1 x SPARC T3-1B
1 x 1.65 GHz SPARC T3
128 GB memory

1 x Sun Blade T6320
1 x 1.4Ghz GHz SPARC T2
64 GB memory

1 x SPARC Enterprise M5000
8 x 2.53 SPARC64 VII
128 GB memory

Software Configuration:

Avitek Medical Records
Oracle Database 10g Release 2
Oracle WebLogic Server 11g R1 version 10.3.3 (Oracle Fusion Middleware)
Oracle Solaris 10 9/10
HP Mercury LoadRunner 9.5

Benchmark Description

Avitek Medical Records (or MedRec) is an Oracle WebLogic Server 11g sample application suite that demonstrates all aspects of the J2EE platform. MedRec showcases the use of each J2EE component, and illustrates best practice design patterns for component interaction and client development. Oracle WebLogic server 11g is a key component of Oracle Fusion Middleware 11g.

The MedRec application provides a framework for patients, doctors, and administrators to manage patient data using a variety of different clients. Patient data includes:

  • Patient profile information: A patient's name, address, social security number, and log-in information.

  • Patient medical records: Details about a patient's visit with a physician, such as the patient's vital signs and symptoms as well as the physician's diagnosis and prescriptions.

MedRec comprises of two main Java EE applications supporting different user scenarios:

medrecEar – Patients log in to the web application (patientWebApp) to register their profile or edit. Patients can also view medical records or their prior visits. Administrators use the web application (adminWebApp) to approve or deny new patient profile requests. medrecEar also provides all of the controller and business logic used by the MedRec application suite, as well as the Web Service used by different clients.

physicianEar – Physicians and nurses login to the web application (physicianWebApp) to search and access patient profiles, create and review medical records, and prescribe medicine to patients. The physician application is designed to communicate using the Web Service provided in the medrecEar.

The medrecEAR and physicianEar application are deployed to Oracle WebLogic Server 11g instance called MedRecServer. The physicianEAR application communicates with the controller components of medrecEAR using Web Services.

The workload injected into the MedRec applications measures the average transactions per second for the following sequence:

  1. A client opens page http://{host}:7011/Start.jsp (MedRec)
  2. Patient completes Registration process
  3. Administrator login, approves the patient profile, and logout
  4. Physician connect to the on-line system and logs in
  5. Physician performs search for a patient and looks up patient's visit information
  6. Physician logs out
  7. Patient logs in and reviews the profile
  8. Patient makes changes to the profile and updates the information
  9. Patient logs out

Each of the above steps constitutes a single transaction.

Key Points and Best Practices

Please see the Oracle documentation on the Oracle Technical Network for tuning your Oracle WebLogic Server 11g deployment.

See Also

Disclosure Statement

Copyright 2011, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 3/22/2011.

Thursday Feb 17, 2011

SPARC T3-1 takes JD Edwards "Day In the Life" benchmark lead, beats IBM Power7 by 25%

Oracle's SPARC T3-1 server, running the application, together with Oracle's SPARC Enterprise M3000 server running the database, have achieved a record result of 5000 users, with 0.523 seconds of average transaction response time, for the online component of the "Day in the Life" JD Edwards EnterpriseOne benchmark.

  • The "Day in the Life" benchmark tests the Oracle JD Edwards EnterpriseOne applications, running Oracle Fusion Middleware WebLogic Server 11g R1, Oracle Fusion Middleware Web Tier Utilities 11g HTTP server and JD Edwards EnterpriseOne 9.0.1 in Oracle Solaris Containers, together with the Oracle Database 11g Release 2.

  • The SPARC T3-1 server is 25% faster and has better response time than the IBM P750 POWER7 system, when executing the JD Edwards EnterpriseOne 9.0.1 Day in the Life test, online component.

  • The SPARC T3-1 server had 25% better space/performance than the IBM P750 POWER7 server.

  • The SPARC T3-1 server is 5x faster than the x86-based IBM x3650 M2 server system, when executing the JD Edwards EnterpriseOne 9.0.1 Day in the Life test, online component.

  • The SPARC T3-1 server had 2.5x better space/performance than the x86-based IBM x3650 M2 server.

  • The SPARC T3-1 server consolidated the application/web tier of the JD Edwards EnterpriseOne 9.0.1 application using Oracle Solaris Containers. Containers provide flexibility, easier maintenance and better CPU utilization of the server leaving processing capacity for additional growth.

  • The SPARC Enterprise M3000 server provides enterprise class RAS features for customers deploying the Oracle 11g Release 2 database software.

  • To obtain this leading result, a number of Oracle advanced technology and features were used: Oracle Solaris 10, Oracle Solaris Containers, Oracle Java Hotspot Server VM, Oracle Fusion Middleware WebLogic Server 11g R1, Oracle Fusion Middleware Web Tier Utilities 11g, Oracle Database 11g Release 2, the SPARC T3 and the SPARC64 VII based servers.

Performance Landscape

JD Edwards EnterpriseOne DIL Online Component Performance Chart

System Memory OS #user JD Edwards
Version
Rack
Units
Response
Time
(sec)
SPARC T3-1, 1x1.65 GHz SPARC T3 128 Solaris 10 5000 9.0.1 2U 0.523
\*IBM Power 750, 1x3.55 GHz POWER7 120 IBM i7.1 4000 9.0 4U 0.61
IBM Power 570, 4x4.2 GHz POWER6 128 IBM i6.1 2400 8.12 4U 1.129
IBM x3650M2, 2x2.93 GHz X5570 64 OVM 1000 9.0 2U 0.29

\* from http://www-03.ibm.com/systems/i/advantages/oracle/, IBM used Websphere

Configuration Summary

Hardware Configuration:

1 x SPARC T3-1 server
1 x 1.65 GHz SPARC T3
128 GB memory
16 x 300 GB 10000 RPM SAS
1 x 1 GbE NIC
1 x SPARC Enterprise M3000
1 x 2.75 SPARC 64 VII
64 GB memory
1 x 1 GbE NIC
2 x StorageTek 2540/2501

Software Configuration:

JD Edwards EnterpriseOne 9.0.1 with Tools 8.98.3.3
Oracle Database 11g Release 2
Oracle Fusion Middleware 11g WebLogic server 11g R1 version 10.3.2
Oracle Fusion Middleware Web Tier Utilities 11g
Oracle Solaris 10 9/10
Mercury LoadRunner 9.10 with Oracle DIL kit for JD Edwards EnterpriseOne 9.0 update 1

Benchmark Description

Oracle's JD Edwards EnterpriseOne is an integrated applications suite of Enterprise Resource Planning software.

  • Oracle offers 70 JD Edwards EnterpriseOne application modules to support a diverse set of business operations.
  • Oracle 's Day-In-Life (DIL) kit is a suite of scripts that exercises most common transactions of J.D. Edwards EnterpriseOne applications including business processes such as payroll, sales order, purchase order, work order, and other manufacturing processes, such as ship confirmation. These are labeled by industry acronyms such as SCM, CRM, HCM, SRM and FMS.
  • Oracle's DIL kit's scripts execute transactions typical of a mid-sized manufacturing company.
  • The workload consists of online transactions. It does not include the batch processing job components.
  • LoadRunner is used to run the workload and collect the users' transactions response times against increasing numbers of users from 500 to 5000.
  • Key metric used to evaluate performance is the transaction response time which is reported by LoadRunner.

Key Points and Best Practices

Two JD Edwards EnterpriseOne and two Oracle Fusion Middleware WebLogic Servers 11g R1 coupled with two Fusion Middleware 11g Web Tier HTTP Servers instances on the SPARC T3-1 server were hosted in four separate Oracle Solaris Containers to demonstrate consolidation of multiple application and web servers.

  • Each Oracle Solaris container was bound to a separate processor set with 40 virtual processors allocated to each EnterpriseOne Server, 16 virtual processors allocated to each WebServer container and 16 to the default set. This was done to improve performance by using the physical memory closest to the processors, thereby, reducing memory access latency and reducing processor cross calls. The default processor set was used for network and disk interrupt handling.

  • The applications were executed in the FX scheduling class to improve performance by reducing the frequency of context switches.

  • A WebLogic Vertical cluster was configured on each WebServer container with seven managed instances each to load balance users' requests and to provide the infrastructure that enables scaling to high number of users with ease of deployment and high availability.

  • The database server was run in an Oracle Solaris Container hosted on the Oracle's SPARC Enterprise M3000 server.

See Also

Disclosure Statement

Copyright 2011, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 2/16/2011.

Tuesday Sep 28, 2010

SPARC T3-2 Delivers First Oracle E-Business X-Large Benchmark Self-Service (OLTP) Result

With Oracle's SPARC T3-2 server running the application and Oracle's Sun SPARC Enterprise M5000 server running the database, Oracle set a world record result for the Oracle E-Business Standard X-Large HR Self-Service (OLTP) benchmark.

  • The combination of a SPARC T3-2 server for the application and a Sun SPARC Enterprise M5000 server for the database achieved a result of 4000 HR Self-Service Online users on the Oracle E-Business X-Large benchmark dataset.

  • Oracle's Sun Storage F5100 Flash Array storage which was utilized in the benchmark was instrumental in obtaining an average transaction response time as low as 1.2 seconds.

  • Oracle has published the first Oracle E-Business R12.1.2 XL benchmark for 4000 HR Self-Service online users on a SPARC T3-2 server for the application tier and a Sun SPARC Enterprise M5000 server on database tier with Oracle 11g R2 database. Both servers ran with the Oracle Solaris 10 operating system.

  • The combination of the SPARC T3-2 server and Oracle E-Business R12.1.2 in the application tier with low CPU utilization provides headroom for growth.

  • The Sun Storage F5100 Flash Array storage provides higher performance with smaller footprint and lower power/cooling costs.

  • The result shows that the SPARC T3-2 server works well as a high capacity application server.

Performance Landscape

This is the FIRST published result for this X-large benchmark.

Workload HR Self-Service
X-Large Configuration
Vendor/System OS Users
SPARC T3-2 Oracle Solaris 10 9/10 4000

Results and Configuration Summary

Application Tier Configuration:

1 x SPARC T3-2 server
2 x SPARC T3 processors, 1.65 GHz
128 GB memory
Oracle Solaris 10 9/10
Oracle E-Business Suite 12.1.2

Database Tier Configuration:

1 x Sun SPARC Enterprise M5000 server
4 x SPARC64 VII processors, 2.53 GHz
128 GB memory
Oracle Solaris 10 10/09
Oracle Database 11g Release 2

Storage Configuration:

1 x Sun Storage F5100 Flash Array storage
1 x StorageTek 2540 array
300 GB

Benchmark Description

The Oracle R12 E-Business Standard Benchmark combines online transaction execution by simulated users with concurrent batch processing to model a typical scenario for a global enterprise. This benchmark includes one online component and 2 batch components. The goal is to obtain reference response times and throughput for Oracle EBS R12. Results can be published in four configurations:

  • X-large: Maximum online users running all business flows between 10,000 to 20,000; 750,000 order to cash lines per hour and 250,000 payroll checks per hour.
    • HR Self-Service Online -- 4000 users
      • The percentage across the 4 transactions in HR Self-Service module is:
        • Create Query Cash Expense -- 20%
        • Create Query Credit Expense -- 20%
        • View Payslip -- 30%
        • Create TimeCard -- 30%
    • Customer Support Flow -- 8000 users
    • Procure to Pay -- 2000 users
    • Order to Cash -- 2400 users
  • Large: 10,000 online users; 100,000 order to cash lines per hour and 100,000 payroll checks per hour.
  • Medium: up to 3000 online users; 50,000 order to cash lines per hour and 10,000 payroll checks per hour.
  • Small: up to 1000 online users; 10,000 order to cash lines per hour and 5,000 payroll checks per hour.

See Also

Disclosure Statement

Copyright 2010, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 9/20/2010.

Wednesday Sep 22, 2010

SPARC T3-1 Supports 13,000 Users on Financial Services and Enterprise Application Integration Running Siebel CRM 8.1.1

The Siebel CRM 8.1.1 benchmark was run on a configuration including Oracle's SPARC T3-1 server and Oracle's Sun SPARC Enterpise T5240 server running Siebel CRM 8.1.1 application along with Oracle Database 11g Release 2 on the Oracle Solaris 10 operating system.

  • Using a SPARC T3-1 server for the application tier and a Sun SPARC Enterprise T5240 server for the web and database tier, Oracle achieved 13,000 Siebel benchmark concurrent users using the customer representative Siebel workload with Siebel CRM 8.1.1 industry applications.

  • The SPARC T3-1 server in the application tier was able to support 13,000 concurrent users with a rate of 116,449 business transactions throughput per hour for Enterprise Application Integration.

  • The SPARC T3-1 server in the application tier was able to support 13,000 concurrent users with a rate of 48,409 business transactions throughput per hour for Siebel Financial Services.

  • The average response time for the Financial Services portion was 0.4 seconds and for the Web Services was 0.2 seconds.

  • Oracle Solaris Containers were used to virtualize the Web and DB tier of the benchmark.

Performance Landscape

Results are presented in separate sections based upon the version of the benchmark.

Sponsor Users Version Servers
Oracle 13,000 8.1.1 App: SPARC T3-1 (1.65GHz/SPARC T3)
Web/DB: Sun SPARC Enterprise T5240 (1.6GHz/UltraSPARC T2 Plus)
Oracle 14,000 8.0
App/Web/DB: Sun SPARC Enterprise T5440 (1.4GHz/UltraSPARC T2 Plus)

HP 12,000 8.0 App: 3 x HP BL460c (3.16GHz/Xeon)
Web: HP BL460c (3.0GHz/Xeon)
DB: HP BL680c (2.4GHz/Xeon)
Oracle 10,000 8.0 App: Sun SPARC Enterprise T5240 (1.2GHz/UltraSPARC T2 Plus)
Web: Sun SPARC Enterprise T5120 (1.2GHz/UltraSPARC T2)
DB: Sun SPARC Enterprise T5120 (1.2GHz/UltraSPARC T2)
HP 10,000 8.0 App: 2 x HP BL460c (3.16GHz/Xeon)
Web: HP BL460c (3.16GHz/Xeon)
DB: HP BL680c (2.4GHz/Xeon)
IBM 7,000 8.0 App: IBM P570 (4.7GHz/P6)
Web: IBM P570 (1.9GHz/P5+)
DB: IBM P570 (4.7GHz/P6)

Results and Configuration Summary

Hardware Configuration:

1 x SPARC T3-1 server (App server)
1 x SPARC T3 processor, 1.65 GHz
64 GB memory
1 x Sun SPARC Enterprise T5240 server (Web server and DB server)
2 x UltraSPARC T2 Plus processors
64 GB memory
1 x Sun Storage 6180 array

Software Configuration:

Oracle Solaris 10 9/10
Siebel 8.1.1 SIA
Oracle Database 11g Release 2 11.2.0.1 64-bit
Sun Java Web Server 7.0 update 8

Benchmark Description

Siebel CRM 8.1.1 workload is a multi-tier benchmark designed to stress the Siebel CRM Release 8.1.1 architecture and to demonstrate that large customers can successfully deploy many thousands of concurrent users.

The test simulates real-world requirements of a large organization, consisting of thousands of concurrent, active users from multiple departments accessing a call center. Test conditions simulate service representatives running Siebel Financial Services Call Center and Enterprise Application Integration component with Siebel Web Services.

Among the Siebel CRM Release 8.1 architecture features exercised are the following:

  • Smart Web Architecture: Takes advantage of the newest Web browser technology to deliver a highly interactive experience. The interaction model, which is similar to Windows-based applications, also improves productivity. Utilization rates on the web server are low, allowing customers to retain existing Web server infrastructure.

  • Smart Network Architecture: Allows Siebel CRM Release 8.1.1 customers to leverage their existing network infrastructure by compressing and caching user interface components, so that browser/Web server interaction occurs only when the application requests data. This allows customers to avoid expensive network upgrades that can be necessary with competing products.

  • Server Connection Broker: The Siebel Connection Broker (SCBroker) is a server component that provides intraserver loadbalancing. SCBroker distributes server requests across multiple instances of Application Object Managers (AOMs) running on a Siebel server.

  • Smart Database Connection Pooling and Multiplexing: Allows customers to scale their database without introducing expensive and complex transaction-processing monitors.

  • Server Request Broker: Server Request Broker (SRBroker) processes synchronous server requests - requests that must be run immediately, and for which the calling process waits for completion.

  • Enterprise Application Integration: Allows customers to integrate their existing systems with Siebel CRM applications.

  • eScript: eScript is a scripting or programming language that application developers use to write simple scripts to extend Siebel applications. Javascript, a popular scripting language used primarily on Web sites, is its core language.

See Also

Disclosure Statement

Copyright 2010, Oracle and/or its affiliates. All rights reserved. Oracle and Java are registered trademarks of Oracle and/or its affiliates. Other names may be trademarks of their respective owners. Results as of 9/20/2010.

Monday Sep 20, 2010

Sun Fire X4470 4 Node Cluster Delivers World Record SAP SD-Parallel Benchmark Result

Oracle delivered an SAP enhancement package 4 for SAP ERP 6.0 Sales and Distribution – Parallel (SD-Parallel) Benchmark world record result using four of Oracle's Sun Fire X4470 servers, Oracle Solaris 10 and Oracle 11g Real Application Clusters (RAC) software.

  • The Sun Fire X4470 servers delivered 8% more performance compared to the IBM Power 780 server running the SAP enhancement package 4 for SAP ERP 6.0 Sales and Distribution benchmark.

  • The Sun Fire X4470 servers result of 40,000 users delivered 2.2 times the performance of the HP ProLiant DL980 G7 result of 18,180 users.

  • The Sun Fire X4470 servers result of 40,000 users delivered 2.5 times the performance of the Fujitsu PRIMEQUEST 1800E result of 16,000 users.

This result shows that a complete software and hardware solution from Oracle using Oracle RAC, Oracle Solaris and Sun servers provides a superior performing solution.

Performance Landscape

Selected SAP Sales and Distribution benchmark results are presented in decreasing order in performance. All benchmarks were using SAP enhancement package 4 for SAP ERP 6.0 (Unicode) except the result marked with an asterix (\*) which was achieved with SAP ERP 6.0.

System OS
Database
Users SAPS Type Date
Four Sun Fire X4470
4xIntel Xeon X7560 @2.26GHz
256 GB
Solaris 10
Oracle 11g Real Application Clusters
40,000 221,014 Parallel 20-Sep-10
Five IBM System p 570 (\*)
8xPOWER6 @4.7GHz
128 GB
AIX 5L Version 5.3
Oracle 10g Real Application Clusters
37,040 187,450 Parallel "non-Unicode" 25-Mar-08
IBM Power 780
8xPOWER7 @3.8GHz
1 TB
AIX 6.1
DB2 9.7
37,000 202,180 2-Tier 7-Apr-10
Two Sun Fire X4470
4xIntel Xeon X7560 @2.26GHz
256 GB
Solaris 10
Oracle 11g Real Application Clusters
21,000 115,300 Parallel 28-Jun-10
HP DL980 G7
8xIntel Xeon X7560 @2.26GHz
512 GB
Win Server 2008 R2 DE
SQL Server 2008
18,180 99,320 2-Tier 21-Jun-10
Fujitsu PRIMEQUEST 1800E
8xIntel Xeon X7560 @2.26GHz
512 GB
Win Server 2008 R2 DE
SQL Server 2008
16,000 87,550 2-Tier 30-Mar-10
Four Sun Blade X6270
2xIntel Xeon X5570 @2.93GHz
48 GB
Solaris 10
Oracle 10g Real Application Clusters
13,718 75,762 Parallel 12-Oct-09
HP DL580 G7
4xIntel Xeon X7560 @2.26GHz
256 GB
Win Server 2008 R2 DE
SQL Server 2008
10,445 57,020 2-Tier 21-Jun-10
Two Sun Blade X6270
2xIntel Xeon X5570 @2.93GHz
48 GB
Solaris 10
Oracle 10g Real Application Clusters
7,220 39,420 Parallel 12-Oct-09
One Sun Blade X6270
2xIntel Xeon X5570 @2.93GHz
48 GB
Solaris 10
Oracle 10g Real Application Clusters
3,800 20,750 Parallel 12-Oct-09

Complete benchmark results and a description can be found at the SAP benchmark website http://www.sap.com/solutions/benchmark/sd.epx.

Results and Configuration Summary

Hardware Configuration:

4 x Sun Fire X4470 servers, each with
4 x Intel Xeon X7560 2.26 GHz (4 chips, 32 cores, 64 threads)
256 GB memory

Software Configuration:

Oracle 11g Real Application Clusters (RAC)
Oracle Solaris 10

Results Summary:

Number of SAP SD benchmark users:
40,000
Average dialog response time:
0.86 seconds
Throughput:

Dialog steps/hour:
13,261,000

SAPS:
221,020
SAP Certification:
2010039

Benchmark Description

SAP is one of the premier world-wide ERP application providers and maintains a suite of benchmark tests to demonstrate the performance of competitive systems running the various SAP products.

The SAP Standard Application SD Benchmark represents the critical tasks performed in real-world ERP business environments. The SAP Standard Application Sales and Distribution - Parallel (SD-Parallel) Benchmark is a two-tier ERP business test that is indicative of full business workloads of complete order processing and invoice processing and demonstrates the ability to run both the application and database software on a single system.

The SD-Parallel Benchmark consists of the same transactions and user interaction steps as the SD Benchmark. This means that the SD-Parallel Benchmark runs the same business processes as the SD Benchmark. The difference between the benchmarks is the technical data distribution.

The additional rule for parallel and distributed databases is one must equally distribute the benchmark users across all database nodes for the used benchmark clients (round-robin method). Following this rule, all database nodes work on data of all clients. This avoids unrealistic configurations such as having only one client per database node.

The SAP Benchmark Council agreed to give the parallel benchmark a different name so that the difference can be easily recognized by any interested parties - customers, prospects, and analysts. The naming convention is SD-Parallel for Sales & Distribution - Parallel.

In January 2009, a new version of the SAP enhancement package 4 for SAP ERP 6.0 (Unicode) Sales and Distribution (SD) Benchmark was released. This new release has higher cpu requirements and so yields from 25-50% fewer users compared to the previous (non-unicode) Standard Sales and Distribution (SD) Benchmark. Between 10-30% of this greater load is due to the extra overhead from the processing of the larger character strings due to Unicode encoding.

Unicode is a computing standard that allows for the representation and manipulation of text expressed in most of the world's writing systems. Before the Unicode requirement, this benchmark used ASCII characters meaning each was just 1 byte. The new version of the benchmark requires Unicode characters and the Application layer (where ~90% of the cycles in this benchmark are spent) uses a new encoding, UTF-16, which uses 2 bytes to encode most characters (including all ASCII characters) and 4 bytes for some others. This requires computers to do more computation and use more bandwidth and storage for most character strings. Refer to the above SAP Note for more details.

See Also

Disclosure Statement

SAP enhancement package 4 for SAP ERP 6.0 (Unicode) Sales and Distribution Benchmark, results as of 9/19/2010. For more details, see http://www.sap.com/benchmark. SD-Parallel, Four Sun Fire X4470 (each 4 processors, 32 cores, 64 threads) 40,000 SAP SD Users, Cert# 2010039. SD-Parallel, Two Sun Fire X4470 (each 4 processors, 32 cores, 64 threads) 21,000 SAP SD Users, Cert# 2010029. SD 2-Tier, HP ProLiant DL980 G7 (8 processors, 64 cores, 128 threads) 18,180 SAP SD Users, Cert# 2010028. SD 2-Tier, Fujitsu PRIMEQUEST 1800E (8 processors, 64 cores, 128 threads) 16,000 SAP SD Users, Cert# 2010010. SD-Parallel, Four Sun Blade X6270 (each 2 processors, 8 cores, 16 threads) 13,718 SAP SD Users, Cert# 2009041. SD 2-Tier, HP ProLiant DL580 G7 (4 processors, 32 cores, 64 threads) 10,490 SAP SD Users, Cert# 2010032. SD 2-Tier, IBM System x3850 X5 (4 processors, 32 cores, 64 threads) 10,450 SAP SD Users, Cert# 2010012. SD 2-Tier, Fujitsu PRIMERGY RX600 S5 (4 processors, 32 cores, 64 threads) 9,560 SAP SD Users, Cert# 2010017. SD-Parallel, Two Sun Blade X6270 (each 2 processors, 8 cores, 16 threads) 7,220 SAP SD Users, Cert# 2009040. SD-Parallel, Sun Blade X6270 (2 processors, 8 cores, 16 threads) 3,800 SAP SD Users, Cert# 2009039. SD 2-Tier, Sun Fire X4270 (2 processors, 8 cores, 16 threads) 3,800 SAP SD Users, Cert# 2009033.

SAP ERP 6.0 (Unicode) Sales and Distribution Benchmark, results as of 9/19/2010. SD-Parallel, Five IBM System p 570 (each 8 processors, 16 cores, 32 threads) 37,040 SAP SD Users, Cert# 2008013.

Wednesday Aug 25, 2010

Transparent Failover with Solaris MPxIO and Oracle ASM

Recently I had a nice experience configuring failover with Solaris MPxIO (Multipathing), Oracle ASM (Automatic Storage Management) and Oracle's Sun Storage 6180 array. In this configuration failover was completely transparent to the Oracle database.

MPxIO is part of the Solaris 10 distribution. Multipathing provides redundant paths and eliminates a single point of failure by failing over to an alternate path automatically when one of the paths fails. Oracle ASM was used for volume manager to manage Oracle database files and it works very well with Multipathing. The Sun Storage 6180 array configuration had two controllers and each controller was connected to a different adapter on the host for redundant path. Multipathing generates a pseudo device and by providing Oracle ASM a pseudo device name at ASM diskgroup creation it makes failover transparent to the Oracle database.

HW/SW Configuration

Sun SPARC Enterprise M9000, 2x 8Gb FC HBA
1xSun Storage 6180 array, 16x 300G 15K RPM disks, two controllers, Firmware 07.60.18.13
Solaris 10, Oracle 11g with Oracle ASM
MPxIO enabled

When one of the paths fails (e.g. Cable, adapter, controller) system messages in /var/adm/messages file show the following sequence of events.

    Link down      <- the path is down
    offlining lun= ...       <- associated devices are offline
    multipath status: degraded      <- multipath status is degraded
    Initiating failover for device       <- Initiate failover
    Failover operation completed successfully      <- failover was successful
However during this period Oracle database transactions continue without any interruption and the Oracle database doesn't even know this is happening. The failover is transparent to the Oracle database and there was no error or message in Oracle log files during this period.

The following shows the steps and examples to set up MPxIO and Oracle ASM. At the bottom of the page it shows system messages from failover.

Sun Storage 6180 array

First you need to configure two paths physically. In this case, one controller was connected to one of the HBAs on the host and the other controller was connected to the other HBA.


Use Sun StorageTEK Common Array Manager GUI tool or command tool 'sscs' (StorEdge Systems Element Manager Command Line Interface) to manage 6180.
6180 'os-type' needs to be set properly to support MPxIO.
To view the current 'os-type' setting run 'sscs list array array_name'.
'Default Host Type' should be 'SOLARIS_MPXIO - Solaris (with Traffic Manager)'.
Run 'sscs list -a array_name os-type' to view Operating Systems supported and 'sscs modify -o os-type_name array array_name' to modify 'os-type'.

Host

Solaris 'stmsboot -e' command enables MPxIO. It will prompt you to reboot the host.


After the reboot 'format' shows a long pseudo name such as c29t60080E500018142E000003574C5 EBF38d0
But 'stmsboot -L' provides both before and after MPxIO. This mapping is very useful for any debugging you might need.

#stmsboot -L

 non-STMS device name                    STMS device name
------------------------------------------------------------------
/dev/rdsk/c6t0d0        /dev/rdsk/c29t60080E5000184350000003614C5EBE23d0
/dev/rdsk/c6t0d2        /dev/rdsk/c29t60080E5000184350000003654C5EBE44d0
/dev/rdsk/c6t0d1        /dev/rdsk/c29t60080E50001845C6000003524C5EBE2Ed0
/dev/rdsk/c6t0d3        /dev/rdsk/c29t60080E50001845C6000003564C5EBE4Cd0
/dev/rdsk/c9t4d0        /dev/rdsk/c29t60080E5000184350000003614C5EBE23d0
/dev/rdsk/c9t4d2        /dev/rdsk/c29t60080E5000184350000003654C5EBE44d0
/dev/rdsk/c9t4d1        /dev/rdsk/c29t60080E50001845C6000003524C5EBE2Ed0
/dev/rdsk/c9t4d3        /dev/rdsk/c29t60080E50001845C6000003564C5EBE4Cd0
Notice c6t0 and c9t4 point to the same STMS device name (e.g. c6t0d0 and c9t4d0 point to c29t60080E5000184350000003614C5EBE23d0).

Oracle ASM

Create a ASM diskgroup by providing MPxIO pseudo device names.

Initialize disks by doing 'dd if=/dev/zero of=/dev/rdsk/c29t60080E50001843500 00003614C5EBE23d0s0 bs=1204k count=10.


Set owner to Oracle 'chown oracle:dba /dev/rdsk/c29t60080E500018435000000 3614C5EBE23d0s0'
create diskgroup diskgroup_name external redundancy disk
   '/dev/rdsk/c29t60080E5000184350000003614C5EBE23d0s0' size 10g,
   '/dev/rdsk/c29t60080E5000184350000003654C5EBE44d0s0' size 10g,
   '/dev/rdsk/c29t60080E50001845C6000003524C5EBE2Ed0s0' size 10g,
   '/dev/rdsk/c29t60080E50001845C6000003564C5EBE4Cd0s0' size 10g;

System messages from failover

When one of the paths fails at the array side 'sscs list -a array_name fcport ' shows one of the ports is down.

And /var/adm/messages file reports:

Aug 10 11:17:33 host1 emlxs: [ID 349649 kern.info] [ 5.0314]emlxs0: NOTICE: 710: Link down.
Aug 10 11:19:03 host1 fctl: [ID 517869 kern.warning] WARNING: fp(2)::OFFLINE timeout
Aug 10 11:19:22 host1 scsi: [ID 243001 kern.info] /pci@6,600000/SUNW,emlxs@0/fp@0,0 (fcp2):
Aug 10 11:19:22 host1    offlining lun=3 (trace=0), target=ef (trace=2800004)
Aug 10 11:19:22 host1 scsi: [ID 243001 kern.info] /pci@6,600000/SUNW,emlxs@0/fp@0,0 (fcp2):
Aug 10 11:19:22 host1    offlining lun=2 (trace=0), target=ef (trace=2800004)
Aug 10 11:19:22 host1 scsi: [ID 243001 kern.info] /pci@6,600000/SUNW,emlxs@0/fp@0,0 (fcp2):
Aug 10 11:19:22 host1    offlining lun=1 (trace=0), target=ef (trace=2800004)
Aug 10 11:19:22 host1 scsi: [ID 243001 kern.info] /pci@6,600000/SUNW,emlxs@0/fp@0,0 (fcp2):
Aug 10 11:19:22 host1    offlining lun=0 (trace=0), target=ef (trace=2800004)
Aug 10 11:19:22 host1 genunix: [ID 408114 kern.info] /pci@6,600000/SUNW,emlxs@0/fp@0,0/ssd@w2014
0080e51845c6,1f (ssd0) offline
Aug 10 11:19:22 host1 genunix: [ID 834635 kern.info] /scsi_vhci/ssd@g60080e50001845c6000003564c5
ebe4c (ssd165) multipath status: degraded, path /pci@6,600000/SUNW,emlxs@0/fp@0,0 (fp2) to targe
t address: w20140080e51845c6,3 is offline Load balancing: round-robin
Aug 10 11:19:22 host1 genunix: [ID 834635 kern.info] /scsi_vhci/ssd@g60080e5000184350000003654c5
ebe44 (ssd166) multipath status: degraded, path /pci@6,600000/SUNW,emlxs@0/fp@0,0 (fp2) to targe
t address: w20140080e51845c6,2 is offline Load balancing: round-robin
Aug 10 11:19:22 host1 genunix: [ID 834635 kern.info] /scsi_vhci/ssd@g60080e50001845c6000003524c5
ebe2e (ssd167) multipath status: degraded, path /pci@6,600000/SUNW,emlxs@0/fp@0,0 (fp2) to targe
t address: w20140080e51845c6,1 is offline Load balancing: round-robin
Aug 10 11:19:22 host1 genunix: [ID 834635 kern.info] /scsi_vhci/ssd@g60080e5000184350000003614c5
ebe23 (ssd168) multipath status: degraded, path /pci@6,600000/SUNW,emlxs@0/fp@0,0 (fp2) to targe
t address: w20140080e51845c6,0 is offline Load balancing: round-robin
Aug 10 11:19:22 host1 scsi: [ID 243001 kern.info] /scsi_vhci (scsi_vhci0):
Aug 10 11:19:22 host1    Initiating failover for device ssd (GUID 60080e5000184350000003614c5ebe
23)
Aug 10 11:19:23 host1 scsi: [ID 243001 kern.info] /scsi_vhci (scsi_vhci0):
Aug 10 11:19:23 host1    Failover operation completed successfully for device ssd (GUID 60080e50
00184350000003614c5ebe23): failed over from primary to secondary
Aug 10 11:23:10 host1 emlxs: [ID 349649 kern.info] [ 5.0536]emlxs0: NOTICE: 720: Link up. (8Gb,
loop, initiator)
Aug 10 11:23:10 host1 genunix: [ID 936769 kern.info] ssd0 is /pci@6,600000/SUNW,emlxs@0/fp@0,0/s
sd@w20140080e51845c6,1f
Aug 10 11:23:11 host1 genunix: [ID 408114 kern.info] /pci@6,600000/SUNW,emlxs@0/fp@0,0/ssd@w2014
0080e51845c6,1f (ssd0) online
Aug 10 11:23:11 host1 genunix: [ID 834635 kern.info] /scsi_vhci/ssd@g60080e50001845c6000003564c5
ebe4c (ssd165) multipath status: optimal, path /pci@6,600000/SUNW,emlxs@0/fp@0,0 (fp2) to target
 address: w20140080e51845c6,3 is standby Load balancing: round-robin
Aug 10 11:23:11 host1 genunix: [ID 834635 kern.info] /scsi_vhci/ssd@g60080e5000184350000003654c5
ebe44 (ssd166) multipath status: optimal, path /pci@6,600000/SUNW,emlxs@0/fp@0,0 (fp2) to target
 address: w20140080e51845c6,2 is online Load balancing: round-robin
Aug 10 11:23:11 host1 genunix: [ID 834635 kern.info] /scsi_vhci/ssd@g60080e50001845c6000003524c5
ebe2e (ssd167) multipath status: optimal, path /pci@6,600000/SUNW,emlxs@0/fp@0,0 (fp2) to target
 address: w20140080e51845c6,1 is standby Load balancing: round-robin
Aug 10 11:23:11 host1 genunix: [ID 834635 kern.info] /scsi_vhci/ssd@g60080e5000184350000003614c5
ebe23 (ssd168) multipath status: optimal, path /pci@6,600000/SUNW,emlxs@0/fp@0,0 (fp2) to target
 address: w20140080e51845c6,0 is standby Load balancing: round-robin
Aug 10 11:23:11 host1 scsi: [ID 243001 kern.info] /scsi_vhci (scsi_vhci0):
Aug 10 11:23:11 host1    Initiating failover for device ssd (GUID 60080e5000184350000003614c5ebe
23)
Aug 10 11:23:13 host1    Failover operation completed successfully for device ssd (GUID 60080e50
00184350000003614c5ebe23): failed over from secondary to primary
Aug 10 11:23:13 host1 scsi: [ID 243001 kern.info] /scsi_vhci (scsi_vhci0):
Aug 10 11:23:13 host1    Auto failback operation succeeded for device ssd (GUID 60080e5000184350
000003614c5ebe23)
That's it. Try this for yourself.

Monday Jun 28, 2010

Sun Fire X4470 2-Node Configuration Sets World Record for SAP SD-Parallel Benchmark

Using two of Oracle's Sun Fire X4470 servers to run the SAP Enhancement Package 4 for SAP ERP 6.0 (Unicode) Sales and Distribution – Parallel (SD-Parallel) standard application benchmark, Oracle delivered a world record result. This was run using Oracle Solaris 10 and Oracle 11g Real Application Clusters (RAC) software.

  • The Sun Fire X4470 servers result of 21,000 users delivered more than twice the performance of the IBM System x3850 X5 system result of 10,450 users.

  • The Sun Fire X4470 servers result of 21,000 users beat the HP ProLiant DL980 G7 system result of 18,180 users. Both solutions used 8 Intel Xeon X7560 processors.

  • The Sun Fire X4470 servers result of 21,000 users beat the Fujitsu PRIMEQUEST 1800E system result of 16,000 users. Both solutions used 8 Intel Xeon X7560 processors.

  • This result shows how a compete software and hardware solution from Oracle, using Oracle RAC, Oracle Solaris and along with Oracle's Sun servers, can provide a superior performing solution when compared to the competition.

Performance Landscape

SAP Enhancement Package 4 for SAP ERP 6.0 (Unicode) Sales and Distribution Benchmark, select results presented in decreasing performance order. Both Parallel and 2-Tier solution results are listed in the table.

System OS
Database
Users SAPS Type Date
Two Sun Fire X4470
4xIntel Xeon X7560 @2.26GHz
256 GB
Solaris 10
Oracle 11g Real Application Clusters
21,000 115,300 Parallel 28-Jun-10
HP DL980 G7
8xIntel Xeon X7560 @2.26GHz
512 GB
Win Server 2008 R2 DE
SQL Server 2008
18,180 99,320 2-Tier 21-Jun-10
Fujitsu PRIMEQUEST 1800E
8xIntel Xeon X7560 @2.26GHz
512 GB
Win Server 2008 R2 DE
SQL Server 2008
16,000 87,550 2-Tier 30-Mar-10
Four Sun Blade X6270
2xIntel Xeon X5570 @2.93GHz
48 GB
Solaris 10
Oracle 10g Real Application Clusters
13,718 75,762 Parallel 12-Oct-09
IBM System x3850 X5
4xIntel Xeon X7560 @2.26GHz
256 GB
Win Server 2008 EE
DB2 9.7
10,450 57,120 2-Tier 30-Mar-10
HP DL580 G7
4xIntel Xeon X7560 @2.26GHz
256 GB
Win Server 2008 R2 DE
SQL Server 2008
10,445 57,020 2-Tier 21-Jun-10
Fujitsu PRIMERGY RX600 S5
4xIntel Xeon X7560 @2.26GHz
512 GB
Win Server 2008 R2 DE
SQL Server 2008
9,560 52,300 2-Tier 06-May-10
Two Sun Blade X6270
2xIntel Xeon X5570 @2.93GHz
48 GB
Solaris 10
Oracle 10g Real Application Clusters
7,220 39,420 Parallel 12-Oct-09
One Sun Blade X6270
2xIntel Xeon X5570 @2.93GHz
48 GB
Solaris 10
Oracle 10g Real Application Clusters
3,800 20,750 Parallel 12-Oct-09
Sun Fire X4270
2xIntel Xeon X5570 @2.93GHz
48 GB
Solaris 10
Oracle 10g
3,800 21,000 2-Tier 21-Aug-09

Complete benchmark results may be found at the SAP benchmark website http://www.sap.com/solutions/benchmark/sd.epx.

Results and Configuration Summary

Hardware Configuration:

2 x Sun Fire X4470 servers, each with
4 x Intel Xeon X7560 2.26 GHz (4 chips, 32 cores, 64 threads)
256 GB memory

Software Configuration:

Oracle 11g Real Application Clusters (RAC)
Oracle Solaris 10

Results Summary:

Number of SAP SD benchmark users:
21,000
Average dialog response time:
0.93 seconds
Throughput:

Dialog steps/hour:
6,918,000

SAPS:
115,300
SAP Certification:
2010029

Benchmark Description

The SAP Standard Application Sales and Distribution - Parallel (SD-Parallel) Benchmark is a two-tier ERP business test that is indicative of full business workloads of complete order processing and invoice processing, and demonstrates the ability to run both the application and database software on a single system. The SAP Standard Application SD Benchmark represents the critical tasks performed in real-world ERP business environments.

The SD-Parallel Benchmark consists of the same transactions and user interaction steps as the SD Benchmark. This means that the SD-Parallel Benchmark runs the same business processes as the SD Benchmark. The difference between the benchmarks is the technical data distribution.

An additional rule for parallel and distributed databases is one must equally distribute the benchmark users across all database nodes for the used benchmark clients (round-robin-method). Following this rule, all database nodes work on data of all clients. This avoids unrealistic configurations such as having only one client per database node.

The SAP Benchmark Council agreed to give the parallel benchmark a different name so that the difference can be easily recognized by any interested parties - customers, prospects, and analysts. The naming convention is SD-Parallel for Sales & Distribution - Parallel.

SAP is one of the premier world-wide ERP application providers, and maintains a suite of benchmark tests to demonstrate the performance of competitive systems on the various SAP products.

See Also

Disclosure Statement

SAP Enhancement Package 4 for SAP ERP 6.0 (Unicode) Sales and Distribution Benchmark, results as of 6/22/2010. For more details, see http://www.sap.com/benchmark. SD-Parallel, Two Sun Fire X4470 (each 4 processors, 32 cores, 64 threads) 21,000 SAP SD Users, Cert# 2010029. SD 2-Tier, HP ProLiant DL980 G7 (8 processors, 64 cores, 128 threads) 18,180 SAP SD Users, Cert# 2010028. SD 2-Tier, Fujitsu PRIMEQUEST 1800E (8 processors, 64 cores, 128 threads) 16,00o SAP SD Users, Cert# 2010010. SD-Parallel, Four Sun Blade X6270 (each 2 processors, 8 cores, 16 threads) 13,718 SAP SD Users, Cert# 2009041. SD 2-Tier, IBM System x3850 X5 (4 processors, 32 cores, 64 threads) 10,450 SAP SD Users, Cert# 2010012. SD 2-Tier, Fujitsu PRIMERGY RX600 S5 (4 processors, 32 cores, 64 threads) 9,560 SAP SD Users, Cert# 2010017. SD-Parallel, Two Sun Blade X6270 (each 2 processors, 8 cores, 16 threads) 7,220 SAP SD Users, Cert# 2009040. SD-Parallel, Sun Blade X6270 (2 processors, 8 cores, 16 threads) 3,800 SAP SD Users, Cert# 2009039. SD 2-Tier, Sun Fire X4270 (2 processors, 8 cores, 16 threads) 3,800 SAP SD Users, Cert# 2009033.

About

BestPerf is the source of Oracle performance expertise. In this blog, Oracle's Strategic Applications Engineering group explores Oracle's performance results and shares best practices learned from working on Enterprise-wide Applications.

Index Pages
Search

Archives
« August 2016
SunMonTueWedThuFriSat
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
19
20
21
22
23
24
25
26
27
28
29
30
31
   
       
Today