dimanche sept. 29, 2013

#OOW2013: Internet of Things... and Big Data

As promised in my first entry few weeks ago, in preparing Oracle OpenWorld, I am coming back to IoT: Internet of Things... and Big Data. As this was the closing topic develop by Edward Screven, Chris Baker and Deutche Telekom, Dr. Thomas Kiessling. Of course, Big Data and Internet of Things (or M2M - Machine2Machine) have been topics not only covered the last day, but all along the conference, including in JavaOne, with 2 interesting sessions from Gemalto. Gemalto even developed a kit to test your own use cases for M2M. Internet of Things is opening new opportunities but also challenges to overcome to get it right, that at Oracle we classify in 3 categories : Acquire & Transmit, Integrate & Secure, and Analyze & Act.

Acquire & Transmit

Just think of potentially billions of devices that you need to remotely deploy, maintain, update, insure proper transmission of data (the right data at the right time - as your power budget is constrain) and even extend decision making closer to the source. With standards-based Java platform optimized for devices, we are already covering today all those requirements, and are already involved in major Internet of Things projects, like the Smart Grids or Connected Cars.

Integrate & Secure

Of course integrating -securely- all the pieces together is key. As you want it 1) to reliably work with potentially a very large amount of devices and 2) not be compromised by any means. Here again, at the device level, Java is providing the intrinsic security functions that you need, from secure code loading, verification, and execution, confidentiality of data handling, storage, and communication, up to authentication of entities involved in secure operations. And we are driving this secured integration up to the Datacenter, thanks to our comprehensive Identity and Access Management system, up to Data masking, fraud detection, and built-in network security and encryption.

Analyze & Act

Last but not least, is to analyze and correlate those Data and take appropriate actions. This is where M2M and Internet of Things link to Big Data. There are different things that characterize "Big Data" : Volume, Velocity (time & speed), Variety (data format), Value (what is really interesting in those data related to my business), Vizualization (how do I find something in this, of value ?), Veracity (insure that what I will add into my trusted data (DWH...) coming from those new sources is something validated. In M2M, we don't always have Volume, but we still have the other "Vs" to take care. To handle all this IoT generated information  inside the Datacenter, and do correlation with existing Data relevant to your customer business (being ERP, Supply Chain, quality tracking of supplier, improving purchasing process, etc...), you may need need tools. That's why Oracle developed the Oracle Big Data Appliance to build an "HPC for Data" grid including Hadoop & NoSQL to capture those IoT data, and Oracle Exalytics/Oracle Endeca Information Discovery, to enable the vizualisation/discovery phase. Once you pass the discovery phase we can act automatically in real time ! on the specific triggers that you will have identified, thanks to Oracle Event Processing solution.

Deliver

As you see, Oracle Internet of Things platform enables you to quickly develop and deliver, securely, an end-to-end solution.

The end result is a quick time-to-market for an M2M project like the one presented on stage and used live during the conference. This project was develop in 4 weeks, with 6 individuals ! The goal was to control the room capacity and in/out doors live control depending on the participants flow in the room. And as you can see in the architecture diagram we are effectively covering from Java on the device up to Exalytics in the Datacenter.

mercredi sept. 25, 2013

#OOW2013: Jump into the Cloud...

Today we went into the Cloud, with 3 major announcements delivered by Thomas Kurian: a full Database as a Service, a full Java as a Service and a full Infrastructure as a Service in Oracle Cloud, guaranteed, backup and operated by Oracle, depending on different level of services.

Database as a Service

You will be able to provision inside Oracle Cloud a full Oracle Database (12c or 11g) either in single node or in highly available RAC cluster. This Database will be accessible in full SQL*NET, with Root access. This service will be offer in 3 different models :

Basic Service: pre-configured, automatically installed Database Software, managed by you through Enterprise Manager Express.

Managed Service: Oracle Databases managed by Oracle, including :

  • Quarterly Patching and Upgrades with SLA
  • Automated Backup and Point-in-Time Recovery
  • Elastic Compute and Storage

Maximum Availability Service: Oracle manages an highly available Database, including:

  • Real Application Cluster (RAC)
  • Data Guard for Maximum Availability
  • More flexible upgrade schedule

Of course you will be able to move your Data or even you entire Database between your Enterprise Datacenter and Oracle Cloud by leveraging regular tools like SQL loader or Data Pump for example.

Java as a Service

In the same model as the Database as a Service, you will be able to deploy dedicated Weblogic cluster(s) on our Compute Service. Full WLST, JMX and Root access will be provided as well. The 3 different models of services will be the following:

Basic Service: pre-configured, automatically installed weblogic software, with a single node Weblogic Suite (12c or 11g), managed by you using Enterprise Manager.

Managed Service: Oracle manages one or more Weblogic domains in the same way as the Database as a Service's Managed Service.

Maximum Availability Service: Oracle Manage an Highly Available environment, with the following characteristics :

  • Weblogic cluster integrated with RAC
  • Automated Disaster Recovery and Failover
  • More flexible upgrade schedules
  • Additional staging environment

So now let's have a quick look at the constituents of our Infrastructure as a Service layer.

Infrastructure as a Service

Compute Service: will provide an elastic compute capacity in Oracle Cloud, based on 3 different type of requirements : Standard, Compute Intensive or Memory Intensive. The management will be based on REST API, and providing as well Root VM access. This Compute Service will provide network isolation and elastic IP addresses. And of course, it will be highly available.

Storage Service: will store and manage digital content. The management will be through Java and REST API (OpenStack Swift). It has been designed for performance, scalability and availability.

 All those new or enhanced services, are complementing all the Oracle Software as a Services already existing and adopted with success by many of our customers, like was shown in many testimonies during Thomas Key Notes. This provides a Platform for our partners who are leveraging our technologies to build their own services in Oracle Cloud. That's why we also created an Oracle Cloud Market Place, enabling the delivery of our partners applications, as well as their combination/integration tailor to your specific needs directly in Oracle Cloud.

Let's Jump into the Cloud....

lundi sept. 23, 2013

#OOW2013: All your Database in-memory for All your existing applications... on Big Memory Machines

Many announcements have been made today by Larry Ellison, during his opening of Oracle OpenWorld. To begin with, Americas Cup is still running, as Oracle won today's races.  I must admit that seeing those boats racing at such a speed and crossing each other at few meters was really impressive. On OpenWorld side, it was also very impressive. More people this year are attending the event : 60 000 ! And in terms of big numbers, we saw very impressive results of the new features and products that have been announced today by Larry: Database 12c in-memory option, M6-32 Big Memory Machine, M6-32 SuperCluster and Oracle Database Backup, Logging, Recovery Appliance (yes, I am not joking, that's its real product name !).

Database 12c in-memory option: both row and column in-memory formats for same data/table

This new option will benefit all your existing applications unchanged. We are leveraging the memory to store both formats at the same time. This enable us to drop all the indexes that are usually necessary to process queries, for a design target of x100 improvement on performance for real-time analytic. As you will see later, we can achieve even more, especially if we are running on an M6-32 Big Memory Machine. At the same time the goal was also to improve transactions x2 !

The nice thing of this option is that it will benefit to all your existing applications running on top of Oracle Database 12c: no change required.

On stage, Juan Loaiza, did a small demonstration of this new option on a 3 billions row database, representing wikipedia research query. On a regular database, without this option, after identifying (or guessing) the query that will most likely be used by users, you put in place appropriate indexes (from 10 to 20 indexes), then you can run you query with acceptable performance, in this case: 2005 Million Rows Scanned / Sec instead of  5 Million Rows Scanned / Sec. Not too bad... Now if we replace the indexes required by the new Column formats store in-memory, we achieved in this case: 7151 Million Rows Scanned / Sec ! Something people looking into Big Data, and real-time decisions, will surely have a look at it.

 The second announcement was a new processor, and a new system associated with it: the M6 chip and the M6-32 Big Memory Machine... available now !

M6-32 Big Memory Machine: Terabyte Scale Computing

This system is compatible with the previous generation of M5 chips, protecting existing investment, and can host as well the new M6 12x cores, 96 threads processor. All in this system is about Terabytes : up to 32 TB of memory, 3 Terabytes/sec of system bandwidth, 1.4 Terabytes/sec of memory bandwidth, 1 Terabyte per second of I/O bandwidth !

This new machine is also the compute node of the new M6-32 SuperCluster announced also today.

M6-32 SuperCluster: In-Memory Database & Application System

That's our fastest Database Machine, with big memory for Column store and integrated Exadata Storage ! Juan Loaiza did also the same demonstration of the wikipedia search on this system... but not with 3 billions rows, but 218 billions rows ! The result speaks by itself: 341 072 Million Rows Scanned / Sec !

With all those critical systems hosting such amount of Data, it is also very important to provide a powerful Database Backup and Restore Solution... And that's all the latest Appliance announced today is about.

Oracle Database Backup, Logging, Recovery Appliance

By just reading its name you get nearly all the capabilities this new appliance will provide to you. First, it is specialized to backup Oracle Database of ALL your systems running an Oracle Database (Engineered Systems, like the lastest M6-32 SuperCluster or Exadata, as well as your regular servers). Second, it also captures all your Database logs. So not only do you have a backup but also the deltas between now and your latest backup. This is allowing you to come back at the point you want when recovering your database.

It can even be coupled with our new Database Backup service on Oracle Public Cloud, for an extra secure copy.

With this new appliance you can now be confident in securing your Oracle Database data.

Building your future datacenter

Today, not only did we see the new Oracle Database 12c enabling to work on memory for all you application, we also saw the associated M6-32 server and associated Engineered Systems M6-32 SuperCluster to run the stack with Big Memory capacity... all being secured by Oracle Database backup, Logging, Recovery Appliance. All of those innovations contributing to build your Datacenter of the future, where all is engineered to work together at the factory.

dimanche sept. 22, 2013

GRTgaz new Information System on Oracle SuperCluster

This testimony from Mr Sébastien Flourac, Head of Strategy for GRTgaz IT, concluded the last week SPARC Showcase event. Mr Flourac highlighted why he selected an Oracle SuperCluster, Engineered Systems, over a more traditional build it yourself approach, that he also studied.

Due to EEC regulation, GRTgaz a subsidary of GDF-Suez, has to be externalized including of course all its applications and existing IT in less than 2 years. But, of course, the current platforms are shared with other GDF-Suez services, which means for GRT gaz, that they have to build entirely a new platform to migrate their existing application with the lowest associated risks. As a major part of the technologies supporting GRT gaz applications was running on Oracle Database and Oracle Weblogic, either on IBM AIX or SPARC Solaris, GRT gaz had a closer look on what Oracle has to propose to simplify Oracle software on Oracle Hardware, compatible with the existing GRT gaz environment. And it became obvious to Mr Flourac that Oracle SuperCluster was the best fit for his project and for the future for several reasons.

Simplicity and lower cost

With Oracle Engineered Systems, all the complexity and cost of traditional build it yourself solutions has been taken care of at Oracle Engineering level. All the configurations and setup have been defined and integrated at all levels (software, virtualization and hardware) to offer the best SLA (performance and availability). This was concurring to simplify its externalization project, and was also bringing additional benefits on the storage layer for the future.

It was the best financial scenario in their project context.

Lower risks

Not only does the SuperCluster offer the best SLA by design. It also provides a very important feature in this complex applications migration : a full compatibility to run existing Oracle software versions. This was very important for Mr Flourac to avoid in his project to do both : migrate and upgrade.
It is also providing :

  • an integrated stack of Oracle Software and Hardware
  • an upgrade process tested by Oracle
  • a better support of the entire stack

Build for the future

Oracle SuperCluster provides to GRT gaz a consolidated, homogeneous and extremely scalable platform, which not only enable this externalization project but also will be able to host the new business requests.

With this new platform in place, Mr Flourac already knows that in the next phases he will be able to leverage additional integrated and unique features that running Oracle Softwares on Oracle SuperCluster provides:

  • Exadata integration and acceleration for Oracle Database starting with 11gR2
  • Exalogic integration and acceleration for Oracle Weblogic starting with 10.3.4

Of course the SuperCluster is a key enabler, but such a project requires also a team to manage the migration, the transition and the run. This is done through the support of Oracle ACS (transition), Fujitsu (migration) and Euriware (run).


T5-2 for high-performance financial trading

On this post, I will focus on the second testimony reported by François Napoleoni. Here,the goal was to select the next best platform to face the growth and upgrade of a financial trading application. The comparaison has been made between :
  1. T5-2, running Oracle VM, Solaris, Sybase and the financial application
  2. x86, running VMWare, Redhat, Sybase and the financial application

The decision criteria being :

  • Simplified architecture
  • Systems performance in real life (what has been tested and measured)
  • Platform stability
  • Single point of support
  • Leverage internal skills
  • Strong security enforced between the different virtual environments
  • ROI of the solution

For those of you understanding french, I will let you listen to this few minutes video below. And for the English readers, go through more details in this post.

Oracle VM virtualization on T4-4 : architecture and service catalog

Last Tuesday, during the SPARC Showcase event, Jean-François Charpentier, François Napoleoni and Sébastien Flourac delivered very interesting uses cases of deployment of latest Oracle Systems at work : Oracle VM virtualization project on T4-4, intensive financial trading on T5-2 and a complete Information System migration to Oracle SuperCluster.

I will start to cover in this post the main points that Jean-François Charpentier focus on to build a consolidated T4-4 platform, effective for his Business users.

Oracle VM virtualization on T4-4 Architecture

As often, Mr Charpentier had to handle existing environment, that needed to be taken into account when building this new virtual platform. First he had to be able to provide a platform that could consolidate all the existing environments, the main driver here being :

  • total memory requirement of existing asset
  • multi-tenant capability to share the platform securely between several different networks
  • comply to strong SLA

T4-4 was the best building block to cover them :

  • memory : up to 2 TB
  • multiple networks connection : up to 16x PCIe extension
  • Oracle VM built-in with redundant channels capability and live migration
  • Solaris binary compatibility to enable easy consolidation of existing environments

Overall Architecture Design

The deployment choice have been to setup 2x Oracle VM T4-4 clusters per sites as follow:


To cover his SLA requirements, Mr Charpentier built redundancy not only by providing multiple T4-4 nodes per Oracle VM clusters, but also at the Oracle VM itselfs. For Oracle VM, he chose to make redundant storage and network virtual access layer as display in the following 2 diagrams.

Oracle VM Virtual Mutlipathing with alternate I/O Domain


Oracle VM Virtual Network Access through IPMP


 All of this virtual network layer being link to different network back-bones, thanks to the 16x PCIe extension of the T4-4, as illustrated in the following diagram.


Another option could have been to deploy Oracle Virtual Network, to enable the disk and network access with only 2x PCIe slots at the server layer.

Oracle VM on T4-4 Service Catalog

Beside the architecture choice, that needs to comply with strong SLA. The development of a service catalogue is also very key to bring the IT toward a service provider. And it is exactly what have been put in place by Jean-François Charpentier, as follow :

 By putting in place this new virtual platform with its associated service catalog, Mr Charpentier was able to provide to his Business better agility thanks to easier and faster deployment. This platform has become the standard for all Solaris deployment for his business unit, and they expect to reach a 90% to 100% Solaris virtualization by 2014.

lundi sept. 09, 2013

SPARC Showcase in Paris : September 17th

I recently posted about preparing Oracle Open World. A good start, for those of you in Paris, on September 17th, would be to come to the SPARC Showcase, where our customers will develop why and where they are leveraging latest SPARC technologies T4, T5 and Oracle SuperCluster to their IT and Business benefits : 

  • Mr Jean-Marc Jacquot, from Mysis
  • Mr Jean-François Charpentier, Technical Architect from a leader in RH services solutions
  • Mr Sébastien Flourac, Head of Strategy for GRTgaz IT

You can register here.

jeudi sept. 05, 2013

Preparing for #OOW: DB12c, M6, In-memory, Clouds, Big Data... and IoT

It’s always difficult to fit the upcoming Oracle Open World topics, and all its sessions in one title, even if "Simplifying IT. Enabling Business Transformation." makes it clear on what Oracle is focusing on, I wanted to be more specific on the "How". At least for those of you who attended Hot Chips conference, some of the acronyms will be familiar to you, some may not (I will come back to "IoT" later). For those of you attending, or those of you who will get the sessions presentations once available online, here are few things that you don't want to miss which will give you not only what Oracle R&D has done for you since last year, but also what customers -like you- have implemented thanks to the red-stack and its partners, being ISVs or SIs.

First, don't miss Oracle Executives Key notes, second, have a look into the general sessions delivered by VPs of Engineering to get a more in-deep direction, and last but not least, network with your peers, being on specifics deep-dive sessions, experience sharing or even the demo ground where you will be able to get the technologies in action with the Oracle developers subject matters experts.You will find hereafter a small selection.

Oracle Strategy and roadmaps

Industry Focus

Projects implementation feedbacks & lessons learn

Deep-dive with the Experts

Learn how to do it yourself (in 1 hour): Hands-on-Labs

Watch the technologies at work : Demos Ground

This digest is an extract of the many valuable sessions you will be able to attend to accelerate your projects and IT evolution.

About

Eric Bezille

Search

Archives
« septembre 2013 »
lun.mar.mer.jeu.ven.sam.dim.
      
1
2
3
4
6
7
8
10
11
12
13
14
15
16
17
18
19
20
21
24
26
27
28
30
      
Today