X

Secure Kafka Cluster

A while ago I've wrote Oracle best practices for building secure Hadoop cluster and you could find details here. In that blog I intentionally didn't mention Kafka's security, because this topic deserved dedicated article. Now it's time to do this and this blog will be devoted by Kafka security only.  Kafka Security challenges 1) Encryption in motion. By default you communicate with Kafka cluster over unsecured network and everyone, who can listen network between your client...

Tuesday, June 19, 2018 | Read More

Big Data SQL 3.2.1 is Now Available

Just wanted to give a quick update.  I am pleased to announce that Oracle Big Data SQL 3.2.1 is now available.   This release provides support for Oracle Database 12.2.0.1.  Here are some key details: Existing customers using Big Data SQL 3.2 do not need to take this update; Oracle Database 12.2.0.1 support is the reason for the update. Big Data SQL 3.2.1 can be used for both Oracle Database 12.1.0.2 and 12.2.0.1 deployments For Oracle Database 12.2.0.1, Big Data SQL 3.2.1...

Thursday, May 31, 2018 | Big Data | Read More

Event Hub Cloud Service. Hello world

In early days, I've wrote a blog about Oracle Reference Architecture and concept of Schema on Read and Schema on Write. Schema on Read is well suitable for Data Lake, which may ingest any data as it is, without any transformation and preserve it for a long period of time.  At the same time you have two types of data - Streaming Data and Batch. Batch could be log files, RDBMS archives. Streaming data could be IoT, Sensors, Golden Gate replication logs. Apache Kafka is...

Saturday, May 19, 2018 | Read More

Autonomous Data Warehouse is LIVE!

That’s right: Autonomous Data Warehouse Cloud is LIVE and available in the Oracle Cloud. ADWC Launch Event at Oracle Conference CenterWe had a major launch event on Thursday last week at the Oracle Conference center in Redwood Shores which got a huge amount of press coverage. Larry Ellison delivered the main keynote covering how our next-generation cloud service is built on the self-driving Oracle Autonomous Database technology which leverages machine learning to deliver...

Thursday, April 5, 2018 | Data Warehousing | Read More

Object Store Service operations. Part 1 - Loading data

One of the most common and clear trends in the IT market is Cloud and one of the most common and clear trends in the Cloud is Object Store. Some introduction information you may find here. Many Cloud providers, including Oracle, assumes, that data lifecycle starts from Object Store: You land it there and then either read or load it by different services, such as ADWC or BDCS, for example. Oracle has two flavors of Object Store Services (OSS), OSS on OCI (Oracle...

Thursday, April 5, 2018 | Read More

Loading Data to the Object Store for Autonomous Data Warehouse Cloud

So you got your first service instance of your autonomous data warehouse set up, you experienced the performance of the environment using the sample data, went through all tutorials and videos and are getting ready to rock-n-roll. But the one thing you’re not sure about is this Object Store. Yes, you used it successfully as described in the tutorial, but what’s next?. And what else is there to know about the Object Store? First and foremost, if you are interested in...

Thursday, April 5, 2018 | Data Warehousing | Read More

Roadmap Update: What you need to know about Big Data Appliance 4.12

As part of our continuous efforts to ensure transparency in release planning and availability to our customers for our big data stack, below is an update to the original roadmap post. Current Release As discussed the 4.11 release delivered the following: Released with updates CDH and now deliver 5.13.1 Updates to the Operating System (OL6) with security updates to said OS Java updates The release consciously pushed back some of the features to ensure the Oracle environments pick up...

Monday, March 12, 2018 | The Data Warehouse Insider | Read More

Big Data SQL Quick Start. Multi-user Authorization - Part 25

One of the major Big Data SQL benefits is security. You deal with the data, which you store in HDFS or other sources, through Oracle Database, which means, that you could apply many Database features, such as Data Redaction, VPD or Database Vault. These features in conjunction with database schema/grant privileged model, allows you to protect data from Database side (when intruder tries to reach data from database side). But it's also important to keep in mind, that Data...

Monday, March 12, 2018 | Big Data SQL | Read More

Demoing OAC Querying ADWC Massive Data Volumes

If you missed all the major announcements from OpenWorld 2017 about Oracle Autonomous Data Warehouse Cloud then take some time to review this blog post Review of Big Data Warehousing at OpenWorld 2017 - Now Available. This great new video below is an excerpt from the Thomas Kurian's Keynote session at Oracle Cloud World earlier this month (February). In this video, George Lumpkin, VP of Product Management, shows how easy it is to create an new data warehouse inside the Oracle...

Thursday, February 22, 2018 | Data Warehousing | Read More

Advanced Data Protection using Big Data SQL and Database Vault - Introduction

According to latest analysts reports and data breaches statistics, Data Protection is rising up to be the most important IT issue in the coming years! Due to increasing threats and cyber-attacks, new privacy regulations are being implemented such as the European Union (EU) General Data Protection Regulation (GDPR), are also being enforced and the increasing adoption of Public Cloud also legitimate these new Cyber Security requirements. Data Lake/Hub environments can be a...

Wednesday, February 7, 2018 | Big Data | Read More

Learn more about using Big Data Manager - importing data, notebooks and other useful things

In one of the previous posts on this blog (See How Easily You Can Copy Data Between Object Store and HDFS) we discussed some functionality enabled by a tool called Big Data Manager, based upon the distributed (Spark based) copy utility. Since then a lot of useful features have been added to Big Data Manager, and to share with the world, these are now recorded and published on YouTube. The library consists of a number of videos with the following topics (video library is here): W...

Monday, January 29, 2018 | Read More

Oracle Big Data Lite 4.11 is Available

The latest release of Oracle Big Data Lite is now available for download on OTN.  Version 4.11 has the following products installed and configured: Oracle Enterprise Linux 6.9 Oracle Database 12c Release 1 Enterprise Edition (12.1.0.2) - including Oracle Big Data SQL-enabled external tables, Oracle Multitenant, Oracle Advanced Analytics, Oracle OLAP, Oracle Partitioning, Oracle Spatial and Graph, and more. Cloudera Distribution including Apache Hadoop (CDH5.13.1) Cloudera Manager...

Tuesday, January 16, 2018 | Big Data | Read More

Free new tutorial: Quickly uploading files with Big Data Manager in Big Data Cloud Service

Sometimes the simplest tasks make life (too) hard. Consider simple things like uploading some new data sets into your Hadoop cluster in the cloud and then getting to work on the thing you really need to do: analyzing that data. This new free tutorial shows you how to easily and quickly do the grunt work with Big Data Manager in Big Data Cloud Service (learn more here) enabling you to worry about analytics, not moving files. The approach taken here is to take a file that resides...

Friday, January 12, 2018 | Big Data | Read More

New Release: BDA 4.11 is now Generally Available

As promised, this update to Oracle Big Data Appliance would come fast. We just uploaded the bits and are in process of uploading both documentation and configurator. You can find the latest software on MyOracleSupport. So what is new: BDA Software 4.11.0 contains few new things, but is intended to keep our software releases close to the Cloudera releases, as discussed in this roadmap post. This latest version uptakes: Cloudera CDH 5.13.1 and Cloudera Manager 5.13.1 Parcels for...

Thursday, December 21, 2017 | Big Data | Read More

SQL Pattern Matching Deep Dive - the book

Those of you with long memories might just be able to recall a whole series of posts I did on SQL pattern matching which were taken from a deep dive presentation that I prepared for the BIWA User Group Conference. The title of each blog post started with SQL Pattern Matching Deep Dive...and covered a set of 6 posts: Part 1 - Overview Part 2 - Using MATCH_NUMBER() and CLASSIFIER() Part 3 - Greedy vs. reluctant quantifiers Part 4 - Empty matches and unmatched rows? Part 5 - SKIP TO...

Monday, November 27, 2017 | Data Warehousing | Read More

Using Materialized Views with Big Data SQL to Accelerate Performance

One of Big Data SQL’s key benefits is that it leverages the great performance capabilities of Oracle Database 12c.  I thought it would be interesting to illustrate an example – and in this case we’ll review a performance optimization that has been around for quite a while and is used at thousands of customers:  Materialized Views (MVs). For those of you who are unfamiliar with MVs – an MV is a precomputed summary table.  There is a defining query that describes that summary. ...

Wednesday, November 22, 2017 | Big Data SQL | Read More

Big Data SQL Quick Start. Big Data SQL over Kafka – Part 23

G Data SQL 3.2 version brings a few interesting features. Among those features, one of the most interesting is the ability to read Kafka. Before drilling down into details, I'd like to explain in the nutshell what Kafka is. What is Kafka? The full scope of the information about Kafka you may find here, but in the nutshell, it's distributed fault tolerant message system. It allows you to connect many systems in an organized fashion. Instead, connect each system peer to peer: you...

Wednesday, November 15, 2017 | Big Data SQL | Read More

Big Data SQL Quick Start. Correlate real-time data with historiacal benchmarks – Part 24

In Big Data SQL 3.2 we have introduced new capability - Kafka as a data source. Some details about how it works with some simple examples, I've posted over here. But now I want to talk about why do you want to run queries over Kafka. Here is Oracle concept picture on Datawarehouse: You have some stream (real-time data), data lake where you land raw information and cleaned Enterprise data. This is just a concept, which could be implemented in many different ways, one of this...

Wednesday, November 15, 2017 | Big Data SQL | Read More

Oracle Big Data SQL 3.2 is Now Available

Big Data SQL 3.2 has been released and is now available for download on edelivery.  This new release has many exciting new features – with a focus on simpler install and configuration, support for new data sources, enhanced security and improved performance. Big Data SQL has expanded its data source support to now include querying data streams – specifically Kafka topics: This enables streaming data to be joined with dimensions and facts in Oracle Database or HDFS.  It’s never...

Wednesday, November 15, 2017 | Read More

Roadmap Update for Big Data Appliance Releases (4.11 and beyond)

With the release of BDA version 4.10 we added a number of interesting features, but for various reasons we did slip behind our targets in up taking the Cloudera updates within a reasonable time. To understand what we do before we ship the latest CDH on BDA and why we think we should spend that time, review this post. That all said, we have decide to rejigger the releases and do the following: Focus BDA 4.11 solely on up taking the latest CDH 5.13.1 and related OS and Java...

Wednesday, November 8, 2017 | Big Data | Read More

Review of Big Data Warehousing at OpenWorld 2017 - Now Available

Did you miss OpenWorld 2017? Then my latest book is definitely something you will want to download! If you went to OpenWorld this book is also for you because it covers all the most important big data warehousing messages and sessions during the five days of OpenWorld.Following on from OpenWorld 2017 I have put together a comprehensive review of all the big data warehousing content from OpenWorld 2017. This includes all the key sessions and announcements from this year's...

Wednesday, October 25, 2017 | Big Data | Read More

New Release: BDA 4.10 is now Generally Available

As of today, BDA version 4.10 is Generally Available. As always, please refer to  If You Struggle With Keeping your BDAs up to date, Then Read This to learn about the innovative release process we do for BDA software. This new release includes a number of features and updates: Support for Migration From Oracle Linux 5 to Oracle Linux 6 - Clusters on Oracle Linux 5 must first be upgraded to v4.10.0 on Oracle Linux 5 and can then be migrated to Oracle Linux 6. This process must...

Monday, October 16, 2017 | Read More

Announcing: Big Data Appliance X7-2 - More Power, More Capacity

Big Data Appliance X7-2 is the 6th hardware generation of Oracle's leading Big Data platform continuing the platform evolution from Hadoop workloads to Big Data, SQL, Analytics and Machine Learning workloads. Big Data Appliance combines dense IO with dense Compute in a single server form factor. The single form factor enables our customers to build a single data lake, rather then replicating data across more specialized lakes.  What is New? The current X7-2 generation is based...

Tuesday, October 10, 2017 | Read More

OpenWorld 2017: Must-See Sessions for Day 4 - Wednesday

Day 4 is here which makes today #Autonomous Wednesday. Included my definitive list of Must-See sessions for today are two of THE most important sessions at this years conference. You will not want to miss these two sessions:The rest of list is, of course, packed full of sessions and labs covering our Big Data Warehouse technologies and features. These sessions are what Oracle OpenWorld is all about: the chance to learn about the latest technology from the real...

Wednesday, October 4, 2017 | Big Data | Read More

OpenWorld 2017: Must-See Sessions for Day 3 - Tuesday

Day 3, Tuesday, is here and this is my definitive list of Must-See sessions for today. Today we are focused on the new features in Oracle Database 18c - multitenant, in-memory, Oracle Text, machine learning, Big Data SQL etc etc. These sessions are what Oracle OpenWorld is all about: the chance to learn about the latest technology from the real technical experts. TUESDAY'S MUST-SEE GUIDE Don't worry if you are not able to join us in San Francisco for this year's conference...

Monday, October 2, 2017 | Big Data | Read More

OpenWorld 2017: Must-See Sessions for Day 2 - Monday

Day 2, Monday, is here and this is my definitive list of Must-See sessions for today. The list is packed full of sessions and labs that follow on from yesterday's (Sunday) big announcements around Oracle Autonomous Database and Oracle Autonomous Data Warehouse Cloud. These sessions are what Oracle OpenWorld is all about: the chance to learn about the latest technology from the real technical experts. MONDAY's MUST-SEE GUIDE Don't worry if you are not able to join us in San...

Monday, October 2, 2017 | Big Data | Read More

OpenWorld 2017 - Must-See Sessions for Day 1

It all starts today -  OpenWorld 2017. Each day I will provide you with a list of must-see sessions and hands-on labs. This is going to be one of the most exciting OpenWorlds ever! Today is Day 1 so here here is my definitive list of Must-See sessions for the opening day. The list is packed full of really excellent speakers such as Franck PachotAmi AharonovichGalo Balda and Rich Niemiec. These sessions are what Oracle OpenWorld is all about: the chance to learn from the...

Saturday, September 30, 2017 | Data Warehousing | Read More

#oow17 BDW Smartphone App Now Live

It’s only 4 days and counting until OpenWorld 2017 starts. If you are coming to this year’s conference then you will definitely want to use our completely free #oow17 online BDW app for smartphones and tablets which is now live: https://keithlaker.github.io/Storyboard.html#LandingPad. The app includes a day-by-day calendar of all the most important sessions covered in the comprehensive Big Data Warehousing Must-See Guide, see here: https://oracle-big-data.blogspot.co.uk/2017/08/...

Wednesday, September 27, 2017 | Data Warehousing | Read More

UPDATED: Big Data Warehousing Must See Guide for Oracle OpenWorld 2017

** NEW ** Chapter 5   *** UPDATED *** Must-See Guide now available as PDF and via Apple iBooks Store This updated version now contains details of all the most important hands-on labs AND a day-by-day calendar. This means that our comprehensive guide now covers absolutely everything you need to know about this year’s Oracle OpenWorld conference. Now, when you arrive at Moscone Conference Center you are ready to get the absolute most out of this amazing conference.The updated, and...

Tuesday, September 12, 2017 | Big Data | Read More

Secure your Hadoop Cluster

Security is a very important aspect of many projects and you must not underestimate it, Hadoop security is very complex and consist of many components, it's better to enable one by one security features. Before starting the explanation of different security options, I'll share some materials that will help you to get familiar with the foundation of algorithms and technologies that underpin many security features in Hadoop. Foundation. First of all, I recommend that you watch...

Tuesday, September 12, 2017 | Read More

How Enabling CDSW Will Help You Make Better Use of Your Big Data Appliance

No one has to elaborate on the interest and importance of Data Science, so we won't go into why you should be looking at frameworks and tools to enable AI/ML and more fun things on your Hadoop infrastructure. One way to do this on Oracle Big Data Appliance is to use Cloudera Data Science Workbench (CDSW). See at the end of this post for some information on CDSW and its benefits. How does it work? Assuming you want to go with CDSW for your data science needs, here is what is...

Tuesday, August 8, 2017 | Big Data | Read More

If You Struggle With Keeping your BDAs up to date, Then Read This

[Updated on October 15th to reflect the release of BDA 4.10, with CDH 5.12.1] One of the interesting aspects of keeping your Oracle Big Data infrastructure up to date (Hadoop, but also the OS and the JDK) is trying to get a hold of the latest information enabling everyone to plan their upgrades and see what is coming. The following is a list of versions released over the past quarters and a look ahead to what is coming. What is the Schedule? The intention is to release a...

Tuesday, August 1, 2017 | Big Data | Read More

MATCH_RECOGNIZE and predicates - everything you need to know

  MATCH_RECOGNIZE and predicates At a recent user conference I had a question about when and how  predicates are applied when using MATCH_RECOGNIZE so that’s the purpose of this blog post. Will this post cover everything you will ever need to know for this topic? Probably! Where to start….the first thing to remember is that the table listed in the FROM clause of your SELECT statement acts as the input into the MATCH_RECOGNIZE pattern matching process and this raises the question...

Wednesday, July 19, 2017 | Data Warehousing | Read More

See How Easily You Can Copy Data Between Object Store and HDFS

Object Stores tend to be a place where people put there data in the cloud (see also The New Data Lake - You Need More Than HDFS). Add data here and then share it, load it or use it across various other services. Here we won't discuss the architecture and whether or not the data lake now is the object store (hint: not yet...), but instead focus on how to easily move data back and forth between object stores and your Big Data Cloud Service (BDCS) cluster(s). ODCP The underlying...

Monday, July 17, 2017 | Big Data | Read More

Big Data Warehousing Must See Guide for Oracle OpenWorld 2017

  It’s here - at last! I have just pushed my usual must-see guide to the Apple iBooks Store.   The free big data warehousing Must-See guide for OpenWorld 2017 is now available for download from the Apple iBooks Store - click here, and yes it’s completely free. This comprehensive guide covers everything you need to know about this year’s Oracle OpenWorld conference so that when you arrive at Moscone Conference Center you are ready to get the most out of this amazing conference....

Wednesday, July 12, 2017 | Big Data | Read More

Big Data SQL Quick Start. Binary Images and Big Data SQL – Part 22

Big Data SQL Quick Start. Binary Images and Big Data SQL – Part 22 Many thanks to Dario Vega, who is the actual author of this content, I'm just publishing it in the Big Data SQL blog. Create a hive table with a binary field and cast to BLOB type in RDBMS when using big data sql For text files, hive is storing in a base64 representation the binary fields. Normally, there is no problem with newline character and not extra work inside the Oracle database the conversion is done...

Wednesday, July 5, 2017 | Big Data SQL | Read More

Big Data SQL Quick Start. Complex Data Types – Part 21

Big Data SQL Quick Start. Complex Data Types – Part 21 Many thanks to Dario Vega, who is the actual author of this content. I'm just publishing it on this blog. A common potentially mistaken approach that people take regarding the integration of NoSQL, Hive and ultimately BigDataSQL is to use only a RDBMS perspective and not an integration point of view. People generally think about all the features and data types they're already familiar with from their experience using one...

Wednesday, July 5, 2017 | Big Data SQL | Read More

Big Data SQL Quick Start. Custom SerDe – Part 20

Big Data SQL Quick Start. Custom SerDe – Part 20 Many thanks to Bilal Ibdah, who is actual author of this content, I'm just publishing it in the Big Data SQL blog. A modernized data warehouse is a data warehouse augmented with insights and data from a Big Data environment, typically Hadoop, now rather than moving and pushing the Hadoop data to a database, companies tend to expose this data through a unified layer that allows access to all data storage platforms, Hadoop, Oracle...

Wednesday, July 5, 2017 | Big Data SQL | Read More

Connecting Apache Zeppelin to your Oracle Data Warehouse

In my last posts I provided an overview of the Apache Zeppelin open source project which is a new style of application called a “notebook”. These notebook applications typically runs within your browser so as an end user there is no desktop software to download and install.Interestingly, I had a very quick response to this article asking about how to setup a connection within Zeppelin to an Oracle Database. Therefore, in this post I am going to look at how you can install...

Tuesday, May 23, 2017 | Data Warehousing | Read More

New Look Blog - Site-Under-Construction

Welcome to our new look blog. We are currently in the process of moving all of our blog posts from the old blogging platform to our completely new Oracle blogging platform.  As the title of this posts suggests, we are having some teething issues....what this all means is that we are currently in "site-under-construction" mode. This new platform offers us a lot of significant improvements over the old blogging software: 1) Posts will display correctly on any size of screen so...

Monday, May 15, 2017 | Read More

The latest in Oracle Partitioning - Part 3: Auto List Partitioning

This is the third blog about new partitioning functionality in Oracle Database 12c Release 2. It's about the new and exciting Auto List Partitioning, an extension to List Partitioning. And yes, it works for both "old" single column list and the new multi column list partitioned tables ... As the name already suggests, something is done 'automatic'. In this . That's it. So conceptually auto-list partitioning is a similar extension to list partitioning than interval partitioning...

Thursday, April 20, 2017 | Functionality | Read More

Using Zeppelin Notebooks with your Oracle Data Warehouse - Part 1

Over the past couple of weeks I have been looking at one of the Apache open source projects called Zeppelin. It’s a new style of application called a “notebook” which typically runs within your browser. The idea behind notebook-style applications like Zeppelin is to deliver an adhoc data-discovery tool - at least that is how I see it being used. Like most notebook-style applications, Zeppelin provides a number of useful data-discovery features such as: a simple way to ingest...

Thursday, April 20, 2017 | Data Warehousing | Read More

The latest in Oracle Partitioning - Part 2: Multi Column List Partitioning

This is the second blog about new partitioning functionality in Oracle Database 12c Release 2, available on-premise for Linux x86-64, Solaris Sparc64, and Solaris x86-64 and for everybody else in the Oracle Cloud . This one will talk about multi column list partitioning, a new partitioning methodology in the family of list partitioning. There will be more for this method, coming in a future blog post (how about that for a teaser?). Just like read only partitions, this...

Monday, March 20, 2017 | Data Warehousing | Read More

The latest in Oracle Partitioning - Part 1: Read Only Partitions

Now that Oracle Database12c Release 2 is available on-premise for  Linux x86-64, Solaris Sparc64, and Solaris x86-64 and on the Oracle Cloud for everybody else - the product we had the pleasure to develop and to play with for quite some time now - it's also time to introduce some of the new functionality in more detail to the broader audience. This blog post will be the first of hopefully many over the course of the next months (time permits) to specifically highlight...

Monday, March 6, 2017 | Functionality | Read More

Database 12c Release 2 available for download

Database 12c Release 2 available for download  Yes, it’s the moment the world has been waiting for: the latest generation of the world’s most popular database, Oracle Database 12c Release 2 (12.2) is now available everywhere - in the Cloud and on-premises. You can download this latest version from the database home page on OTN - click on the Downloads tab. So What’s New in 12.2 for Data Warehousing? This latest release provides some incredible new features for data warehouse and...

Thursday, March 2, 2017 | Data Warehousing | Read More

The first really hidden gem in Oracle Database 12c Release 2: runtime modification of external table parameters

We missed to document some functionality !!! With the next milestone for Oracle Database 12c Release 2 just taking place - the availability on premise for Linux x86-64, Solaris Sparc64, and Solaris x86-64, in addition to the Oracle Cloud - I managed to use this as an excuse to play around with it for a bit .. and found that we somehow missed to document new functionality. Bummer. But still better than the other way around .. ;-) We missed to document the capability to override...

Thursday, March 2, 2017 | Functionality | Read More

Data loading into HDFS - Part3. Streaming data loading

In my previous blogs, I already told about data loading into HDFS. In the first blog, I covered data loading from generic servers to HDFS. The second blog was devoted by offloading data from Oracle RDBMS. Here I want to explain how to load into Hadoop streaming data. Before all, I want to note that I will now explain Oracle Golden Gate for Big Data just because it already has many blogposts. Today I'm going to talk about Flume and Kafka. What is Kafka?  Kafka is distributed...

Friday, February 24, 2017 | Big Data | Read More

Big Data Lite 4.7.0 is now available on OTN!

The latest release of Big Data Lite is now available on OTN!  This 4.7 release contains key components of Oracle's big data platform.  It has demos, tutorials and more.  Listed below are the products/features that are installed: Oracle Enterprise Linux 6.8 Oracle Database 12c Release 1 Enterprise Edition (12.1.0.2) - including Oracle Big Data SQL-enabled external tables, Oracle Multitenant, Oracle Advanced Analytics, Oracle OLAP, OraclePartitioning, Oracle Spatial and Graph,...

Thursday, February 16, 2017 | Big Data | Read More

Big Data SQL Quick Start. Machine Learning and Big Data SQL – Part 19

It's very frequent case when somebody talks about Big Data he or she also wants to know how to apply Machine Learning algorithms over this data sets. Oracle Big Data SQL provides the easy and seamless way to do this. Secret of this in Oracle Advanced Analytics (OAA) option, which has been existing for many years. This is the set of the existing algorithms, together with SQL Developer Data Minner interface, allow easily in drag and drop style create advanced models. OAA...

Saturday, February 11, 2017 | Big Data SQL | Read More

Big Data SQL Quick Start. Oracle Text Integration – Part 18

Today, we’ll focus on the integration of Oracle Text, the Full-Text Indexing capabilities from the Oracle database with documents residing on HDFS.     Oracle Text has been available since years and has evolved to address today’s needs regarding indexing:       150+ document formats supported (PDF, DOCX, XLSX, PPTX, XML, JSON…) Dozens of languages Files can be stored inside the database (SecureFiles), or outside on File System or accessible through URL Advanced search functions:Approxim...

Saturday, February 11, 2017 | Big Data SQL | Read More

How to intelligently aggregate approximations

The growth of low-cost storage platforms has allowed many companies to actively seeking out new external data sets and combine them with internal historical data that goes back over a very long time frame. Therefore, as both the type of data and the volume of data continue to grow the challenge for many businesses is how to process this every expanding pool of data and at the same time, make timely decisions based on all the available data.(Image above courtesy of ...

Tuesday, January 31, 2017 | Functionality | Read More

Dealing with very very long string lists using Database 12.2

Oracle RDBMS 11gR2 introduced the LISTAGG function for working with string values. It can be used to aggregate values from groups of rows and return a concatenated string where the values are typically separated by a comma or semi-colon - you can determine this yourself within the code by supplying your own separator symbol. Based on the number of posts across various forums and blogs, it is widely used by developers. However, there is one key issue that has been highlighted...

Wednesday, January 18, 2017 | Functionality | Read More

Oracle Big Data SQL: Simplifying Information Lifecycle Management

For many years, Oracle Database has provided rich support for Information Lifecycle Management (ILM).  Numerous capabilities are available for data tiering – or storing data in different media based on access requirements and storage cost considerations.  These tiers may scale from in-memory for real time data analysis – to Database Flash for frequently accessed data – to operational data captured in Database Storage and Exadata Cells.   Hadoop offers yet another storage layer...

Tuesday, January 17, 2017 | Big Data | Read More

Are approximate answers the best way to analyze big data

Image courtesy of pixabay.com In my previous post I reviewed some reasons why people seem reluctant to accept approximate results as being correct and useful. The general consensus is that approximate results are wrong which is very strange when you consider how often we interact with approximations as part of our everyday life. Most of the use cases in my first post on this topic covered situations where distinct counts were the primary goal - how many click throughs did an...

Friday, January 13, 2017 | Functionality | Read More

SQL Pattern Matching Deep Dive - Part 6, state machines

The obvious way to start this particular post is to pose a couple of simple questions: what is a state machine and why should you care? In general I would say that you don't need to know about or care about state machines. That's the beauty of using SQL for pattern matching. The MATCH_RECOGNIZE clause encapsulates all the deep technical modelling and processing that has to be performed to run pattern matching on a data set. However, there are times when it is useful, probably...

Thursday, January 12, 2017 | Functionality | Read More

Exploring the interfaces for User Defined Aggregates

image courtesy of wikipedia Whilst I was working on the functional specification for the LISTAGG extensions that we implemented in 12c Release 2, I came across Tom Kyte’s stragg function which uses the User Defined Aggregate API introduced in database 9i. Tom’s comprehensive answer covers the two important areas that need to considered when using the data cartridge API: 1) a UDA can run as serial process and 2) a UDA can run as a parallel process. Therefore, you need to code...

Friday, January 6, 2017 | Functionality | Read More

Simplifying your data validation code with Database 12.2

Image courtesy of pixabay.com Doesn’t matter who much testing you do (well, it actually does but that’s a whole different issue) you can almost guarantee that at some point your beautiful data validation code, that parses data input from a web form or loads data from some external file, will pop up with the error: SQL Error: ORA-01722: invalid number 01722. 00000 - "invalid number"*Cause: The specified number was invalid.*Action: Specify a valid number. Of course, what’s is...

Thursday, January 5, 2017 | Functionality | Read More

Big Data SQL Quick Start. Add SerDe in Big Data SQL classpath. - Part17.

today I'm going to write about how to add custom SerDe in Big Data SQL. SerDe is one of the most powerful features of Hadoop and Big Data SQL in particular. It allows you to read any type of data as structured, you just need to explain how to do parse it.  Let's imagine, that we have JSON file: {"wr_returned_date_sk":38352,"wr_returned_time_sk":46728,"wr_item_sk":561506,"wr_refunded_customer_sk":1131210} {"wr_returned_date_sk":38380,"wr_returned_time_sk":78937,"wr_item_sk":100...

Wednesday, December 7, 2016 | Big Data SQL | Read More

Big Data SQL Quick Start. Big Data SQL over complex data types in Oracle NoSQL. - Part16.

Today I'm going to publish the blogpost, which has been written by Javier De La Torre Medina. Thanks to him for the great research! All below is his article. Oracle Big Data SQL over complex data types in Oracle NoSQL When working with Oracle NoSQL databases, we have the flexibility to choose complex data types like arrays, records and map. In this example we are going to show you how to use Oracle Big Data SQL over these complex data types. Let’s use one Oracle NoSQL table...

Tuesday, November 22, 2016 | Big Data SQL | Read More

Big Data SQL Quick Start. Big Data SQL over nested tables in Oracle NoSQL. - Part15.

Today I'm going to publish the blogpost, which has been written by Javier De La Torre Medina. Thanks to him for the great research! All below is his article.   Big Data SQL over nested tables in Oracle NoSQL. In the Oracle NoSQL database, customers can take advantage of the table model. The table model simplifies application data modeling by leveraging existing customer skills: SQL. The table model is built on top of the distributed key-value structure, inheriting all its...

Tuesday, November 22, 2016 | Big Data SQL | Read More

SQL Pattern Matching Deep Dive - Part 5, SKIP TO where exactly?

So far in this series we looked at how to ensure query consistency, how correctly use predicates, managing sorting, using the built-in measures to help with optimise your code and the impact of different types of quantifiers: SQL Pattern Matching deep dive - Part 1 SQL Pattern Matching Deep Dive - Part 2, using MATCH_NUMBER() and CLASSIFIER() SQL Pattern Matching Deep Dive - Part 3, greedy vs. reluctant quantifiers SQL Pattern Matching Deep Dive - Part 4, Empty matches...

Wednesday, November 9, 2016 | Data Warehousing | Read More

Parameter Changes for Parallel Execution in Oracle Database 12c Release 2

As our new database release, Oracle Database 12c Release 2, is now available on the Exadata Express Cloud Service, the Exadata Cloud Service, and the Database Cloud Service, we can start talking about the new features and changes it brings. In regards to Parallel Execution let me start with the initialization parameter changes in this new release. Obsoleted and desupported parameters The following parameters were deprecated long time ago but were still there prior to Oracle...

Tuesday, November 8, 2016 | Read More

Data Warehousing in the Cloud - Part 3

In my last post I looked at Oracle’s Cloud Services for data warehousing and described how they are based around engineered systems running the industry’s #1 database for data warehousing, fully optimised for data warehousing workloads and providing 100% compatibility with existing workloads. Most importantly, Oracle customers can run their data warehouse services on-premise, in the Cloud or using hybrid Cloud using the same management and business tools. I also looked at how...

Wednesday, November 2, 2016 | Read More

Big Data Lite 4.6 Now Available!

Version 4.6 Oracle Big Data Lite VM v 4.6 is now available on OTN (download here).  The VM is packed with all of the latest capabilities from Oracle's big data platform - helping you get started with the tech stack: Oracle Enterprise Linux 6.8 Oracle Database 12c Release 1 Enterprise Edition (12.1.0.2) - including Oracle Big Data SQL-enabled external tables, Oracle Multitenant, Oracle Advanced Analytics, Oracle OLAP, Oracle Partitioning, Oracle Spatial and Graph, and more. Clouder...

Monday, October 31, 2016 | Big Data | Read More

Data Warehousing in the Cloud - Part 2

In the last blog post (Data Warehousing in the Cloud - Part 1) I examined why you need to start thinking about and planning your move to the cloud: looking forward data warehousing in the cloud is seen as having the greatest potential for driving significant business impact through increased agility, better cost control and faster data integration via co-location. In the last section I outlined the top 3 key benefits of moving your data warehouse to the Oracle cloud: it...

Friday, October 28, 2016 | Read More

Data Warehousing in the Cloud - Part 1

Why is cloud so important? Data warehouses are currently going through two very significant transformations that have the potential to drive significant levels of business innovation: The first area of transformation is the drive to increase overall agility. The vast majority of IT teams are experiencing a rapid increase demand for data. Business teams want access to more and more historical data whilst at the same time, data scientists and business analysts are exploring ways...

Tuesday, October 25, 2016 | Read More

SQL Pattern Matching Deep Dive - Part 4, Empty matches and unmatched rows?

I have been asked a number of times during and after presenting on this topic (SQL Pattern Matching Deep Dive) what is the difference between the various matching options such as EMPTY MATCHES and UNMATCHED ROWS. This is the area that I am going to cover in this particular blog post, which is No 4 in this deep dive series. When determining the type of output you want MATCH_RECOGNIZE to return most developers will opt for one of the following: ONE ROW PER MATCH - each match...

Monday, October 24, 2016 | Data Warehousing | Read More

Big Data SQL Quick Start. Big Data SQL and YARN on the same cluster. - Part14.

today I'm going to explain how to multitenant on the same cluster Big Data SQL and YARN. I think it's a quite common scenario - you may want to store historical data and query it with the Big Data SQL. As well you may want to perform the ETL job within the same cluster. If so, resource management became one of the main requirement for this. In other words, you have to warranty certain performance despite on other jobs. For example: 1) You may need to finish your ETL  as fast...

Tuesday, October 18, 2016 | Big Data SQL | Read More

Big Data SQL Quick Start. My query is running too slow or how to tune Big Data SQL. - Part13.

In my previous posts, I was talking about different features of the Big Data SQL. Everything is clear (I hope), but when you start to run real queries you may have doubts - is it a maximum performance which I could get from this Cluster? In this article, I would like to explain steps which are required for the performance tuning of the Big Data SQL. SQL Monitoring. First of all, the Big Data SQL is the Oracle SQL. You may use to start to debug Oracle SQL performance/other...

Wednesday, October 5, 2016 | Big Data SQL | Read More

The complete review of data warehousing and big data content from Oracle OpenWorld 2016

The COMPLETE REVIEW of OpenWorld covers all the most important sessions and related content from this year's conference, including Oracle's key data warehouse and big technologies: Oracle Database 12c Release 2, Oracle Cloud, engineered systems, partitioning, parallel execution, Oracle Optimizer, analytic SQL, analytic views, in-memory, spatial, graph, data mining, multitenant, Big Data SQL, NoSQL Database and industry data models. The COMPLETEreview covers the...

Tuesday, October 4, 2016 | Read More

Big Data SQL Quick Start. Semi-structured data. - Part12.

In my previous blogpost, I was talking about Schema on Read and Schema on Write advantages and disadvantages. As a conclusion, we found that HDFS could be quite suitable for data in the original format. Very often customers have data in a semi-structure format like XML or JSON. In this post, I will show how to work with it. Use case for storing semi-structure data. One of the most common use case for storing semi-structure data in the HDFS could be desire to store all original...

Friday, September 30, 2016 | Big Data SQL | Read More

Your Chance To Meet the Analytic SQL Development Team at OpenWorld

Wow, it’s only just under two weeks to go until this year’s OpenWorld kicks off on September 18th at Moscone Center in San Francisco. The analytic SQL development team will be available on the demo booth (id ref 1635) in the database area, Moscone South, to help with any technical questions and provide general guidance and using analytic features of 12c Release 2. Obviously we would love to meet you all and we will keep you up-to-date with the latest data warehouse news...

Tuesday, September 6, 2016 | Read More

OpenWorld on your smartphone and tablet

Hopefully, you have seen the comprehensive data warehouse guide for #oow16 in iBook and PDF formats which contains links to the latest data warehouse videos, a calendar for the most important sessions and labs and a section that provides profiles and relevant links for all the most important data warehouse presenters. For this year’s conference I have made all this information available in an HTML app that runs on most smartphones and tablets. The pictures below show the menu...

Tuesday, September 6, 2016 | Read More

Big Data SQL Quick Start. Schema on Read and Schema on Write - Part11.

Schema on Read vs Schema on Write So, when we talking about data loading, usually we do this with a system that could belong on one of two types. One of this is schema on write. With this approach, we have to define columns, data formats and so on. During the reading, every user will observe the same data set. As soon as we performed ETL (transform data in the format that most convenient to some particular system), reading will be pretty fast and overall system performance...

Wednesday, August 31, 2016 | Big Data SQL | Read More

The Complete Guide To Data Warehousing and Big Data at Oracle OpenWorld 2016

The COMPLETE guide for OpenWorld provides a comprehensive day-by-day list of all the most important sessions and hands-on labs for Oracle's data warehouse and big technologies: Oracle Database 12c Release 2, Oracle Cloud, engineered systems, partitioning, parallel execution, Oracle Optimizer, analytic SQL, analytic views, in-memory, spatial, graph, data mining, multitenant, Big Data SQL, NoSQL Database and industry data models. The COMPLETEguide covers the following areas: Ke...

Tuesday, August 30, 2016 | Read More

Your Essential Online Session and Hands-on Lab Calendars for #oow16

It’s almost time for OpenWorld. Only three weeks to go! With so much to see and learn at Oracle OpenWorld we are doing our best to make sure that everyone get the most from this year’s conference. Therefore, to help you get prepared and organized we have created a series of online calendars which list all the must-see data warehousing and big data sessions, labs and key events. Just look at the agenda below - we have packed this year’s schedule with the very best must-see...

Tuesday, August 23, 2016 | Read More

Parallel PL/SQL Functions and Global Temporary Tables... and Wrong Results

Recently I got a question from a customer about a parallel query which produces wrong results. The query involves a global temporary table and a parallel enabled PL/SQL function. Before talking about this specific query I want to briefly show the effect of using PL/SQL functions in a parallel query. PL/SQL functions in parallel queries When you use a PL/SQL function, as a predicate for example, in a parallel query the function is executed by the query coordinator (QC). This can...

Thursday, August 18, 2016 | Read More

Must-See Guide for Data Warehousing and Big Data #OOW16

There's so much to see and learn at Oracle OpenWorld - it provides more educational and networking opportunities than any other conference dedicated to Oracle business and technology users. To help you get the most from this year's event I have prepared a comprehensive guide which lists all the must-see data warehousing and big data sessions and labs along with lots of related content. It is divided into the following chapters: Key Highlights from 2015 List of Data Warehouse...

Friday, August 12, 2016 | Read More

Big Data SQL Quick Start. Storage Indexes - Part10.

Today I'm going to explain very powerfully Big Data SQL feature - Storage Indexes. Before all, I want to note, that name "Storage Index" could mislead you. In fact, it's dynamic structure that automatically built over your data after you scan it. There is no any specific command or something that user have run. There is nothing that you have to maintain (like Btree index), rebuild. You just run your workload and after a while, you may note better performance. Storage Indexes...

Sunday, August 7, 2016 | Big Data SQL | Read More

Updates to Big Data Lite 4.5

There are some updates to Big Data Lite 4.5 that you will want to take advantage of:Updated Big Data SQL configuration enables predicate pushdown to engage - which can have pretty dramatic performance benefits Update to the Big Data Spatial framework (fixes a configuration issue) New samples for the Oracle Shell for Hadoop Loaders To get the latest updates, simply double click on the Refresh Samples icon on the desktop: As an example of the BDS performance improvement (and -...

Friday, June 24, 2016 | Big Data | Read More

Adaptive Distribution Methods in Oracle Database 12c

In my post about common distribution methods in Parallel Execution I talked about a few problematic execution plans that can be generated when the optimizer statistics are stale or non-existent. Oracle Database 12c brings some adaptive execution features that can fix some of those issues at runtime by looking at the actual data rather than statistics. In this post we will look at one of these features which is about adapting the distribution method on the fly during statement...

Thursday, June 2, 2016 | Read More

Big Data SQL Quick Start. NoSQL databases - Part9.

It's not a secret that lately IT people are talking a lot about NoSQL. Some even use it. NoSQL databases could have some advantages over RDBMS (like scalability), but many of NoSQL databases don't have many features that are quite common for RDBMS (like transaction support, maturity for backup and recovery tools). Also, many NoSQL databases are schema-less, which can be an advantage (in sense of application development agility), but it also could be a disadvantage (in sense ...

Wednesday, June 1, 2016 | Big Data SQL | Read More

Big Data SQL Quick Start. Data types - Part8.

Today I'm going to share one of the easiest way to improve overall Big Data SQL performance. Big Data SQL is the complex system, which contains two main pieces - Database and Hadoop. Each system has own datatypes - Oracle RDBMS and Java. Every time when you query data from Oracle RDBMS you do data conversion. Data conversion is very expensive CPU operation.   AVRO, RCFile, ORC and Parquet files. Let's zoom Hadoop server piece and find there few components - Hadoop part (Data...

Thursday, May 19, 2016 | Big Data SQL | Read More

Single PX Server Execution

I recently helped a customer tune a few queries in their Oracle Database In Memory POC. I want to talk about a simplified version of one of their queries as it is a nice tuning example and also a good opportunity to talk about a new parallel execution feature introduced in Oracle Database 12c. Let me start with what the feature is and then look at the performance problem this particular customer was having. Single PX server execution SQL constructs like rownum can introduce...

Tuesday, May 17, 2016 | Read More

New pattern matching tutorial on LiveSQL

If you always wanted to try our new SQL pattern matching feature, MATCH_RECOGNIZE, but never had access to a Database 12c instance then you really need to checkout the our great new LiveSQL playground environment. LiveSQL is a great place to learn about all the new features of Database 12c along with all the existing features from earlier releases. The new tutorial is called “Log file sessionization analysis with MATCH_RECOGNIZE” and you can view by clicking here. This...

Friday, May 13, 2016 | Data Warehousing | Read More

Big Data SQL Quick Start. Partition Pruning - Part7.

partitioning is the very common technique in data warehousing and all kind of databases. I assume that reader know what partitioning is and I will not explain theoretical part. If you want you could consider Oracle RDBMS example. But I directly start with the practical. Hive partitioning. Hive originally was created as an easy way to write MapReduces over  HDFS. HDFS is a file system, which has Linux like structure. So, it's easy to assume that partition, in this case, is just...

Monday, May 2, 2016 | Big Data SQL | Read More

Big Data SQL Quick Start. Predicate Push Down - Part6.

Since 3.0 version Big Data SQL has awesome feature - Predicate Push Down (PPD). It works over ORC files, Parquet Files and few NoSQL Databases (like HBase). I plan to devote separate topic for NoSQL databases and Big Data SQL, today i'm going to explain ORC and Parquet file case. For both files there is using the same approach and I'll explain ORC case (Parquet is the same). If you are not familiar with those filetypes I advice you to read this documentation about Parquet...

Friday, April 29, 2016 | Big Data SQL | Read More

SQL Pattern Matching Deep Dive - Part 3, greedy vs. reluctant quantifiers

Picture courtesy of Pixabay Welcome to the third post in this deep-dive series on SQL pattern matching using the MATCH_RECOGNIZE feature that is part of Database 12c.  In the first part of this series we looked at a wide range of topics including ensuring query consistency, how to correctly use predicates and how to manage sorting. In the second part we looked at using the built-in measures to understand how a data set is matched to a pattern.  In this post I am going to review...

Monday, April 25, 2016 | Data Warehousing | Read More

Common Distribution Methods in Parallel Execution

Parallel execution uses the producer/consumer model when executing a SQL statement. The execution plan is divided up into DFOs, each DFO is executed by a PX server set. Data is sent from one PX server set (producer) to another PX server set (consumer) using different types of distribution methods. In this post we will look at the most common distribution methods used in joins in Oracle Database 11gR2. The distribution method is chosen by the optimizer and depends on the DOP,...

Thursday, April 21, 2016 | Read More

Big Data SQL Quick Start. Joins. Bloom Filter and other features - Part5.

Big Data SQL has perfect wonderful heritage from the Oracle RDBMS. One of those representative is Bloom Filter. This feature was available in Oracle 10g and Big Data SQL use it for improve Join performance. Before show you couple concrete examples, let me remind (explain) what Bloom Filter is. Bloom Filter abstract example. Input. Independently of any technology speaking, Bloom Filter is data structure that could answer you to one simple question: "Does element X exist in array...

Tuesday, April 19, 2016 | Big Data SQL | Read More

Is an approximate answer just plain wrong?

We are starting to see a significant change in the way we analyze data as a result of the growth of interest in big data and the newer concept of Internet of Things. Ever since databases were first created everyone has been obsessed, quite rightly so, with ensuring queries returned the correct answer - i.e. precise, accurate answers. This key requirement is derived from the need to run operational, transactional applications. If we check our bank balance online we want the...

Friday, April 15, 2016 | Data Warehousing | Read More

Big Data SQL Quick Start. Security - Part4.

Big Data SQL is the way to access data that stored in HDFS through Oracle RDBMS, using Oracle external table mechanism. In a context of security "table" is the key word, which means that you may apply standard security approaches to those tables. Today I want to give you couple examples with: - Oracle Virtual Private Database (VPD) and - Oracle Data Redaction features.  Oracle Data Redaction I don't want to replace documentation of the Oracle Data Redaction within this blog, all...

Friday, April 15, 2016 | Big Data SQL | Read More

Oracle OpenWorld 2016 call for papers is OPEN!

Yes it's that time of year again! If you have a story to tell about data warehousing, big data and SQL analytics then we want to hear from you because the OpenWorld 2016 call for presentations is now open. Mark your calendars: this year Oracle OpenWorld conference will be held in September on 18 - 22, 2016 at the Moscone Center in San Francisco. We are looking for proposals that describe insights and improvements that attendees can put to use in their own jobs: exciting...

Wednesday, April 13, 2016 | Data Warehousing | Read More

SQL Pattern Matching Deep Dive - Part 2, using MATCH_NUMBER() and CLASSIFIER()

Welcome to the second post in this deep dive series on SQL pattern matching using the new MATCH_RECOGNIZE feature that is part of Database 12c. In the first part of this series we looked at the areas of ensuring query consistency, how to correctly use predicates and how to manage sorting. In this post I am going to review the two built-in measures that we have provided to help you understand how your data set is mapped to the pattern that you have defined. This post will...

Tuesday, April 12, 2016 | Data Warehousing | Read More

In-Memory Parallel Query

In an earlier post we looked at how you can see if a query uses Oracle Database In-Memory (DBIM) or In-Memory Parallel Query (IMPQ). In this post let's look at what IMPQ is and how it works in 12c. What is IMPQ? Before 11gR2 Parallel Execution (PX) in the Oracle Database used direct reads to scan large tables and bypassed the buffer cache. As memory sizes got bigger and SGA sizes got bigger as a result, IMPQ was introduced in 11gR2 to make use of the SGA for data...

Thursday, March 31, 2016 | Read More

Data loading into HDFS - Part2. Data movement from the Oracle Database to the HDFS

Some time ago I started to explain how to move data to the Hadoop Distributed File  System (HDFS) from different sources. In my first blogpost about this I told about batch data loading from generic Linux (or even Unix) servers, Today I’m going to explain some best practices about data movement (offloading) from the Oracle Database to the HDFS in batch mode. Generally speaking there are two major ways: Sqoop and Copy2Hadoop. You may also think about Oracle Table Access for...

Wednesday, March 23, 2016 | Data Warehousing | Read More

SQL Pattern Matching deep dive - Part 1

There has been quite a lot of interest in the new 12c MATCH_RECOGNIZE feature for SQL pattern matching. Therefore, I thought now would be the perfect time to start a series of quick deep dive posts that explain how SQL pattern matching works. Over the coming weeks I will cover the following topics in a series of posts. This is the start of a series of posts based on a presentation that I put together for the recent annual BIWA conference at Oracle HQ. The Oracle BI, DW and...

Monday, March 21, 2016 | Data Warehousing | Read More

Big Data SQL 3.0 is now available!

Oracle Big Data SQL 3.0 is now available!  This is an exciting milestone for Oracle.  With support for Cloudera CDH (both on Big Data Appliance and non-Big Data Appliance), Hortonworks HDP and Oracle Database 12c (both Exadata and non-Exadata) - the benefits derived from unified queries across relational, Hadoop and NoSQL stores can now be achieved across a wide breadth of big data deployments. Hadoop and NoSQL are rapidly becoming key components of today's data management...

Thursday, March 17, 2016 | Data Warehousing | Read More

Maximum Availability Architecture for Big Data Appliance

Oracle Maximum Availability Architecture (MAA) is Oracle's best practices blueprint based on proven Oracle high availability technologies, along with expert recommendations and customer experiences. MAA best practices have been highly integrated into the design and operational capability of Oracle Big Data Appliance, and together they provide the most comprehensive highly available solution for Big Data. Oracle MAA papers are published at the MAA home page of the Oracle...

Wednesday, March 16, 2016 | Data Warehousing | Read More

Big Data SQL Quick Start. Parallel Query - Part3.

Definitely, not all queries could be offloaded and not all part of the SQL could be executed on the cell (storage) side. Some of the queries could be pretty complex and could have PL/SQL inside. For increase performance those type of queries you could definitely use database parallel query and good news that you could use it together with Big Data SQL. And second good news is that for end user or application there is no difference for that. Let’s consider complex query...

Wednesday, March 16, 2016 | Big Data SQL | Read More