By Mike.Hallett-Oracle on Jul 01, 2014
One of the best reasons to start using Hadoop, is to off-load ETL processing away from a potentially higher cost “Data Warehouse staging system” and deploy it onto a platform with a better performance-to-cost ratio for this ETL load.
If you do this, you will likely still want high productivity ETL tools such as Oracle Data Integrator (ODI12c), and if you are handling large volumes of data in a limited batch window, you need fast processing and most importantly high speed loading into the Data Warehouse.
ODI12c on Hadoop gives you this, when combined with the Oracle Big Data Connectors. This works especially well on our engineered systems (Big Data Appliance to Exadata), but is still also the best solution for any ETL work from Hadoop to an Oracle database, even on so called “commodity hardware”.
- To learn more, you might like to check out these articles by Mark Rittman (CTO at partner Rittman Mead) re “End-to-End ODI12c ETL on Oracle Big Data Appliance Pt.1 : Flume to Initial Hive Table” plus the follow on parts 2,3... 4,5... etc.
Mark installed all the software elements directly, but if you need to get going quickly, you may be able to use our downloadable VM ( Demonstration VM “BigDataLite 2.4.1” Available on OTN ... although this is now updated to version 3.0) which works on non-BDA hardware.