Brian Femiano

author
+ Follow
since Apr 24, 2013
Merit badge: grant badges
For More
Cows and Likes
Cows
Total received
0
In last 30 days
0
Total given
0
Likes
Total received
1
Received in last 30 days
0
Total given
0
Given in last 30 days
0
Forums and Threads
Scavenger Hunt
expand Ranch Hand Scavenger Hunt
expand Greenhorn Scavenger Hunt

Recent posts by Brian Femiano

Accumulo can satisfy queries that demand fast response times, but the internal operations are not strictly in-memory. All underlying data structures are persisted to Hadoop HDFS.

It's primary purpose is to enable low-latency fetches over persistant columnar data stored in HDFS.
11 years ago
Hi Pawan. Thanks for considering the book.

With regards to your problem, Hadoop Real world solutions is designed to get people up and rolling with Hadoop once they've decided it's a good fit for their system. There is no one particular recipe that covers a full transition from a legacy warehouse to using Hadoop, but many of the recipes independently help with a critical piece. For example, we cover using Sqoop to migrate data from a RDBMS (like Oracle) into Hadoop HDFS. That would be a good place to start.

Hive/Pig are excellent data warehousing/processing tools designed to run over Hadoop, and we cover how to use them in very solid detail. You can follow the sample analytics to get an idea of how you would port your PL/SQL implementations to a language like Pig Latin or HiveQL.

So to answer your question, if you've decided Hadoop is the right platform moving forward, many of the Real-World solutions recipes can be combined to help with the overall data migration and ETL transition.

Hope this helps!
11 years ago