In his new article Jonathan Natkins explains how to use components of Apache Hadoop, including Flume, Hive and Oozie to implement a typical Data management system. He also gives a practical example of such architecture to measure Twitter user’s influence.
A new Apache HCatalog project is a table and storage management layer for Hadoop that enables different data processing tools – Pig, MapReduce, and Hive – to more easily inter-operate data. HCatalog’s presents users with a relational view of data and ensures that users need not worry about where or in what format their data is stored – RCFile format, text files, or sequence files.
This article contains an interview with Dipti Borkar, Director of Product Management at Couchbase, on the challenges, benefits and the process of migrating from RDBMS to NoSQL. 6
In this article, authors Arun Viswanathan and Shruthi Kumar discuss how to implement common aggregation functions on a MongoDB document database using its MapReduce functionality. 7
Approaches to integrating data are changing with emergence of cloud computing. 2
In this article, Boris Lublinsky shows how to extend Hbase - based Lucene implementation with geospatial search support.
Usage of custom Hadoop OutputFormat allows to produce output data in a form most appropriate for other applications. 2
InputFormat class provides a powerful mechanism for tighter control of Maps execution in Map Reduce jobs. In this article authors show how to leverage this mechanism for solving specific problems. 1
In this article authors show how to extend Oozie by introducing custom actions, specific for a given company/line of business. 4
Complete Oozie example, demonstrating language features and their usage in real world examples 2
Matrix presents a white paper on using the open source tool, Hadoop, to implement the MapReduce strategy and a Cloud computing strategy to solve business intelligence problems. 1