BT

Facilitating the Spread of Knowledge and Innovation in Professional Software Development

Write for InfoQ

Topics

Choose your language

InfoQ Homepage Distributed Systems Content on InfoQ

  • Spoilt for Choice – How to choose the right Big Data / Hadoop Platform?

    In his new article Kai Wähner compares several alternatives for installing a version of Hadoop and realizing big data processes. He compares distributions and tooling from Apache and many other vendors including Cloudera, HortonWorks, MapR, Amazon, IBM, Oracle, Microsoft. He additionally describes pros and cons of every distribution and provides a decision tree for choosing a most appropriate one.

  • Interview and Video Review: Working with Big Data: Infrastructure, Algorithms, and Visualizations

    Paul Dix leads a practical exploration into Big Data in this video training series. The first five lessons of the training span multiple server systems with a focus on the end to end processing of large quantities of XML data from real Stack Exchange posts. He completes the training with a lesson on developing visualizations for gaining insights from the macro level analysis of Big Data.

  • Hadoop Virtual Panel

    In this virtual panel, InfoQ talks to several Hadoop vendors and users about their views at current and future state of Hadoop and the things that are the most important for Hadoop’s further adoption and success.

  • Interview with Arun Murthy on Apache YARN

    Apache Hadoop YARN – a new Hadoop resource manager - has just been promoted to a high level Hadoop subproject. InfoQ had the chance to discuss YARN with Arun Murthy - founder and architect at Hortonworks.

  • Generating Avro Schemas from XML Schemas Using JAXB

    Apache Avro is an up and coming binary marshalling framework. In his new article Benjamin Fagin explains how one can leverage existing XSD tooling to create data definitions and then use XJC plugin to directly generate AVRO schemes and marshaling classes.

  • Exploring Hadoop OutputFormat

    As more companies adopt Hadoop, its integration with other applications is becoming more important. A key to such integration is usage of the appropriate OutputFormat allowing to produce output data in a form most appropriate for other applications.

  • Extending Oozie

    In this article authors show how leverage Oozie extensibility to implement custom language extensions. This approach can be viewed a specializing workflow language for a given company/line of business.

  • An Open, Interoperable Cloud

    This article describes how interoperable clouds can be created, today, through the integration of open standards such as the Open Cloud Compute Interface, the Open Virtualisation Format and CDMI. They provide the means to package virtual infrastructure deployments, an API for the runtime management of storage infrastructure and an API for the runtime management of infrastructure as service.

  • Oozie by Example

    End to end Oozie example, including process design, resource coordinator and workflow implementation

  • Introduction to Oozie

    Basic introduction to Oozie - a framework allowing to combine multiple Map/Reduce jobs into a logical unit of work.

  • Using Apache Avro

    Boris Lublinsky presents an introduction to AVRO and evaluate its usage for Schema componentization, inheritance and polymorphism. He also discusses backward compatibility issues and AVRO solutions for this problem.

  • Data Mining in the Swamp: Taming Unruly Data With Cloud Computing

    Matrix presents a white paper on using the open source tool, Hadoop, to implement the MapReduce strategy and a Cloud computing strategy to solve business intelligence problems.

BT