InfoQ Homepage Distributed Systems Content on InfoQ

  • Costin Leau on Elasticsearch, BigData and Hadoop

    Elasticsearch is an open source, distributed real-time search and analytics engine for the cloud. The first milestone of elasticsearch-hadoop 1.3.M1 was released last month. InfoQ spoke with Costin Leau about Elasticsearch and how it integrates with Hadoop and other Big Data technologies.

  • Spoilt for Choice – How to choose the right Big Data / Hadoop Platform?

    In his new article Kai Wähner compares several alternatives for installing a version of Hadoop and realizing big data processes. He compares distributions and tooling from Apache and many other vendors including Cloudera, HortonWorks, MapR, Amazon, IBM, Oracle, Microsoft. He additionally describes pros and cons of every distribution and provides a decision tree for choosing a most appropriate one.

  • Interview and Video Review: Working with Big Data: Infrastructure, Algorithms, and Visualizations

    Paul Dix leads a practical exploration into Big Data in this video training series. The first five lessons of the training span multiple server systems with a focus on the end to end processing of large quantities of XML data from real Stack Exchange posts. He completes the training with a lesson on developing visualizations for gaining insights from the macro level analysis of Big Data.

  • Hadoop Virtual Panel

    In this virtual panel, InfoQ talks to several Hadoop vendors and users about their views at current and future state of Hadoop and the things that are the most important for Hadoop’s further adoption and success.

  • Interview with Arun Murthy on Apache YARN

    Apache Hadoop YARN – a new Hadoop resource manager - has just been promoted to a high level Hadoop subproject. InfoQ had the chance to discuss YARN with Arun Murthy - founder and architect at Hortonworks.

  • Generating Avro Schemas from XML Schemas Using JAXB

    Apache Avro is an up and coming binary marshalling framework. In his new article Benjamin Fagin explains how one can leverage existing XSD tooling to create data definitions and then use XJC plugin to directly generate AVRO schemes and marshaling classes.

  • Exploring Hadoop OutputFormat

    As more companies adopt Hadoop, its integration with other applications is becoming more important. A key to such integration is usage of the appropriate OutputFormat allowing to produce output data in a form most appropriate for other applications.

  • Extending Oozie

    In this article authors show how leverage Oozie extensibility to implement custom language extensions. This approach can be viewed a specializing workflow language for a given company/line of business.

  • An Open, Interoperable Cloud

    This article describes how interoperable clouds can be created, today, through the integration of open standards such as the Open Cloud Compute Interface, the Open Virtualisation Format and CDMI. They provide the means to package virtual infrastructure deployments, an API for the runtime management of storage infrastructure and an API for the runtime management of infrastructure as service.

  • Oozie by Example

    End to end Oozie example, including process design, resource coordinator and workflow implementation

  • Introduction to Oozie

    Basic introduction to Oozie - a framework allowing to combine multiple Map/Reduce jobs into a logical unit of work.

  • Using Apache Avro

    Boris Lublinsky presents an introduction to AVRO and evaluate its usage for Schema componentization, inheritance and polymorphism. He also discusses backward compatibility issues and AVRO solutions for this problem.


Is your profile up-to-date? Please take a moment to review and update.

Note: If updating/changing your email, a validation request will be sent

Company name:
Company role:
Company size:
You will be sent an email to validate the new email address. This pop-up will close itself in a few moments.