Addison-Wesley Professional Devops in AWS LiveLessons is a 4+ hours video course for infrastructure developers and Sys Ops engineers who aim at creating a fully-automated continuous delivery system in AWS. The course is focused on leveraging key features of AWS while also presenting them in the framework of DevOps best practices and tools. InfoQ has spoken with course author Paul Duvall.
Apache Spark is an open source big data processing framework built around speed, ease of use, and sophisticated analytics. In this article, Srini Penchikala talks about how Apache Spark framework helps with big data processing and analytics with its standard API. He also discusses how Spark compares with traditional MapReduce implementation like Apache Hadoop.
Setting up a new monitoring system might seem daunting at first. Franklin guides us through the first steps and explains the architecture and inner workings of a Graphite-based monitoring system. Key takeaways are understanding time series data and configuration, datapoint formats, aggregation methods and retention.
Delivering the right products fast can be challenging, certainly when there are many unknowns along the way. If you want to build products fast you need to be able to learn fast and efficiently.
An introduction to Puppet, an infrastructure configuration management product. In this article Susannah Axelrod, gives an overview of both Puppet, the language, and Puppet, the platform.
The new book, The Practice of Cloud System Administration: Designing and Operating Large Distributed Systems, looks at a wide range of considerations for cloud-scale systems.
In this article Monica Beckwith, starting from core Hadoop components, investigates the design of a highly available, fault tolerant Hadoop cluster, adding security and data-level isolation.
GridGain announced that the In-Memory Data Fabric has been accepted into Apache Incubator program as Apache Ignite. InfoQ spoke with Nikita Ivanov about their product becoming part of Apache.
Kubernetes is an open source project to manage a cluster of Linux containers as a single system, managing and running Docker containers across multiple hosts. 4
Application Lifecycle Management has traditionally been difficult for databases. Ben Rees, explains why the road ahead is now clear for Database Lifecycle Management.
“Infrastructure as Code” is a tenet of the DevOps community. We’ll make a brief introdution to Chef, a well-known IT automation tool, and use it to illustrate the state of the art.
The new “Hadoop in Practice. 2 Edition" book by Alex Holmes covers a lot of topics building Hadoop code and organizing data to support code simplicity and execution speed.