Daniel Woods discusses Enterprise Operations which is seeing a shift in the management paradigm of infrastructure.
Mike Keane presents how Conversant migrated to Flume, managing 1000 agents across 4 data centers, processing over 50B log lines per day with peak hourly averages of over 1.5 million log lines/sec.
Lisa Van Gelder provides simple tips and tricks for improving delivery without investing lots of time up front creating complex deployment frameworks.
Melody Meckfessel explores how Google's engineering teams use CD to build products and scale them, and how their strain of DevOps speeds launches and helps their engineering culture thrive.
Jayesh Thakrar shows what can be done with irb, how to exploit JRuby-Java integration, and demonstrates how the Shell can be used in Hadoop streaming to perform complex and large volume batch jobs.
Janne Valkealahti shows how Spring provides a simple programming model to develop applications that can easily be tested and deployed as either a YARN application or a traditional application.
Steve Hoffman, Ken Dallmeyer share their experience integrating Hadoop into the existing environment at Orbitz, creating a reusable data pipeline, ingesting, transporting, consuming and storing data.
Mitchell Hashimoto introduces Vagrant, Packer, Consul, Serf, explaining how they can help DevOps streamline the entire process from development through to production.
Cornelia Davis takes the Spring Trader application and makes (a few) modifications (mostly to config) to get it running on the same components in the cloud, specifically on the Cloud Foundry PaaS.
Barry Jaspan describes how they test Acquia Cloud, a large PaaS and DevOps project, and what they have learned over several years of developing those tests.
Damien Dallimore demonstrates some of the ways in which Splunk software can be used to collect and correlate data throughout the various stages of the lifecycle of your code.