Sam McAfee outlines an approach they are using at Neo to help their clients integrate business metrics and probabilistic modeling into their prioritization process.
Sangeeta Narayanan goes over how Netfix got to the current continuous delivery state, the lessons they learnt and the successes they enjoyed along the way.
Daniel Woods discusses Enterprise Operations which is seeing a shift in the management paradigm of infrastructure.
Mike Keane presents how Conversant migrated to Flume, managing 1000 agents across 4 data centers, processing over 50B log lines per day with peak hourly averages of over 1.5 million log lines/sec.
Lisa Van Gelder provides simple tips and tricks for improving delivery without investing lots of time up front creating complex deployment frameworks.
Melody Meckfessel explores how Google's engineering teams use CD to build products and scale them, and how their strain of DevOps speeds launches and helps their engineering culture thrive.
Jayesh Thakrar shows what can be done with irb, how to exploit JRuby-Java integration, and demonstrates how the Shell can be used in Hadoop streaming to perform complex and large volume batch jobs.
Janne Valkealahti shows how Spring provides a simple programming model to develop applications that can easily be tested and deployed as either a YARN application or a traditional application.
Steve Hoffman, Ken Dallmeyer share their experience integrating Hadoop into the existing environment at Orbitz, creating a reusable data pipeline, ingesting, transporting, consuming and storing data.
Mitchell Hashimoto introduces Vagrant, Packer, Consul, Serf, explaining how they can help DevOps streamline the entire process from development through to production.
Cornelia Davis takes the Spring Trader application and makes (a few) modifications (mostly to config) to get it running on the same components in the cloud, specifically on the Cloud Foundry PaaS.