Sharan Kalwani presents the history of HPC and the technologies and trends which have contributed to creating the world of big data, covering applications of HPC resulting in big data technologies.
Daniel Woods discusses Enterprise Operations which is seeing a shift in the management paradigm of infrastructure.
John Leach explains using HBase co-processors to support a full ANSI SQL RDBMS without modifying the core HBase source, showing how Hadoop/HBase can replace traditional RDBMS solutions.
Roman Shaposhnik discusses more advanced features of HDFS, in addition to how YARN has enabled businesses to massively scale their systems beyond what was previously possible.
Sergi Almar reviews the architecture behind WebSocket and tries to show how to build highly scalable applications with it.
Yann Yu discusses how Solr and Hadoop complement each other, and how to use Solr as a real-time, analytical, full-text search front-end to data stored in Hadoop.
John Bunting talks about different services Tumblr has built and how their architecture helps them be fault tolerant as they continue to grow.
Andy Vaughn gives attendees a case study of how changing the development model and release cycle of a 5 year old software product to continuous delivery greatly improved the product.
Mike Keane presents how Conversant migrated to Flume, managing 1000 agents across 4 data centers, processing over 50B log lines per day with peak hourly averages of over 1.5 million log lines/sec.
Jeremy Stieglitz discusses best practices for a data-centric security , compliance and data governance approach, with a particular focus on two customer use cases.
Lisa Van Gelder provides simple tips and tricks for improving delivery without investing lots of time up front creating complex deployment frameworks.
Melody Meckfessel explores how Google's engineering teams use CD to build products and scale them, and how their strain of DevOps speeds launches and helps their engineering culture thrive.