Dean Wampler argues that Spark/Scala is a better data processing engine than MapReduce/Java because tools inspired by mathematics, such as FP, are ideal tools for working with data.
Bob Kelly presents case studies on how Platfora uses Hadoop to do analytics for several of their customers.
Jayesh Thakrar shows what can be done with irb, how to exploit JRuby-Java integration, and demonstrates how the Shell can be used in Hadoop streaming to perform complex and large volume batch jobs.
Carlos Queiroz introduces the lambda architecture and showcases how it can be implemented with SpringXD, GemFireXD and Hadoop in a CDR(Call Detail Record) mining application.
The authors explain how the Pivotal team leveraged familiar SQL-based queries to analyze fine-grained cluster utilization using Spring XD.
Mohammad Quraishi presents implementing a Big Data initiative, detailing preparation, goal evaluation, convincing executives, and post implementation evaluation.
Ted Dunning discusses the different options for running SQL on Hadoop including pros and cons.
Jim Scott keynotes on the history of Hadoop, the difficulties that this technology has gone through, exploring the reasons why enterprises need to evaluate their targets and prepare for the future.
Randy Shoup describes KIXEYE's analytics infrastructure from Kafka queues through Hadoop 2 to Hive and Redshift, built for flexibility, experimentation, iteration, testability, and reliability.
Marcel Kornacker presents a case study of an EDW built on Impala running on 45 nodes, reducing processing time from hours to seconds and consolidating multiple data sets into one single view.
In this solutions track talk, sponsored by Cloudera, Eva Andreasson discusses how search and Hadoop can help with some of the industry's biggest challenges. She introduces the data hub concept.
In this solutions track talk, sponsored by MongoDB, Matt Asay discusses the differences between some of the NoSQL and SQL databases and when Hadoop makes sense to be used with a NoSQL solution.