Vaclav Petricek discusses how to train models, architect and build a scalable system powered by Storm, Hadoop, Spark, Spring Boot and Vowpal Wabbit that meets SLAs measured in tens of milliseconds.
Piotr Kołaczkowski discusses how they integrated Spark with Cassandra, how it was done, how it works in practice and why it is better than using a Hadoop intermediate layer.
Eugene Mandel discusses challenges of conforming data sources and compares processing stacks: Hadoop+Redshift vs Spark, showing how the technology drives the way the problem is modeled.
Michael Minella uses Spring XD and Spring Batch to orchestrate the full lifecycle of Hadoop processing and uses Apache Mahout to provide the audience with the recommendation processing.
Matt Stine presents how combine Spring Boot, Spring Data, Spring Reactor, Spring XD, Hadoop and run them in the cloud.
Thomas Risberg introduces the Spring for Apache Hadoop project and discusses integration with Spring XD, batch jobs and external data sources.
John Leach explains using HBase co-processors to support a full ANSI SQL RDBMS without modifying the core HBase source, showing how Hadoop/HBase can replace traditional RDBMS solutions.
Roman Shaposhnik discusses more advanced features of HDFS, in addition to how YARN has enabled businesses to massively scale their systems beyond what was previously possible.
Yann Yu discusses how Solr and Hadoop complement each other, and how to use Solr as a real-time, analytical, full-text search front-end to data stored in Hadoop.
Jeremy Stieglitz discusses best practices for a data-centric security , compliance and data governance approach, with a particular focus on two customer use cases.
Dean Wampler argues that Spark/Scala is a better data processing engine than MapReduce/Java because tools inspired by mathematics, such as FP, are ideal tools for working with data.
Bob Kelly presents case studies on how Platfora uses Hadoop to do analytics for several of their customers.