Martin Kleppmann explores using event streams and Kafka for keeping data in sync across heterogeneous systems, and compares this approach to distributed transactions.
Peter Bakas presents in detail how Netflix has used Kafka, Samza, Docker, and Linux to implement a multi-tenant pipeline processing 700B events/day in the Amazon AWS cloud.
David Talby demos using Python libraries to build a ML model for fraud detection, scaling it up to billions of events using Spark, and what it took to make the system perform and ready for production.
Sid Anand discusses how Agari is applying big data best practices to the problem of securing its customers from email-born threats, presenting a system that leverages big data in the cloud.
Irad Ben-Gal discusses Big Data analytics misconceptions, presenting a technology predicting consumer behavior patterns that can be translated into wins, revenue gains, and localized assortments.
Jan Neumann presents how Comcast uses machine learning and big data processing to facilitate search for users, for capacity planning, and predictive caching.
Mathieu Bastian explores the mechanics of unit, integration, data and performance testing for large, complex data workflows, along with the tools for Hadoop, Pig and Spark.
Adam Miskiewicz goes beyond the React Native docs and talks about best practices for building responsive and production-ready React Native applications with Redux, Relay, and GraphQL.
Janet Wiener discusses using a data pipeline and graphic visualizations to extract and analyze the Chorus – the aggregated, anonymized voice of the people communicating on Facebook - in real time.
Sudhir Tonse discusses using stream processing at Uber: indexing and querying of geospatial data, aggregation and computing of streaming data, extracting patterns, TimeSeries analyses and predictions.
Robert Metzger provides an overview of the Apache Flink internals and its streaming-first philosophy, as well as the programming APIs.
Karthik Ramasamy presents the design and implementation of Heron, the new de facto stream data processing engine at Twitter. Ramasamy shares Twitter’s experience of running Heron in production.