Neha Narkhede shares the experience at LinkedIn moving from ETL to real-time streams, the challenges of scaling Kafka to hundreds of billions of events/day, supporting thousands of engineers, etc.
Tim Wagner defines server-less computing, examines the key trends and innovative ideas behind the technology, and looks at design patterns for big data, event processing, and mobile using AWS Lambda.
Neha Narkhede explains how Apache Kafka was designed to support capturing and processing distributed data streams by building up the basic primitives needed for a stream processing system.
Martin Kleppmann explores using event streams and Kafka for keeping data in sync across heterogeneous systems, and compares this approach to distributed transactions.
Peter Bakas presents in detail how Netflix has used Kafka, Samza, Docker, and Linux to implement a multi-tenant pipeline processing 700B events/day in the Amazon AWS cloud.
Gian Merlino discusses stream processors and a common use case - keeping databases up to date-, the challenges they present, with examples from Kafka, Storm, Samza, Druid, and others.
Yongsheng Wu talks about how to build highly-resilient systems at scale. Wu presents also failure cases that prompted engineers at Pinterest to build such systems, and how they test these systems.
Neha Narkhede describes Apache Kafka and Samza: scalability and parallelism through data partitioning, fault tolerance, order guarantees, stateful processing, and stream processing primitives.
Marius Bogoevici demoes how to unleash the power of Kafka with Spring XD, by building a highly scalable data pipeline with RxJava and Kafka, using Spring XD as a platform.
Danny Yuan discusses how Uber uses stream processing to solve a wide range of problems, including real-time aggregation and prediction on geospatial time series, and much more.
Sharad Murthy & Tony Ng present Pulsar, a real-time streaming system which can scale to millions of events per second with high availability and 4GL language support.
Bart de Smet discusses how the reactive programming paradigm can be used for event stream processing and how it can be applied from small devices all the way to cloud-scale infrastructures.