Michael Wise discusses the journey from having data integrated across an organization, to employing data science to make good use of it.
Dylan Beattie discusses the challenges faced restructuring Spotlight so their organization aligns with the natural domain boundaries of their business.
Paul Payne explains the benefits of containerization of a Go web service, discussing testing, integration, canary deploys and how they achieve 20 minute development cycles with zero downtime.
Giannakakis and Dalkitsis present how Shazam releases faster, more predictably and with more features by using BDD and automation testing, without slowing down or hindering the development process.
Maxime Porhel discusses possible integration paths between Sirius and Xtext with demos based on Xtext DSLs, and the latest progress made in Sirius 3.0 regarding this integration.
Piotr Kołaczkowski discusses how they integrated Spark with Cassandra, how it was done, how it works in practice and why it is better than using a Hadoop intermediate layer.
Stuart Williams examines some of the problems faced building an application processing billions of events/day with Spring Integration, Spring Expression Language, Reactor and the LMAX Disruptor.
Shobana Radhakrishnan shares details about best practices adopted in implementing API integration with third party services, how to manage change and deal with failures.
Josh Long and Phillip Webb present what Spring Boot is, why it's turning heads, why you should consider it for your next application and how to get started.
Sam McAfee outlines an approach they are using at Neo to help their clients integrate business metrics and probabilistic modeling into their prioritization process.
Garrett Wampole describes an experimental methodology of applying Enterprise Integration Patterns to the near real-time processing of surveillance radar data, developed by MITRE.
Yann Yu discusses how Solr and Hadoop complement each other, and how to use Solr as a real-time, analytical, full-text search front-end to data stored in Hadoop.