Nick Kolegraff discusses common problems and architecture to support all the phases of data science and how to start a data science initiative, sharing lessons from Accenture, Best Buy, and Rackspace.
Sebastian Kanthak details how Spanner relies on GPS and atomic clocks to provide two of its innovative features: Lock-free strong reads and global snapshots consistent with external events.
Dan Frank discusses stream data processing and introduces NSQ – Bitly’s open source queuing system – and other new technologies used for communication between streaming programs.
Alex Miller discusses Clojure’s approach to data, comparing it with OOP’s approach, and covering various related topics such as mutation, state vs. value, primitive and composite data.
Andrew Clegg overviews methods and provides use cases for performing data sets operations like membership testing, distinct counts, and nearest-neighbour finding more efficiently.
Scott Vokes presents some lesser-known data structures and shows how probability distributions and content-addressable storage can become tools to shape global system behavior.
The panelists discuss their approaches in using APIs and open standards and data in the education sector.
Ian Plosker shares a number of techniques for establishing the data query patterns from the outset of application development, designing a data model to fit those patterns.
Jeni Tennison explains how to evaluate an organization's data assets as potential sources of open data, and how to deal with the thorny issues of derived and personal data.
Steve Miner discusses tag literals and data readers, what’s new in Clojure 1.5, and EDN (Extensible Data Notation).
Stephane Dubois shares insight in Xignite’s road building a business model providing APIs for accessing financial data.