Rich Hickey discusses the complexity introduced by a database into a system, and a way to deal with it by using Datomic. He also discusses immutability, epochal time, and persistent data structures.
Stuart Sierra discusses using a data-oriented programming approach in order to create programs that are easier to write and test. The session is accompanied with Clojure code samples.
Frank Tarsillo , John Davies, Jon Vernon and Ari Zilka (moderator) discuss the technologies and architectures used these days to manage large amounts of sensitive data in top financial institutions.
Jim Webber talks about the data of these days, how integrated data looks, how to model it using actual data stores and the implications of this modeling.
James Spooner discusses the need to make good use of the underlying silicon using Dataflow computing and parallelism to improve throughput and latency for optimized data processing performance.
Ashish Thusoo presents the data scalability issues at Facebook and the data architecture evolution from EDW to Hadoop to Puma.
Baishampayan Ghose discusses creating custom data types in Clojure, covering: types vs. records, interfaces and corresponding protocol, mutable types, and example implementations.
Paul Sanford presents the transformations supported by data throughout its life cycle, and how that can be better done with Splunk, an engine for monitoring and analyzing machine-generated data.
Sastry Malladi discusses the performance implications of using various data formats and versioning across eBay, showing the results of certain benchmarks concluding that JSON is the best format.