InfoQ Homepage Database Content on InfoQ
-
Taking the Pain out of Real-time Mobile Back-end Development
Mandy Waite shows how to get started with Firebase before walking through a live demo of building a multi-user, collaborative mobile app that provides real-time updates to its users.
-
Interactive Analytics at Scale with Druid
Julien Lavigne du Cadet discusses how Criteo uses Druid: an open-source, real-time data store designed to power interactive applications at scale, covering Druid's architecture and internals.
-
Mini-talks: Machine Intelligence, Algorithms for Anti-Money Laundering, Blockchain
Mini-talks: The Machine Intelligence Landscape: A Venture Capital Perspective. The future of global, trustless transactions on the largest graph: blockchain. Algorithms for Anti-Money Laundering
-
Mini-talks: OS/App Inversion, Testing & Living Databases
Mini-talks on: OS/application inversion, testing, living databases, and rogue protocols.
-
Applying Reactive Programming to Existing Applications
Ben Christensen discusses the mental shift from imperative to declarative programming, working with blocking IO such as JDBC and RPC, service composition, debugging and unit testing.
-
Financial Modeling with Apache Spark: Calculating Value at Risk
Sandy Ryza aims to give a feel for what it is like to approach financial modeling with modern big data tools, using the Monte Carlo method for a a basic VaR calculation with Spark.
-
LDAP at Lightning Speed
Howard Chu covers highlights of the LMDB design and discusses some of the internal improvements in slapd due to LMDB, as well as the impact of LMDB on other projects.
-
Product thru the Looking Glass
Chris Matts discusses how to manage product mastery, how do we decide whether to use analysis or product management techniques, and what does an end-to-end process looks like.
-
The SenseMaker® Method
Tony Quinlan introduces the SenseMaker® method from preparing the ground through gathering experiences and qualitative material to analysis and action planning.
-
Lightning Fast Cluster Computing with Spark and Cassandra
Piotr Kołaczkowski discusses how they integrated Spark with Cassandra, how it was done, how it works in practice and why it is better than using a Hadoop intermediate layer.
-
Translating Imperative Code to MapReduce
The authors present an approach for automatic translation of sequential, imperative code into a parallel MapReduce framework using Mold, translating Java code to run on Apache Spark.
-
The Deep Learning Revolution: Rethinking Machine Learning Pipelines
Soumith Chintala introduces deep learning, what it is, why it has become popular, and how it can be fitted into existing machine learning solutions.