InfoQ Homepage Database Content on InfoQ
-
Amazon DynamoDB Design Patterns & Best Practices
Siva Raghupathy discusses DynamoDB Design Patterns & Best Practices for realizing DynamoDB benefits at the right cost.
-
How Draw Something Scaled To 50 million New Users, in 50 Days, with Zero Downtime
Robin Johnson discusses using a data management model for games that can be scaled, and the bottlenecks and challenges met by OMGPOP scaling to millions of users.
-
How a Graph Database Allows Shutl to Deliver Even Faster
Volker Pacher explains why Shutl chose Neo4j when faced with the need of building a new API meant to support business growth, the challenges met during implementation and solutions applied.
-
Approximate Methods for Scalable Data Mining
Andrew Clegg overviews methods and provides use cases for performing data sets operations like membership testing, distinct counts, and nearest-neighbour finding more efficiently.
-
Data Science of Love
Vaclav Petricek digs some of the romantic interactions nuggets hidden in eHarmony's large collection of human relationships.
-
Leveraging Your Hadoop Cluster Better - Running Performant Code at Scale
Michael Kopp explains how to run performance code at scale with Hadoop and how to analyze and optimize Hadoop jobs.
-
Rebuilding Your Engine at 200 Miles per Hour
Michael Brunton-Spall shares his experience re-architect The Guardian’ Content API from a system based on Solr to a message queue cloud service based upon Elastic Search, without any downtime.
-
Core.logic and SQL Killed my ORM
Craig Brozefsky presents the tradeoffs involved with moving to a purely SQL relational model, instead of using an ORM, along with some of the tools built to facilitate this.
-
The Technology behind an Equity Trade
John O’Hara discusses banking business and technology integration, covering: low-latency, high-frequency trading, in-memory caches, multi-terabyte time-series databases, and contracts in NoSQL stores.
-
Equity – Transparent and Live Risk Assessment
Tormod Varhaugvik provides a design and rationale for an In Memory and Big Data architecture for live equity and risk assessment, using Tax Norway’ new architecture as an example.
-
Lessons Learned Building Storm
Nathan Marz shares lessons learned building Storm, an open-source, distributed, real-time computation system.
-
Building Applications using Apache Hadoop
Eli Collins overviews how to build new applications with Hadoop and how to integrate Hadoop with existing applications, providing an update on the state of Hadoop ecosystem, frameworks and APIs.