InfoQ Homepage Database Content on InfoQ
-
Evolving Panorama of Data
Rebecca Parsons reviews some of the changes in how data is used and analyzed, looking at how data is used to track violence, and attempts to predict famine and other crises before they happen.
-
Leveraging Scriptable Infrastructures, Towards a Paradigm Shift in Software for Data Science
Karim Chine introduces Elastic-R, demonstrating some of its applications in bioinformatics and finance.
-
Amazon DynamoDB Design Patterns & Best Practices
Siva Raghupathy discusses DynamoDB Design Patterns & Best Practices for realizing DynamoDB benefits at the right cost.
-
How Draw Something Scaled To 50 million New Users, in 50 Days, with Zero Downtime
Robin Johnson discusses using a data management model for games that can be scaled, and the bottlenecks and challenges met by OMGPOP scaling to millions of users.
-
How a Graph Database Allows Shutl to Deliver Even Faster
Volker Pacher explains why Shutl chose Neo4j when faced with the need of building a new API meant to support business growth, the challenges met during implementation and solutions applied.
-
Approximate Methods for Scalable Data Mining
Andrew Clegg overviews methods and provides use cases for performing data sets operations like membership testing, distinct counts, and nearest-neighbour finding more efficiently.
-
Data Science of Love
Vaclav Petricek digs some of the romantic interactions nuggets hidden in eHarmony's large collection of human relationships.
-
Leveraging Your Hadoop Cluster Better - Running Performant Code at Scale
Michael Kopp explains how to run performance code at scale with Hadoop and how to analyze and optimize Hadoop jobs.
-
Rebuilding Your Engine at 200 Miles per Hour
Michael Brunton-Spall shares his experience re-architect The Guardian’ Content API from a system based on Solr to a message queue cloud service based upon Elastic Search, without any downtime.
-
Core.logic and SQL Killed my ORM
Craig Brozefsky presents the tradeoffs involved with moving to a purely SQL relational model, instead of using an ORM, along with some of the tools built to facilitate this.
-
The Technology behind an Equity Trade
John O’Hara discusses banking business and technology integration, covering: low-latency, high-frequency trading, in-memory caches, multi-terabyte time-series databases, and contracts in NoSQL stores.
-
Equity – Transparent and Live Risk Assessment
Tormod Varhaugvik provides a design and rationale for an In Memory and Big Data architecture for live equity and risk assessment, using Tax Norway’ new architecture as an example.