InfoQ Homepage Database Content on InfoQ
-
Building Scalable Applications in .NET: Introducing the FatDB Distributed Computing Platform
Justin Weiler introduces FatDB, a NoSQL DB and a distributed platform built on Mission Oriented Architecture meant to abstract and generalize the essential characteristics of enterprise applications.
-
Exploring the Architecture of the NuoDB Database, Part 2
In Part 2 of this article the author takes a look at how the transaction system is implemented, the role of the administrative layer, how all components work together and what to expect in the future.
-
Exploring the Architecture of the NuoDB Database, Part 1
In Part 1 of this article the author introduces NuoDB and covers some of its main features: 3-tiered architecture, nodes are equal peers, Atoms - the fundamental data unit, and the versioning and concurrency system used to handle data update conflicts and implement consistency.
-
Spoilt for Choice – How to choose the right Big Data / Hadoop Platform?
In his new article Kai Wähner compares several alternatives for installing a version of Hadoop and realizing big data processes. He compares distributions and tooling from Apache and many other vendors including Cloudera, HortonWorks, MapR, Amazon, IBM, Oracle, Microsoft. He additionally describes pros and cons of every distribution and provides a decision tree for choosing a most appropriate one.
-
Apache MetaModel – Providing Uniform Data Access Across Various Data Stores
MetaModel - an Apache Incubator project – is a Java library used to browse, query and update various types of data stores including traditional SQL databases, unusual stores such as CSV or Excel, or the more modern NoSQL stores in a uniform and programmatic way.
-
Cassandra Mythology
In this article, author Jonathan Ellis addresses the concerns of using Apache Cassandra NoSQL database, in terms of architecture, deployment and configuration, performance, query language (CQL), and database maturity.
-
Jepsen: Testing the Partition Tolerance of PostgreSQL, Redis, MongoDB and Riak
Distributed systems are characterized by exchanging state over high-latency or unreliable links. The system must be robust to both node and network failure if it is to operate reliably--however, not all systems satisfy the safety invariants we'd like. In this article, we'll explore some of the design considerations of distributed databases, and how they respond to network partitions.
-
Graph Databases - Book Review and Interview
"Graph Databases" book covers the Graph based NoSQL database technology and different options available for storing "Connected Data" in the real world applications. InfoQ spoke with co-authors Ian Robinson and Jim Webber about the book, role of Graph Databases in the NoSQL database space, and what’s coming up in the Graph Databases.
-
Mike Barlow on Real-Time Big Data Analytics
"Real-Time Big Data Analytics: Emerging Architecture" white paper authored by Mike Barlow covers big data analytics topic and how real-time big data analytics (RTBDA) are different from traditional analytics. InfoQ spoke with Mike about the current state of real-time big data analytics and the emerging trends in the Big Data space like Decision Science.
-
Interview and Book Review: The LogStash Book, Log Management Made Easy
James Turnbull makes a compelling case for using Logstash for centralizing logging by explaining the implementation details of Logstash within the context of a logging project. The book targets both small companies and large enterprises through a two sided case; both for the low barrier to entry and the scaling capabilities.
-
Interview and Video Review: Working with Big Data: Infrastructure, Algorithms, and Visualizations
Paul Dix leads a practical exploration into Big Data in this video training series. The first five lessons of the training span multiple server systems with a focus on the end to end processing of large quantities of XML data from real Stack Exchange posts. He completes the training with a lesson on developing visualizations for gaining insights from the macro level analysis of Big Data.
-
Refactoring Legacy Applications: A Case Study
To refactor legacy code, the ideal is to have a suite of unit tests to prevent regressions. However it's not always that easy. This article describes a methodology to safely refactor legacy code.