InfoQ Homepage Data Partitioning Content on InfoQ
News
RSS Feed-
From Outages to Order: Netflix’s Approach to Database Resilience with WAL
Netflix uses a Write-Ahead Log (WAL) system to improve data platform resilience, addressing data loss, replication entropy, multi-partition failures, and corruption. WAL decouples producers and consumers, leverages SQS/Kafka with dead-letter queues, and supports delay queues, cross-region replication, and multi-table mutations for high-throughput, consistent, and recoverable database operations.
-
Data Preparation Pipelines: Strategy, Options and Tools
Data preparation is an important aspect of data processing and analytics use cases. Business analysts and data scientists spend about 80% of their time gathering and preparing the data rather than analyzing it or developing machine learning models. Kelly Stirman spoke last week at Enterprise Data World 2017 Conference about the data preparation best practices.
-
VMware Releases SQLFire 1.0
VMware releases SQLFire 1.0 a distributed SQL database geared towards high availability and horizontal scalability which offers table replication, table partitioning and parallel execution of queries.
-
JBoss Releases Hibernate 4.0
JBoss Releases Hibernate 4.0 which comes with Multi-tenancy support, the introduction of a standard mechanism for writing Hibernate extensions, initial refactorings towards OSGI and several other cleanups.
-
Facebook on Hadoop, Hive, HBase, and A/B Testing
The Hadoop Summit of 2010 included presentations from a number of large scale users of Hadoop and related technologies. Notably, Facebook presented a keynote and details information about their use of Hive for analytics. Mike Schroepfer, Facebook's VP of Engineering delivered a keynote describing the scale of their data processing with Hadoop.
-
Databases Roundup: Data Sharding for ActiveRecord and Faster Postgres IO
In this databases roundup we take a look at DataFabric, FiveRun's recently open sourced data sharding plug-in for ActiveRecord. Also: a look at speeding up Postgres data access using the asynchronous client API and Ruby 1.9's Fibers.