InfoQ Homepage AI, ML & Data Engineering Content on InfoQ
-
Big Data in Memory
John Davies shows a Spring work-flow consuming 7.4kB XML messages, binding them to 25kB Java but storing them in just 450 bytes each, 10 million derivative contracts in-memory on a laptop.
-
Gobblin: A Framework for Solving Big Data Ingestion Problem
Lin Qiao discusses the architecture of Gobblin, LinkedIn’s framework for addressing the need of high quality and high velocity data ingestion.
-
Better Together - Using Spark and Redshift to Combine Your Data with Public Datasets
Eugene Mandel discusses challenges of conforming data sources and compares processing stacks: Hadoop+Redshift vs Spark, showing how the technology drives the way the problem is modeled.
-
Become a Data-driven Organization with Machine Learning
Peter Harrington explains what you do with machine learning, and what are the building blocks for an application that uses machine learning from collected data to creating predictions for customers.
-
Machine Learning for Programming
Peter Norvig keynotes on using machine learning techniques to solve more general software problems, helping both the advanced programmer and the novice one.
-
SQL Strikes Back! Recent Trends in Data Persistence and Analysis
Dean Wampler takes a look at SQL’s resurgence and specific example technologies, including: NewSQL, Hybrid SQL, SQL abstractions on top of file-based data, SQL as a functional programming language.
-
NoSQL Is Dead
Eric Redmond explains the differences and commonalities amongst many kinds of databases and takes a stab at the marketing term “NoSQL.”
-
Intro to Datomic
Stuart Sierra provides an introduction to Datomic's data model, architecture, query syntax, and transactions.
-
High Performance Computing Contributions to the World of Big Data
Sharan Kalwani presents the history of HPC and the technologies and trends which have contributed to creating the world of big data, covering applications of HPC resulting in big data technologies.
-
A Distributed Transactional Database on Hadoop
John Leach explains using HBase co-processors to support a full ANSI SQL RDBMS without modifying the core HBase source, showing how Hadoop/HBase can replace traditional RDBMS solutions.
-
Why Would You Integrate Solr and Hadoop?
Yann Yu discusses how Solr and Hadoop complement each other, and how to use Solr as a real-time, analytical, full-text search front-end to data stored in Hadoop.
-
1.5 Million Log Lines Per Second: Building and Maintaining Flume Flows at Conversant
Mike Keane presents how Conversant migrated to Flume, managing 1000 agents across 4 data centers, processing over 50B log lines per day with peak hourly averages of over 1.5 million log lines/sec.