In this article, author discusses the evolving technologies like Machine Learning and Internet of Things and how to exploit them for data analytics. He also talks about how organizations can benefit from these new sources of information and intelligence embedded in their environments.
DevOps is the industrialization of IT, says Nati Shalom. Organizations that wish to optimize for speed and cost cannot afford silos anymore."Doing DevOps" is not adding new features to existing tools. In this article, Shalom takes us through the differences between management solutions in a pre and post DevOps world.
Graph NoSQL databases support data models with connected data and relationships. In this article, author discusses the security implications of graph database technology. He talks about the privacy and security concerns in use cases like graph discovery, knowledge management, and prediction.
When building a web application there are lot of choices for the database. In this article, author discusses why Neo4j is a good choice if data model contains lot of connected data and relationships.
In this article, Carlos Bueno describes a method for analyzing constraints on the shape and flow of data in systems. He also discusses details of two use cases, movie streaming and face recognition.
Spark SQL, part of Apache Spark, is used for structured data processing by running SQL queries on Spark data. Srini Penchikala discusses Spark SQL module & how it simplifies data analytics using SQL. 1
Bulk data is usually accessed via files & FTP. As collaboration migrates to APIs, what are the requirements for data APIs? Here we describe a meta-data driven architecture for bulk data ingestion.
In this article, Basho Sr. Software Engineer Chris Meiklejohn explores the basic building blocks for crafting deterministic applications that guarantee convergence of data without synchronization. 1
Apache Spark is an open source big data framework built around speed, ease of use, and sophisticated analytics. In this article, Srini Penchikala discusses how Spark helps with big data processing. 3
This article shows how to use Amazon DynamoDB to create a Mars Rover application. You can use the same concepts described in this post to build your own web application. 1
When it comes to database change, agility through automation - the ability to rapidly to accelerate delivery – is what differentiates world-class enterprises from the rest of the crowd. 4
GridGain announced that the In-Memory Data Fabric has been accepted into Apache Incubator program as Apache Ignite. InfoQ spoke with Nikita Ivanov about their product becoming part of Apache.