BT

Facilitating the Spread of Knowledge and Innovation in Professional Software Development

Write for InfoQ

Topics

Choose your language

InfoQ Homepage Big Data Content on InfoQ

  • Data Workflow Management Using Airbnb's Airflow

    Airbnb recently opensourced Airflow, its own data workflow management framework. Airflow is being used internally at Airbnb to build, monitor and adjust data pipelines. Airflow’s creator, Maxime Beauchemin and Agari’s Data Architect and one of the framework’s early adopters Siddharth Anand discuss about Airflow, where it can be of use and future plans.

  • Microsoft Releases Azure Data Factory

    Any cloud provider that believes in data gravity is trying to make it easier to collect and store data in its facilities. To make data movement between cloud and on-premises endpoints easier, Microsoft recently announced the general availability of Azure Data Factory (ADF).

  • Data Quality at Prezi

    For an organization to be data-driven, it's not enough to just dump mountains of data. That data needs to be accurate and meaningful. Julianna Göbölös-Szabó, data engineer at Prezi shared how they improved the quality of its log data. Their solution involved moving from unstructured to structured data with a lightweight, contract-based approach to nudge all teams in the right direction.

  • Basho Data Platform Supports In-Memory Analytics, Caching, Search and Integration with NoSQL

    Basho Data Platform supports integration with NoSQL databases like Redis, in-memory analytics, caching, and search. Basho Technologies, the company behind Riak NoSQL database, announced in May, the availability of the data platform that can be used to deploy and manage Big Data, IoT and hybrid cloud applications.

  • Leveraging Data Science to Improve Monitoring

    At the recent devopsdays Amsterdam 2015, Patrick Roelke contended that monitoring still has lots of issues. Roelke believes that data science can help by eliminating static thresholds and coalescing information from various data sources into a single metric. The talk included a quick overview of monitoring tools that leverage data science: Kale, Bosun and AnomalyDetection.

  • Software Defined Data Mart In The Enterprise Using Metanautix Quest

    Metanautix recently announced the newest version of its product, Quest. Quest allows enterprises to build software defined data marts that can run in virtualized servers. Designed from the ground up with security and auditability in mind, Quest can deal with Big Data workloads and encapsulate it into different logical views, ready for consumption by different departments in enterprise.

  • Developments in IT Project Management

    The demand for IT project managers is increasing. Agile methodologies support collaboration with distributed teams for creative problem solving. The Internet of Things, cloud, big data, and cyber security will continue to dominate the IT landscape. Project managers have to pioneer IOT initiatives, be prepared for the influx of data and ensure that deliverables from their projects are secure.

  • Twitter Has Replaced Storm with Heron

    Twitter has replaced Storm with Heron which provides up to 14 times more throughput and up to 10 times less latency on a word count topology, and helped them reduce the needed hardware to a third.

  • Parquet Becomes Top-Level Apache Project

    Apache Parquet, the open-source columnar storage format for Hadoop, recently graduated from the Apache Software Foundation Incubator and became a top-level project. Initially created by Cloudera and Twitter in 2012 to speed up analytical processing, Parquet is now openly available for Apache Spark, Apache Hive, Apache Pig, Impala, native MapReduce, and other key components of the Hadoop ecosystem.

  • Capgemini Apollo: An Open Source Microservice and Big Data Platform

    Capgemini are currently working on Apollo, an open source application platform built on top of the Apache Mesos cluster manager and Docker, which is designed to power next generation web services, microservices and big data platforms running at scale.

  • MemSQL 4 Database Supports Community Edition, Geospatial Intelligence and Spark Integration

    Latest version of MemSQL, in-memory database with support for transactions and analytics, includes a new Community Edition for free use by organizations. MemSQL 4, released last week, also supports integration with Apache Spark, Hadoop Distributed File System (HDFS), and Amazon S3.

  • Deep Convolutional Networks for Super-Resolution Image Reconstruction at Flipboard

    Flipboard recently reported on an in-house application of deep learning to scale up low-resolution images that illustrates the power and flexibility of this class of learning algorithms.

  • Glenn Tamkin on Applying Apache Hadoop to NASA's Big Climate Data

    NASA Center for Climate Simulation (NCCS) is using Apache Hadoop for high-performance data analytics. Glenn Tamkin from NASA team, recently spoke at ApacheCon Conference and shared the details of the platform they built for climate data analysis with Hadoop.

  • Google Offers Bigtable in the Cloud

    Google is making available to customers Cloud Bigtable, their own database used for more than a decade for services such as Search, GMail, Maps or YouTube. While they are not open sourcing Bigtable as they did with other products, the new cloud service is accessible through an open source interface, the Apache HBase 1.0.1 API.

  • Adatao Launches Full Stack Data Intelligence Platform

    Adatao recently announced the general availability of its Data Intelligence platform. Its platform aims to make data analysis and predictive analytics available to everyone in large organizations. Adatao had secured an investment of $13 million last year from a group of investors including Bloomberg Beta, Lightspeed Venture Partners and Andreessen Horowitz.

BT