Genomic data sequencing and subsequent analysis faces large data volume challenges that several organizations are solving with cloud services. The Broad Institute detailed their experience with petabyte scale sequencing pipelines last month through the Google Research Blog and is detailed here by InfoQ.
NoSQL database adoption in a large organization takes significant effort and time for the transition from using relational database models to NoSQL databases. Mike Bowers, Enterprise Data Architect at LDS Church, spoke at the recent Enterprise Data World Conference about lessons learned from eight years of using NoSQL databases.
After months of awaiting details about the NHS and Google DeepMind partnership InfoQ gains insights into recent claims of widespread patient data access.
SQL Server 2005 has now officially hit its end of life. This means that it will no longer receive security updates and new vulnerabilities that are discovered will go unfixed. Yet a recent survey commissioned by Microsoft showed that 46% of companies using SQL Server had at least one production machine running SQL Server 2005.
Hadoop and other big data technologies revolutionized the way organizations run data analytics but the organizations are still facing challenges with operating costs of using these technologies for on-premise data processing. Ashish Thusoo recently spoke at Enterprise Data World Conference about Hadoop as a service offering that helps organizations bridge the gaps with these capabilities.
AirFlow recently joined the Apache Incubator program. AirFlow is a workflow and scheduling system designed to manage data pipelines. Developed by AirBnb for their internal usage, it was open sourced last September, as previously reported by InfoQ.
In July 2013 Security Explorations discovered a vulnerability in Java by which attackers could elevate their access privileges. Oracle released a patch, but a simple modification was discovered that still makes the attack effective. Once known, Oracle released a patch as part of 8u77. In this article we investigate the little understood class loading process at the heart of the problem.
Operational Data Stream and Batch Processing at Netflix with Mantis
Today at GraphConnect Europe 2016, Neo Technology announced the release of Neo4j 3.0, which includes a new binary protocol for transmitting data between server and client, and a new set of standardised drivers for interacting with the database, along with stored procedure support and higher performance and capacity. InfoQ spoke to Neo Technology to find out more.
Late last month Google released an alpha version of their TensorFlow (TF) integrated cloud machine learning service as a response to a growing need to make their Tensor Flow library to run at scale on the Google Cloud Platform (GCP). Google describes several new feature sets around making TF usage scale by integrating several pieces of the GCP like Dataproc, a managed Hadoop and Spark service.
Recently at the 2016 Build Event in San Francisco, Microsoft announced a change to their Power BI offering. The update comes in the form of giving customers and ISVs with the ability to embed Power BI reports within their own applications. Microsoft is calling this service Power BI Embedded and it is currently in preview.
Microsoft has dramatically changed the support policy for SQL Server. In the past, the policy was essentially "wait for a service pack unless you have a problem". Now they want developers to proactively Cumulative Updates.
SQL Server 2016 is removing a lot of the limitations on memory-optimized tables including the inability to work with the large documents favored by NoSQL style designs.
After a brief beta period that saw customers migrate more than 1,000 on-premises databases to the cloud, AWS formally released their Database Migration Service. This on-demand cloud service supports live migration scenarios, and customers who wish to switch their database platform as part of the migration can do so, thanks to a free schema conversion tool.