InfoQ Homepage AI, ML & Data Engineering Content on InfoQ
-
DataStax Announces Cloud Native Database as a Service and AIOps Tools
DataStax announced last month the release of Astra, a cloud-native Database-as-a-Service (DBaaS) built on Apache Cassandra. They also recently announced an AIOps product called Vector that proactively monitors the health of Apache Cassandra clusters.
-
Google ML Kit SDK Now Focuses on On-Device Machine Learning
Google has introduced a new ML Kit SDK aimed at working in standalone mode without requiring a tight integration with Firebase, as the original ML Kit SDK did. Additionally, it provides limited support for replacing its default models with custom ones for image labeling and object detection and tracking.
-
Facebook Announces TransCoder AI to Translate Code across Programming Languages
Facebook AI Research has announced TransCoder, a system that uses unsupervised deep-learning to convert code from one programming language to another. TransCoder was trained on more than 2.8 million open source projects and outperforms existing code translation systems that use rule-based methods.
-
Uber Open-Sources AI Abstraction Layer Neuropod
Uber open-sourced Neuropod, an abstraction layer for machine learning frameworks that allows researchers to build models in the framework of their choice while reducing the effort of integration, allowing the same production system to swap out models implemented in different frameworks. Neuropod currently supports several frameworks, including TensorFlow, PyTorch, Keras, and TorchScript.
-
Paddle Quantum: Bringing Baidu Deep Learning Perform to Quantum Computing
Baidu has announced quantum machine learning toolkit Paddle Quantum, which makes it possible to build and train quantum neural network models. Paddle Quantum aims to support advanced quantum computing applications as well as to allow developers new to quantum machine learning to create their models step-by-step.
-
Google Open-Sources Computer Vision Model Big Transfer
Google Brain has released the pre-trained models and fine-tuning code for Big Transfer (BiT), a deep-learning computer vision model. The models are pre-trained on publicly-available generic image datasets and can meet or exceed state-of-the-art performance on several vision benchmarks after fine-tuning on just a few samples.
-
Microsoft Announces a New Pricing Model Option for Azure Cosmos DB and More Capabilities
During this year's digital Build event, Microsoft announced a new consumption-based pricing model for Azure Cosmos DB called 'serverless' that is ideal for spiky workloads and complements the existing provisioned throughput pricing model.
-
OpenAI Announces GPT-3 AI Language Model with 175 Billion Parameters
A team of researchers from OpenAI recently published a paper describing GPT-3, a deep-learning model for natural-language with 175 billion parameters, 100x more than the previous version, GPT-2. The model is pre-trained on nearly half a trillion words and achieves state-of-the-art performance on several NLP benchmarks without fine-tuning.
-
QCon San Francisco 2020 Announces Program Committee
The QCon team has finalized the Program Committee for QCon San Francisco 2020 (Nov 16-18). The committee works on all aspects of software development. At QCon topics, track hosts and speakers are handpicked to guarantee relevant and timely content.
-
AWS Releases its Machine Learning Powered Enterprise Search Service Kendra into General Availability
Recently Amazon announced the general availability of its enterprise search service Kendra on AWS. With the GA release of Amazon Kendra, the public cloud provider added a few new specialized features and improved service accuracy.
-
Microsoft Build 2020: Highlights
Last week Microsoft held the 10th edition of Build, its annual conference aimed at developers using Microsoft technologies. The online event included multiple important announcements and releases, such as the general availability of Blazor WebAssembly, updates on the upcoming .NET 5, Azure Static Web Apps, and new projects related to IoT and Artificial Intelligence.
-
Google Open-Sources AI for Using Tabular Data to Answer Natural Language Questions
Google open-sourced Table Parser (TAPAS), a deep-learning system that can answer natural-language questions from tabular data. TAPAS was trained on 6.2 million tables extracted from Wikipedia and matches or exceeds state-of-the-art performance on several benchmarks.
-
Jepsen Disputes MongoDB’s Data Consistency Claims
In an article, MongoDB claimed their database passed “the industry’s toughest data safety, correctness, and consistency Tests”. In response, Jepsen published an article stating that MongoDB 3.6.4 had in fact failed their tests; the newer MongoDB 4.2.6 has more problems including “retrocausal transactions” where a transaction reverses order so that a read can see the result of a future write.
-
Amazon Announces General Availability of UltraWarm for Its Elastic Search Service on AWS
Recently, Amazon announced the general availability of UltraWarm for its Elasticsearch Service on AWS. Ultrawarm is a low cost warm storage tier, and extension to the Elasticsearch Service - offering up to three petabytes of storage, at almost a 90% cost reduction over existing options.
-
Google AI Launches NLU-Powered Tool to Help Explore COVID-19 Literature
Google AI launched COVID-19 Research Explorer, which provides a semantic search interface on top of the COVID-19 Open Research Dataset to help scientists and researchers efficiently analyze all of the dataset’s journal articles and preprints.