InfoQ Homepage AI, ML & Data Engineering Content on InfoQ
-
Shopify’s Practical Guidelines from Running Airflow for ML and Data Workflows at Scale
Shopify engineering shared its experience in the company's blog post on how to scale and optimize Apache Airflow for running ML and data workflows. They shared practical solutions for the challenges they faced like slow file access, insufficient control over DAG, irregular level of traffic, resource contention among workloads, and more.
-
Obituary: Alex Blewitt
It is with great sadness that we announce that InfoQ editor Dr. Alex Blewitt has unexpectedly passed away.
-
Google's Image-Text AI LIMoE Outperforms CLIP on ImageNet Benchmark
Researchers at Google Brain recently trained Language-Image Mixture of Experts (LIMoE), a 5.6B parameter image-text AI model. In zero-shot learning experiments on ImageNet, LIMoE outperforms CLIP and performs comparably to state-of-the-art models while using fewer compute resources.
-
PyTorch 1.12 Release Includes Accelerated Training on Macs and New Library TorchArrow
The PyTorch open-source deep-learning framework announced the release of version 1.12 which includes support for GPU-accelerated training on Apple silicon Macs and a new data preprocessing library, TorchArrow, as well as updates to other libraries and APIs.
-
Google AI Developed a Language Model to Solve Quantitative Reasoning Problems
Google AI developed a deep learning language model called Minerva which could solve mathematical quantitative problems. Google AI researchers achieved a state-of-the-art deep learning model by training on a large dataset that contains quantitative reasoning with symbolic expressions. The final model, Minerva, could solve quantitative mathematical problems on STEM reasoning tasks.
-
MLGO Framework Brings Machine Learning in Compiler Optimizations
Google’s new Machine Learning Guided Optimization (MLGO) is an industrial-grade general framework for integrating machine-learning (ML) techniques systematically in a compiler and in particular in LLVM. Compiling faster and smaller code can significantly reduce the operational cost of large data-center applications.
-
OpenAI Releases Minecraft-Playing AI VPT
Researchers from OpenAI have open-sourced Video PreTraining (VPT), a semi-supervised learning technique for training game-playing agents. In a zero-shot setting, VPT performs tasks that agents cannot learn via reinforcement learning (RL) alone, and with fine-tuning is the first AI to craft a diamond pickaxe in Minecraft.
-
Google's BigQuery Introduces Column-Level Encryption Functions and Dynamic Masking of Information
Google recently released new features for its SaaS data warehouse BigQuery which include column level encryption functions and dynamic masking of information. Specifically, dynamic masking of information can be used for real-time transactions whereas column level encryption provides additional security for data at rest or in motion where real-time usability is not required.
-
LinkedIn Open-Sourced Its Feature Store to Evangelize Productive Machine Learning
LinkedIn Engineering recently open-sourced its feature store Feathr, which helps engineers to develop machine Learning products by simplifying feature management and usage in production. It defines features, computes them for training and inference purposes, and makes them discoverable by other machine learning developers.
-
Amazon Unveils ML-Powered Coding Assistant CodeWhisperer
Amazon launched CodeWhisperer, an ML-Powered Coding Companion which provides code recommendations based on developers' comments in natural language and their code in the integrated development environment. The machine learning-powered service increases developer productivity.
-
Adobe Researchers Open-Source Image Captioning AI CLIP-S
Researchers from Adobe and the University of North Carolina (UNC) have open-sourced CLIP-S, an image-captioning AI model that produces fine-grained descriptions of images. In evaluations with captions generated by other models, human judges preferred those generated by CLIP-S a majority of the time.
-
AWS and Microsoft Working Together on PyWhy, the New Home of Causal ML Library DoWhy
AWS in a joint effort with Microsoft have established PyWhy as a fresh GitHub organization to integrate AWS algorithms into DoWhy, a casual ML library from Microsoft, which has moved to PyWhy.
-
Stanford University Open-Sources Controllable Generative Language AI Diffusion-LM
Researchers at Stanford University have open-sourced Diffusion-LM, a non-autoregressive generative language model that allows for fine-grained control of the model's output text. When evaluated on controlled text generation tasks, Diffusion-LM outperforms existing methods.
-
Amazon Released Incremental Training Feature in SageMaker JumpStart
AWS recently released a new feature in SageMaker (AWS Machine Learning Service) JumpStart to incrementally retrain machine-learning (ML) models trained with expanded datasets. By using this feature, developers could fine-tune their models for better performance in production with a couple of clicks. This recent feature is among the series of efforts to add more automation to SageMaker JumpStart.
-
Microsoft Limits Public Access to AI-Powered Facial Analysis Features
Microsoft recently announced phasing out public access to AI-powered Facial Analysis features in several Azure services.