InfoQ Homepage Machine Learning Content on InfoQ
-
Google Releases Major Firebase Studio Updates for Agentic AI Development
At Google Cloud Summit London in early July, Google revealed new capabilities in Firebase Studio that promise to enhance agentic cloud-based development: an autonomous Agent mode, native support for Model Context Protocol (MCP), and Gemini CLI integration. These updates aim to streamline agentic AI development by making AI agents more independent and seamlessly embedded in developer workflows.
-
Databricks Agent Bricks Automates Enterprise AI Development with TAO and ALHF Methods
Databricks introduced Agent Bricks, a new product that changes how enterprises develop domain-specific agents. The automated workflow includes generating task-specific evaluations and LLM judges for quality assessment, creating synthetic data that resembles customer data to supplement agent learning, and searching across optimization techniques to refine agent performance.
-
Microsoft Adds Deep Research Capability in Azure AI Foundry Agent Service
Unlock the future of research with Microsoft’s Azure AI Foundry Agent Service, featuring Deep Research—an innovative tool that empowers knowledge workers in complex fields. This advanced AI capability autonomously analyzes and synthesizes web data, automating rigorous research tasks while ensuring traceability and transparency. Sign up for the public preview today!
-
Arm Scalable Matrix Extension 2 Coming to Android to Accelerate On-Device AI
Available in the Armv9-A architecture, Arm Scalable Matrix Extension 2 (SME2) is a set of advanced CPU instructions designed to accelerate matrix heavy computation. The new Arm technology aims to help mobile developers to run advanced AI models directly on CPU with improved performance and efficiency, without requiring any changes to their apps.
-
The Rise of Energy and Water Consumption Using AI Models, and How It Can Be Reduced
Artificial intelligence's (AI) energy and water consumption has become a growing concern in the tech industry, particularly for large-scale machine learning models and data centers. Sustainable AI focuses on making AI technology more environmentally friendly and socially responsible.
-
QCon AI New York 2025: Program Committee Announced
Meet the QCon AI New York Program Committee, senior software leaders shaping a practical AI conference for engineers building at scale.
-
Google Cloud Run Now Offers Serverless GPUs for AI and Batch Processing
Google Cloud has launched NVIDIA GPU support for Cloud Run, enhancing its serverless platform with scalable, cost-efficient GPU resources. This upgrade enables rapid AI inference and batch processing, featuring pay-per-second billing and automatic scaling to zero. Developers can access seamless GPU support easily, making advanced AI applications faster and more accessible.
-
Virt8ra Sovereign Cloud Expands with Six New European Providers
Virt8ra is a groundbreaking European initiative aiming to establish a sovereign, interoperable cloud ecosystem, countering US cloud dominance. With significant expansion, now inclusive of six new providers, and a focus on open-source technology, Virt8ra promotes data localization and vendor independence, paving the way for an innovative digital future across Europe.
-
Azure AI Search Unveils Agentic Retrieval for Smarter Conversational AI
Microsoft’s Azure AI Search unveils agentic retrieval, a cutting-edge query engine that enhances conversational AI answer relevance by up to 40%. This dynamic system leverages conversation history and parallel subquery execution, paving the way for sophisticated knowledge retrieval. Currently in public preview, it offers adaptive search strategies tailored for evolving enterprise needs.
-
Uber Completes Massive Kubernetes Migration for Microservices and Large-Scale Compute Workloads
Uber has successfully completed a large Kubernetes migration, transitioning its entire compute platform from Apache Mesos to Kubernetes across multiple data centers and cloud environments.
-
Google Enhances LiteRT for Faster On-Device Inference
The new release of LiteRT, formerly known as TensorFlow Lite, introduces a new API to simplify on-device ML inference, enhanced GPU acceleration, support for Qualcomm NPU (Neural Processing Unit) accelerators, and advanced inference features.
-
OpenAI’s Stargate Project Aims to Build AI Infrastructure in Partner Countries Worldwide
OpenAI has announced a new initiative called "OpenAI for Countries" as part of its Stargate project, aiming to help nations develop AI infrastructure based on democratic principles. This expansion follows the company's initial $500 billion investment plan for AI infrastructure in the United States.
-
Google Cloud Enhances AI/ML Workflows with Hierarchical Namespace in Cloud Storage
On March 17, 2025, Google Cloud introduced a hierarchical namespace (HNS) feature in Cloud Storage, aiming to optimize AI and machine learning (ML) workloads by improving data organization, performance, and reliability.
-
DeepSeek Launches Prover-V2 Open-Source LLM for Formal Math Proofs
DeepSeek has released DeepSeek-Prover-V2, a new open-source large language model specifically designed for formal theorem proving in Lean 4. The model builds on a recursive theorem proving pipeline powered by the company's DeepSeek-V3 foundation model.
-
Uber’s Journey to Ray on Kubernetes
Uber has detailed a recent transition to running Ray-based machine learning workloads on Kubernetes. This marks an evolution in its infrastructure, with the aim of enhancing scalability, efficiency, and developer experience. The company recently published a two-part series from Uber Engineering delving into the motivations, challenges, and solutions encountered during this migration.