InfoQ Homepage AI, ML & Data Engineering Content on InfoQ
-
Google Releases MedGemma: Open AI Models for Medical Text and Image Analysis
Google has released MedGemma, a pair of open-source generative AI models designed to support medical text and image understanding in healthcare applications. Based on the Gemma 3 architecture, the models are available in two configurations: MedGemma 4B, a multimodal model capable of processing both images and text, and MedGemma 27B, a larger model focused solely on medical text.
-
Microsoft Announces AI Agent and Platform Updates at Build 2025
At its annual developer conference, Build 2025, Microsoft introduced a set of updates focused on expanding the role of AI agents across Windows, GitHub, Azure, and Microsoft 365.
-
Gemma 3n Available for On-Device Inference Alongside RAG and Function Calling Libraries
Google has announced that Gemma 3n is now available in preview on the new LiteRT Hugging Face community, alongside many previously released models. Gemma 3n is a multimodal small language model that supports text, image, video, and audio inputs. It also supports finetuning, customization through retrieval-augmented generation (RAG), and function calling using new AI Edge SDKs.
-
Google DeepMind Unveils AI Coding Agent AlphaEvolve
Google DeepMind published a paper describing their AlphaEvolve coding agent. AlphaEvolve uses LLMs to discover and optimize algorithms across a range of domains, including hardware design, data center operations, and AI training.
-
Apollo GraphQL Launches MCP Server: a New Gateway Between AI Agents and Enterprise APIs
Apollo GraphQL recently launched its MCP Server, enabling businesses to securely and efficiently integrate AI agents with existing APIs using GraphQL. The platform empowers teams to scale innovation and drive faster time-to-value from AI investments by reducing development overhead, improving governance, and accelerating AI feature delivery.
-
Mistral Releases Devstral, an Open-Source LLM for Software Engineering Agents
Mistral AI announced the release of Devstral, a new open-source large language model designed to improve the automation of software engineering workflows, particularly in complex coding environments that require reasoning across multiple files and components.
-
Uber Completes Massive Kubernetes Migration for Microservices and Large-Scale Compute Workloads
Uber has successfully completed a large Kubernetes migration, transitioning its entire compute platform from Apache Mesos to Kubernetes across multiple data centers and cloud environments.
-
Google Enhances LiteRT for Faster On-Device Inference
The new release of LiteRT, formerly known as TensorFlow Lite, introduces a new API to simplify on-device ML inference, enhanced GPU acceleration, support for Qualcomm NPU (Neural Processing Unit) accelerators, and advanced inference features.
-
Redis Returns to Open Source under AGPL License: Is It Too Late?
Redis 8 has recently hit general availability, switching to the AGPLv3 license. A year after leaving its open source roots to challenge cloud service providers and following the birth of Valkey, Redis has rehired its creator and moved back to an open source license.
-
HashiCorp Releases Terraform MCP Server for AI Integration
HashiCorp has released the Terraform MCP Server, an open-source implementation of the Model Context Protocol designed to improve how large language models interact with infrastructure as code.
-
Prime Intellect Releases INTELLECT-2: a 32B Parameter Model Trained via Decentralized Reinforcement
Prime Intellect has released INTELLECT-2, a 32 billion parameter language model trained using fully asynchronous reinforcement learning across a decentralized network of compute contributors. Unlike traditional centralized model training, INTELLECT-2 is developed on a permissionless infrastructure where rollout generation, policy updates, and training are distributed and loosely coupled.
-
Gemma 3 Supports Vision-Language Understanding, Long Context Handling, and Improved Multilinguality
Google’s generative artificial intelligence (AI) model Gemma 3 supports vision-language understanding, long context handling, and improved multi-linguality. In a recent blog post, Google DeepMind and AI Studio teams discussed the new features in Gemma 3. The model also highlights KV-cache memory reduction, a new tokenizer and offers better performance and higher resolution vision encoders.
-
OpenAI Launches Codex Software Engineering Agent Preview
OpenAI has launched Codex, a research preview of a cloud-based software engineering agent designed to automate common development tasks such as writing code, debugging, testing, and generating pull requests. Integrated into ChatGPT, Codex runs each assignment in a secure sandbox environment preloaded with the user's codebase and configured to reflect their development setup.
-
Windsurf Launches SWE-1 Family of Models for Software Engineering
Windsurf has introduced its first set of SWE-1 models, aimed at supporting the full range of software engineering tasks, not limited to code generation. The lineup consists of three models SWE-1, SWE-1-lite, and SWE-1-mini, each designed for specific scenarios.
-
OpenAI’s Stargate Project Aims to Build AI Infrastructure in Partner Countries Worldwide
OpenAI has announced a new initiative called "OpenAI for Countries" as part of its Stargate project, aiming to help nations develop AI infrastructure based on democratic principles. This expansion follows the company's initial $500 billion investment plan for AI infrastructure in the United States.