InfoQ Homepage Google Content on InfoQ
-
Google Labs Introduces Opal, a Visual Platform for Creating AI Mini-Apps
Google Labs has introduced Opal, an experimental no-code tool that enables users to create AI-powered mini-applications through natural language descriptions and a visual workflow editor.
-
Google Cloud Introduces Non-Disruptive Cloud Storage Bucket Relocation
Google Cloud's innovative Cloud Storage bucket relocation feature enables seamless, non-disruptive data migration across regions while preserving metadata and minimizing application downtime. Maintain governance, enhance lifecycle management, and leverage insights for optimized storage—all without altering access paths. Experience efficient, low-latency solutions tailored for your needs.
-
Gemma 3n Introduces Novel Techniques for Enhanced Mobile AI Inference
Launched in early preview last May, Gemma 3n is now officially available. It targets mobile-first, on-device AI applications, using new techniques designed to increase efficiency and improve performance, such as per-layer embeddings and transformer nesting.
-
Google Launches Gemini CLI: Open-Source Terminal AI Agent for Developers
Google has released Gemini CLI, a new open-source AI command-line interface that brings the full capabilities of its Gemini 2.5 Pro model directly into developers’ terminals. Designed for flexibility, transparency, and developer-first workflows, Gemini CLI provides high-performance, natural language AI assistance through a lightweight, locally accessible interface.
-
Google's Agent2Agent Protocol Enters the Linux Foundation
Recently open-sourced by Google, the Agent2Agent protocol is now part of the Linux Foundation, along with its accompanying SDKs and developer tools.
-
Google Previews Gemini's Agent Mode in Android Studio Narwhal
Google has announced the integration of Gemini in Android Studio's Agent Mode into the latest canary release of Android Studio, Android Studio Narwhal preview. According to Google, the new Agent Mode is designed to handle multi-step development tasks that span across several files.
-
Android 16 Introduces Wide-Ranging Features and Changes
Google has released Android 16, now available for supported Pixel devices. The new release introduces a wide range of new APIs and changes, some of which affect existing apps and require developers to update their code. Material 3 Expressive, the new paradigm for Android user interfaces, is not yet included and will roll out later this year.
-
Google Cloud Run Now Offers Serverless GPUs for AI and Batch Processing
Google Cloud has launched NVIDIA GPU support for Cloud Run, enhancing its serverless platform with scalable, cost-efficient GPU resources. This upgrade enables rapid AI inference and batch processing, featuring pay-per-second billing and automatic scaling to zero. Developers can access seamless GPU support easily, making advanced AI applications faster and more accessible.
-
Google Brings Gemini Nano to ML Kit with New On-Device GenAI APIs
The new GenAI APIs recently added to ML Kit enable developers to use Gemini Nano for on-device inference in Android apps, supporting features like summarization, proofreading, rewriting, and image description.
-
Google Releases LMEval, an Open-Source Cross-Provider LLM Evaluation Tool
LMEval aims to help AI researchers and developers compare the performance of different large language models. Designed to be accurate, multimodal, and easy to use, LMEval has already been used to evaluate major models in terms of safety and security.
-
Google Releases MedGemma: Open AI Models for Medical Text and Image Analysis
Google has released MedGemma, a pair of open-source generative AI models designed to support medical text and image understanding in healthcare applications. Based on the Gemma 3 architecture, the models are available in two configurations: MedGemma 4B, a multimodal model capable of processing both images and text, and MedGemma 27B, a larger model focused solely on medical text.
-
Gemma 3n Available for On-Device Inference Alongside RAG and Function Calling Libraries
Google has announced that Gemma 3n is now available in preview on the new LiteRT Hugging Face community, alongside many previously released models. Gemma 3n is a multimodal small language model that supports text, image, video, and audio inputs. It also supports finetuning, customization through retrieval-augmented generation (RAG), and function calling using new AI Edge SDKs.
-
Google Enhances LiteRT for Faster On-Device Inference
The new release of LiteRT, formerly known as TensorFlow Lite, introduces a new API to simplify on-device ML inference, enhanced GPU acceleration, support for Qualcomm NPU (Neural Processing Unit) accelerators, and advanced inference features.
-
Gemma 3 Supports Vision-Language Understanding, Long Context Handling, and Improved Multilinguality
Google’s generative artificial intelligence (AI) model Gemma 3 supports vision-language understanding, long context handling, and improved multi-linguality. In a recent blog post, Google DeepMind and AI Studio teams discussed the new features in Gemma 3. The model also highlights KV-cache memory reduction, a new tokenizer and offers better performance and higher resolution vision encoders.
-
Anthropic Introduces Web Search Functionality for Claude Models
Anthropic has announced the addition of web search capabilities to its Claude models, available via the Anthropic API. This update enables Claude to access current information from the web, allowing developers to create applications and AI agents that provide up-to-date insights.