InfoQ Homepage Gemma Content on InfoQ
News
RSS Feed-
Google BigQuery Adds SQL-Native Managed Inference for Hugging Face Models
Google has launched SQL-native managed inference for 180,000+ Hugging Face models in BigQuery. The preview release collapses the ML lifecycle into a unified SQL interface, eliminating the need for separate Kubernetes or Vertex AI management. Key features include automated resource governance via endpoint_idle_ttl and secure identity-based execution using existing data warehouse permissions.
-
Google Introduces TranslateGemma Open Models for Multilingual Translation
Google has released TranslateGemma, a set of open translation models based on the Gemma 3 architecture, offering 4B, 12B, and 27B parameter variants designed to support machine translation across 55 languages and to run on platforms ranging from mobile and edge devices to consumer hardware and cloud accelerators.
-
Gemma 3n Available for On-Device Inference Alongside RAG and Function Calling Libraries
Google has announced that Gemma 3n is now available in preview on the new LiteRT Hugging Face community, alongside many previously released models. Gemma 3n is a multimodal small language model that supports text, image, video, and audio inputs. It also supports finetuning, customization through retrieval-augmented generation (RAG), and function calling using new AI Edge SDKs.
-
Gemma 3 Supports Vision-Language Understanding, Long Context Handling, and Improved Multilinguality
Google’s generative artificial intelligence (AI) model Gemma 3 supports vision-language understanding, long context handling, and improved multi-linguality. In a recent blog post, Google DeepMind and AI Studio teams discussed the new features in Gemma 3. The model also highlights KV-cache memory reduction, a new tokenizer and offers better performance and higher resolution vision encoders.
-
Docker Model Runner Aims to Make it Easier to Run LLM Models Locally
Currently in preview with Docker Desktop 4.40 for macOS on Apple Silicon, Docker Model Runner allows developers to run models locally and iterate on application code using the local models- without disrupting their container-based workflows.
-
Google Launches Gemma 3 1B for Mobile and Web Apps
Requiring a "mere" 529MB, Gemma 3 1B is a small language model (SLM) specifically meant for distribution across mobile and Web apps, where models must download quickly and be responsive to keep user engagement high.