InfoQ Homepage Artificial Intelligence Content on InfoQ
-
Buoyant Announces MCP Support for Linkerd, Extending Service Mesh Capabilities to Agentic AI Traffic
Buoyant, the company behind the open-source Linkerd service mesh, announced that Linkerd now supports the Model Context Protocol (MCP), making it the first service mesh to natively manage, secure, and observe agentic AI traffic in Kubernetes environments.
-
LinkedIn’s Migration Journey to Serve Billions of Users by Nishant Lakshmikanth at QCon SF
Engineering Manager Nishant Lakshmikanth showcased LinkedIn's transformation at QCon SF 2025, detailing a shift from legacy batch-based systems to a real-time architecture. By decoupling recommendations and leveraging dynamic scoring techniques, LinkedIn achieved a 90% reduction in offline costs, enhanced session-level freshness, and improved member engagement while future-proofing its platform.
-
SAM 3 Introduces a More Capable Segmentation Architecture for Modern Vision Workflows
Meta has released SAM 3, the latest version of its Segment Anything Model and the most substantial update to the project since its initial launch. Built to provide more stable and context-aware segmentation, the model offers improvements in accuracy, boundary quality, and robustness to real-world scenes, aiming to make segmentation more reliable across research and production systems.
-
Sauce Labs Launches AI Tool for Faster Test Analysis
Sauce Labs has launched Sauce AI for Insights, an AI-driven tool that accelerates test analysis by providing natural-language explanations, visual summaries and faster root cause detection. The company claims that it reduces debugging time, improves release readiness, and addresses the growing complexity of test data.
-
New Token-Oriented Object Notation (TOON) Hopes to Cut LLM Costs by Reducing Token Consumption
The recently released Token-Oriented Object Notation (TOON) aims to be a schema-aware alternative to JSON that significantly reduces token consumption at a similar level of accuracy. While the existence and importance of token saved depend on the data shape, some benchmarks show TOON may use in some cases 40% fewer tokens than JSON, possibly resulting in LLM and inference cost savings.
-
Olmo 3 Release Provides Full Transparency into Model Development and Training
The Allen Institute for AI has unveiled Olmo 3, an open-source language model family that empowers developers with full access to the model lifecycle, from training datasets to checkpoints. Featuring reasoning-focused variants and robust tools for post-training modifications, Olmo 3 promotes transparency, experimentation, and community collaboration, driving innovations in AI.
-
QConSF 2025: Humans in the Loop: Engineering Leadership in a Chaotic Industry
At QCon SF 2025, Michelle Brush of Google explored the evolving landscape of software engineering in her keynote “Humans in the Loop: Engineering Leadership in a Chaotic Industry.” She highlighted the complexities engineers face amid automation and AI, stressing the importance of conscious competence, higher-level problem-solving, and effective leadership in navigating today's challenges.
-
Google Launches Code Wiki, an AI-Driven System for Continuous, Interactive Code Documentation
Google has introduced Code Wiki, a new platform designed to keep software documentation continuously synchronized with the code it describes. The system generates a structured wiki for each repository, automatically updates it after every change, and powers an integrated chat interface that understands the entire codebase.
-
QConSF 2025 - Developing Claude Code at Anthropic at AI Speed
At QCon San Francisco 2025, Adam Wolff showcased Claude Code at Anthropic, where AI powers 90% of production code. With a focus on speed over planning, Claude Code's design evolved through experimentation, addressing challenges like Unicode issues and shell command bottlenecks. Discover successful iterations and lessons learned in real-time software development.
-
Google Announces Gemini 3
Google's Gemini 3, unveiled on November 18, 2025, sets a new standard for multimodal AI, integrating seamlessly across platforms like Search and Vertex AI. With capabilities for text, code, and rich media, it empowers both consumer and enterprise applications. Gemini 3 Pro and its advanced Deep Think mode enhance reasoning and task execution, revolutionizing workflows and analytics.
-
Cloudflare Workflows Adds Python Support for Durable AI Pipelines
Innovative Cloudflare Workflows now supports both TypeScript and Python, enabling developers to orchestrate complex applications seamlessly. With durable execution and state persistence, it simplifies the development of robust data pipelines and AI/ML models. Experience enhanced concurrency and intuitive design, making orchestration effortless for Python enthusiasts.
-
First Keynote at QCon San Francisco 2025: Reducing Friction by Nicole Forsgren
At QCon SF 2025, Dr. Nicole Forsgren highlighted how AI accelerates code generation but reveals deployment bottlenecks, urging a strategic pivot to optimizing Developer Experience (DevEx). With 31% of developer time lost to friction, focusing on effective feedback loops, flow state, and cognitive load management is vital for competitive survival and retention.
-
New IBM Granite 4 Models to Reduce AI Costs with Inference-Efficient Hybrid Mamba-2 Architecture
IBM recently announced the Granite 4.0 family of small language models. The model family aims to deliver faster speeds and significantly lower operational costs at acceptable accuracy vs. larger models. Granite 4.0 features a new hybrid Mamba/transformer architecture that largely reduces memory requirements, enabling Granite to run on significantly cheaper GPUs and at significantly reduced costs.
-
KubeCon NA 2025 - Erica Hughberg and Alexa Griffith on Tools for the Age of GenAI
Generative AI technologies need to support new workloads, traffic patterns, and infrastructure demands and require a new set of tools for the age of GenAI. Erica Hughberg from Tetrate and Alexa Griffith from Bloomberg spoke last week at KubeCon + CloudNativeCon North America 2025 Conference about what it takes to build GenAI platforms capable of serving model inference at scale.
-
Anthropic Adds Sandboxing and Web Access to Claude Code for Safer AI-Powered Coding
Anthropic released sandboxing capabilities for Claude Code and launched a web-based version of the tool that runs in isolated cloud environments. The company introduced these features to address security risks that arise when Claude Code writes, tests, and debugs code with broad access to developer codebases and files.