InfoQ Homepage Artificial Intelligence Content on InfoQ
-
DoorDash Builds LLM Conversation Simulator to Test Customer Support Chatbots at Scale
DoorDash engineers built a simulation and evaluation flywheel to test large language model customer support chatbots at scale. The system generates multi-turn synthetic conversations using historical transcripts and backend mocks, evaluates outcomes with an LLM-as-judge framework, and enables rapid iteration on prompts, context, and system design before production deployment.
-
Advance Your Socio-Technical Architecture Skills with InfoQ’s New Online Cohorts
Enhance your architectural leadership with InfoQ’s new online cohorts starting April 15, May 7, and June 10, 2026. Led by Luca Mezzalira, this 5-week program focuses on socio-technical skills like ADRs, platform engineering, and AI trade-offs. Senior practitioners can apply frameworks to live projects, earn ICSAET certification, and contribute to the InfoQ community.
-
AWS Launches Strands Labs for Experimental AI Agent Projects
Amazon Web Services has introduced Strands Labs, a new GitHub organization created to host experimental projects related to agent-based AI development.
-
Running Ray at Scale on AKS
The Azure Kubernetes Service (AKS) team at Microsoft has shared guidance for running Anyscale's managed Ray service at scale. They focus on three key issues: GPU capacity limits, scattered ML storage, and problems with credential expiry.
-
AI-Powered Bot Compromises GitHub Actions Workflows across Microsoft, DataDog, and CNCF Projects
AI-powered bot hackerbot-claw exploited GitHub Actions workflows across Microsoft, DataDog, and CNCF projects over 7 days using 5 attack techniques. Bot achieved RCE in 5 of 7 targets, stole GitHub token from awesome-go (140k stars), and fully compromised Aqua Security's Trivy. Campaign included first documented AI-on-AI attack where bot attempted prompt injection against Claude Code.
-
GitLab Suggests AI Can Detect Vulnerabilities But it's AI Governance That Determines Risk
Artificial intelligence is rapidly transforming how software vulnerabilities are detected, but questions about who governs the risks AI exposes, and how those risks are acted on, are becoming increasingly urgent, according to a new blog post by GitLab.
-
Cloudflare Releases Experimental Next.js Alternative Built with AI Assistance
Cloudflare released vinext, an experimental Next.js reimplementation built on Vite by one engineer, with AI guidance over one week, for $1,100. Early benchmarks show 4.4x faster builds, but Cloudflare cautions it's untested at scale. Missing static pre-rendering. HN reaction skeptical, noting Vite does the heavy lifting. Already running on CIO.gov despite experimental status.
-
OpenAI Secures AWS Distribution for Frontier Platform in $110B Multi-Cloud Deal
OpenAI's $110B funding includes AWS as the exclusive third-party distributor for the Frontier agent platform, introducing an architectural split: Azure retains stateless API exclusivity; AWS gains stateful runtime environments via Bedrock. Deal expands the existing $38B AWS agreement by $100B and commits 2GW of Trainium capacity.
-
QCon AI Boston’s Early Program Focuses on the Engineering Work behind Production AI
As teams move AI from pilots to production, the hard problems shift from demos to dependability. The first confirmed talks for QCon AI Boston (June 1–2) focus on context engineering, agent explainability, reasoning beyond basic RAG, evaluation, governance, and platform infrastructure needed to run AI reliably under real-world constraints.
-
GitHub Data Shows AI Tools Creating "Convenience Loops" That Reshape Developer Language Choices
GitHub’s Octoverse 2025 report reveals a "convenience loop" where AI coding assistants drive language choice. TypeScript’s 66% surge to the #1 spot highlights a shift toward static typing, as types provide essential guardrails for LLMs. While Python leads in AI research, the industry is consolidating around stacks that minimize AI friction, creating a barrier for new, niche languages.
-
GitHub's Points to a More Global, AI-Challenged Open Source Ecosystem in 2026
GitHub has released its yearly look at open-source trends. They used data from the Octoverse 2025 report to help the open-source community get ready for the coming year. The picture that emerges is one of extraordinary scale and the structural strains that come with it.
-
Microsoft Open Sources Evals for Agent Interop Starter Kit to Benchmark Enterprise AI Agents
Microsoft's Evals for Agent Interop is an open-source starter kit that enables developers to evaluate AI agents in realistic work scenarios. It features curated scenarios, datasets, and an evaluation harness to assess agent performance across tools like email and calendars.
-
Microsoft Agent Framework RC Simplifies Agentic Development in .NET and Python
Microsoft has announced that the Microsoft Agent Framework has reached Release Candidate status for both .NET and Python. This milestone indicates that the API surface is stable and feature-complete for what is planned in version 1.0, setting the stage for an upcoming general availability release.
-
Google Brings its Developer Documentation into the Age of AI Agents
Google has announced the public preview of the Developer Knowledge API. It comes with a Model Context Protocol (MCP) server. This gives AI development tools a simple, machine-readable way to reach Google's official developer documentation.
-
AI "Vibe Coding" Threatens Open Source as Maintainers Face Crisis
Daniel Stenberg shut down cURL's bug bounty after AI submissions hit 20%. Mitchell Hashimoto banned AI code from Ghostty. Steve Ruiz closed all external PRs to tldraw. Economic research shows "vibe coding" weakens the user engagement that sustains open source. As developers delegate to AI agents, documentation visits and bug reports collapse—threatening the ecosystem's viability.