InfoQ Homepage Performance & Scalability Content on InfoQ
-
Enhancing Reliability Using Service-Level Prioritized Load Shedding: Netflix at QCon SF 2025
At QCon San Francisco, Netflix engineers unveiled their advanced Service-Level-Prioritized Load-Shedding strategy, enhancing reliability during traffic spikes. By prioritizing high-value requests and automating management across microservices, they safeguard user experience and system stability. Key insights stress prioritization, automation, and structured load shedding for optimal resilience.
-
Meta Open Sources OpenZL: a Universal Compression Framework for Structured Data
Meta’s OpenZL changes the way data is compressed by maximizing efficiency for structured datasets, outperforming traditional methods like Zstandard. With a universal decompressor and custom compression plans, it simplifies operational deployment while achieving superior compression ratios and speeds, making it an essential tool for modern data infrastructures.
-
PlanetScale Extends Database Platform to PostgreSQL
PlanetScale has announced the general availability of its managed sharded Postgres service, built for performance and reliability on AWS or Google Cloud. The launch extends PlanetScale's offerings to PostgreSQL users, adding to the company's existing popular MySQL-based platform built on top of Vitess.
-
Cloudflare Achieves 99.99% Warm Start Rate for Workers with 'Shard and Conquer' Consistent Hashing
Cloudflare's innovative "Shard and Conquer" technique revolutionizes its serverless platform by slashing cold start rates by 90%. Utilizing a consistent hash ring, it routes traffic efficiently, keeping Workers warm and minimizing latency. Enhanced for larger applications, this approach ensures optimal performance while accommodating user demands for richer functionalities.
-
Uber Achieves 150M Reads per Second with CacheFront Improvements
Uber has updated its CacheFront architecture to handle over 150 million reads per second. The new design improves consistency and reduces stale reads by integrating Flux for MySQL binlog tailing, enhancing the storage engine, and introducing Cache Inspector for monitoring and optimization.
-
Anthropic Reveals Three Infrastructure Bugs behind Claude Performance Issues
Anthropic recently published a postmortem revealing that three distinct infrastructure bugs intermittently degraded the output quality of its Claude models in recent weeks. While the company states it has now resolved those issues and is modifying its internal processes to prevent similar disruptions, the community highlights the challenges of running the service across three hardware platforms.
-
Datadog Launches Monocle, a Unified Rust-Powered Real-Time Metrics Engine
Datadog has launched Monocle, a new real-time time series storage engine written in Rust. The system unifies the company’s metrics storage infrastructure, delivering higher ingestion throughput and lower query latency while reducing operational complexity. Monocle replaces several generations of storage backends, addressing concurrency challenges and scaling limits that accumulated over time.
-
Pinterest Unifies Engineering Tools with New Pinconsole Platform
Pinterest has introduced PinConsole, a unified internal developer platform (IDP) that centralizes engineering workflows. Built to address fragmented tools for deployment, monitoring, and service management, PinConsole provides a consistent layer that lets engineers focus on business logic instead of infrastructure complexity.
-
Impulse, Airbnb’s New Framework for Context-Aware Load Testing
Airbnb has developed Impulse, an internal load testing framework to improve microservice reliability and performance. It enables distributed, large-scale testing and lets teams run self-service, context-aware load tests integrated with CI pipelines. By simulating production-like traffic, Impulse helps engineers identify bottlenecks and errors before changes reach production.
-
Uber Eats Scales Catalog Management from Restaurants to Retail with INCA Framework
Uber Eats introduced INCA (Inventory and Catalog), a scalable system to handle vast product catalogs from supermarkets, pharmacies, and retail partners. Unlike the earlier restaurant-focused setup built for low SKUs and simple pass-through data, INCA supports large-scale inventories, rich metadata, and compliance needs essential for retail operations.
-
AWS Lambda Response Streaming Increases Payload Limit to 200 MB
AWS has revolutionized Lambda with an increased response streaming payload limit from 20 MB to 200 MB. This enhancement allows developers to stream larger data sets, improving Time to First Byte performance. By simplifying response handling and eliminating complex workarounds, AWS empowers developers to deliver rich content seamlessly, transforming serverless applications.
-
Amazon DocumentDB Serverless: Auto-Scaling Database Solution for Variable Workloads
AWS has launched Amazon DocumentDB Serverless, an auto-scaling database solution compatible with MongoDB, tailored for variable workloads. While marketed as "serverless," it functions more like auto-scaling, charging from $30/month. Ideal for enterprises and SaaS vendors, it adeptly handles spikes in demand, particularly for AI-driven applications.
-
Grab Switches from SQS and Redis to Temporal for Its Subscription Platform
Grab based the new architecture for GrabUnlimited on Temporal. The company enhanced user experience and reduced production incidents by 80% for its subscription platform, which serves millions of users. The new architecture significantly improved robustness and scalability, addressing a range of issues with the previous solution.
-
Apple Completes Migration of Key Ecosystem Service to Swift, Gains 40% Performance Uplift
Apple has migrated its global Password Monitoring service from Java to Swift, achieving a 40% increase in throughput and significantly reducing memory usage—freeing up nearly 50% of previously allocated Kubernetes capacity.
-
AWS Promotes Responsible AI in the Well-Architected Generative AI Lens
AWS announced the availability of the new Well-Architected Generative AI Lens, focused on providing best practices for designing and operating generative AI workloads. The lens is aimed at organizations delivering robust and cost-effective generative AI solutions on AWS. The document offers cloud-agnostic best practices, implementation guidance and links to additional resources.