InfoQ Homepage Performance & Scalability Content on InfoQ
-
Eating One's Own Dogfood: GitHub Using Actions and Runners for GitHub.com
To improve how they ship software in a scalable and effective way, GitHub has adopted GitHub Actions for a part of their continuous integration system. In particular, they leveraged the new Actions larger runners to get to run 15,000 CI jobs across 150,000 cores. In the process they also extended larger runners capabilities for all their users.
-
Effective Performance Engineering at Twitter-Scale: Yao Yue at QCon San Francisco
During the second day of QCon San Francisco 2023, Yao Yue, the founder of IOP Systems, presented on performance engineering. In her session Yue discussed the evolving performance engineering in the modern era. For decades, hardware advancements have kept many performance engineers on the sidelines, but now, in a pivotal moment, their skills are more crucial than ever.
-
Hugging Face's Guide to Optimizing LLMs in Production
When it comes to deploying Large Language Models (LLMs) in production, the two major challenges originate from the huge amount of parameters they require and the necessity of handling very long input sequences to represent contextual information. Hugging Face has documented a list of techniques to tackle those hurdles based on their experience serving such models.
-
LinkedIn's Open-Source "iris-message-processor" Achieves 86.6x Faster Escalation Management Speeds
LinkedIn developed a new open-source service called "iris-message-processor" to enhance the performance and reliability of its existing Iris escalation management system. "iris-message-processor" significantly improves processing speeds, being ~4.6x faster under average loads and ~86.6x faster under high loads than its predecessor.
-
Golem Unveils a Resilient Computing Platform for Serverless Workers with WebAssembly Component Model
Recently Golem released its flagship product Golem Cloud, a durable computing platform allowing developers to build and deploy long-running, stateful serverless workers that are resistant to failures, upgrades, and updates. The product is currently in developer preview.
-
Cadence 1.0: Uber Releases Its Scalable Workflow Orchestration Platform
Uber released a major version of its workflow orchestration platform named Cadence after six years in development. Uber and other companies use Cadence to build stateful services at scale using native programming languages.
-
AWS Launches General Availability of Amazon EC2 P5 Instances for AI/ML and HPC Workloads
AWS recently announced the general availability (GA) of Amazon EC2 P5 instances powered by the latest NVIDIA H100 Tensor Core GPUs suitable for users that require high performance and scalability in AI/ML and HPC workloads. The GA is a follow-up to the earlier announcement of the development of the infrastructure.
-
Microsoft Azure Managed Lustre for HPC and AI Workloads Now Generally Available
Microsoft recently announced the general availability (GA) of Azure Managed Lustre, a managed file system for high-performance computing (HPC) and AI workloads.
-
How LinkedIn Serves over 4.8 Million Member Profiles per Second
LinkedIn introduced Couchbase as a centralized caching tier for scaling member profile reads to handle increasing traffic that has outgrown their existing database cluster. The new solution achieved over 99% hit rate, helped reduce tail latencies by more than 60% and costs by 10% annually.
-
New Azure Cosmos DB Features to Boost Performance and Optimize Cost
Microsoft has recently unveiled several new features for Azure Cosmos DB to enhance cost efficiency, boost performance, and increase elasticity. These features are burst capacity, hierarchical partition keys, serverless container storage of 1 TB, and priority-based execution.
-
Datadog Creates Scalable Data Ingestion Architecture
Datadog created a dedicated data ingestion architecture offering exactly-once semantics for their third-generation event store, Husky. The event-driven architecture (EDA) can accommodate bursts in traffic in the multi-tenant platform with reasonable ingestion latency and acceptable operational costs.
-
Real-Time Messaging Architecture at Slack
Slack recently described how it sends millions of messages daily in real-time across the globe. The company provides a comprehensive insight into its architecture, designed to manage real-time messages at scale. It highlights the unique challenges posed by delivering real-time messages across different time zones and regions and how Slack's engineers designed the infrastructure to handle them.
-
Content Discovery at Scale with Hexagons and Elasticsearch at DoorDash
DoorDash recently published an article on how it is solving scaling challenges with content discovery using Elasticsearch and H3, a geospatial indexing system that partitions the world into hexagonal cells.
-
BBC New Serverless Platform Improves Scalability and Performance
One year into the transition to their new WebCore serverless platform, the BBC has started to reap the benefits of an architecture that removes the burden on engineers to solve performance and operational challenges and allows them to focus on the value they deliver to customers.
-
Netflix’s RENO Keeps Experience Consistent across Devices
Netflix has developed the Rapid Event Notification System (RENO) to create a consistent user experience across various platforms and devices. RENO reacts more quickly and consistently than the traditional request/response model to user-generated actions ranging from watching a title to changing profile information.