InfoQ Homepage Memory Content on InfoQ
-
Cloudflare Announces Agent Memory, a Managed Persistent Memory Service for AI Agents
Cloudflare announced Agent Memory in private beta, a managed service that extracts structured memories from AI agent conversations and retrieves them on demand using five-channel parallel retrieval with Reciprocal Rank Fusion. Shared memory profiles let teams of agents access common knowledge. Competitors include Mem0, Zep, LangMem, and Letta.
-
Designing Memory for AI Agents: inside Linkedin’s Cognitive Memory Agent
LinkedIn introduces Cognitive Memory Agent (CMA), generative AI infrastructure layer enabling stateful, context-aware systems. It provides persistent memory across episodic, semantic, and procedural layers, supporting multi-agent coordination, retrieval, and lifecycle management. CMA addresses LLM statelessness and enables production-grade personalization and long-term context in AI applications.
-
Pinterest Reduces Spark OOM Failures by 96% through Auto Memory Retries
Pinterest Engineering cut Apache Spark out-of-memory failures by 96% using improved observability, configuration tuning, and automatic memory retries. Staged rollout, dashboards, and proactive memory adjustments stabilized data pipelines, reduced manual intervention, and lowered operational overhead across tens of thousands of daily jobs.
-
Microsoft Foundry Agent Service Simplifies State Management with Long-Term Memory Preview
Microsoft has launched a public preview of a managed long-term memory store for its Foundry Agent Service. The service automates the extraction, consolidation, and retrieval of user context, providing a native "state layer" that prevents intelligence decay in long-running interactions with AI agents.
-
Go's New Green Tea Garbage Collector May Improve Performance up to 40%
Go 1.25 introduces a new experimental garbage collector that delivers up to 40% faster than the current implementation, bringing a significant performance improvement for GC-heavy workloads.
-
AWS Launches Memory-Optimized EC2 R8i and R8i-flex Instances with Custom Intel Xeon 6 Processors
AWS has launched its eighth-generation Amazon EC2 R8i and R8i-flex instances, powered by custom Intel Xeon 6 processors. Designed for memory-intensive workloads, these instances offer up to 15% better price performance and enhanced memory throughput, making them ideal for real-time data processing and AI applications.
-
Using Machine Learning on Microcontrollers: Decreasing Memory and CPU Usage to Save Power and Cost
According to Eirik Midttun, artificial intelligence (AI) and machine learning (ML) are useful tools for interpreting sensor data, especially when the input is complex, such as vibration, voice, and vision. The main challenges of using machine learning on microcontrollers are the constraints in computing power available and cost-related requirements that come with microcontroller-based designs,
-
Amazon Launches High Memory U7inh EC2 Instance for Enhanced SAP HANA Workloads
AWS has unveiled the Amazon EC2 High Memory U7inh instance, a game-changer for mission-critical in-memory databases like SAP HANA, offering 32 TB of memory and 1,920 vCPUs. Designed with HPE, it doubles the performance of previous models, ensuring seamless integration in AWS. Maximize your SAP workloads in the cloud with enhanced speed and scalability.
-
Java 24 to Reduce Object Header Size and Save Memory
JEP 450 (Compact Object Headers) has been targeted for delivery in JDK 24. This currently experimental feature optimizes heap utilization by shrinking the size of the mandatory object header in HotSpot. This should reduce overall heap size, improve density of deployments, and increase data locality.
-
AWS Launches Graviton4-Powered Memory-Optimized EC2 X8g Instances for High-Memory Workloads
AWS has introduced Graviton4-powered EC2 X8g instances, featuring up to 3 TiB of DDR5 memory and 192 vCPUs for memory-intensive workloads. With 3x improved memory and network bandwidth, plus enhanced security, these instances offer unmatched performance and value. Ideal for databases, big data analytics, and more, X8g sets a new standard in cloud computing.
-
Amazon EC2 R8g Instances with AWS Graviton4 Processors Generally Available
AWS has announced the general availability of Amazon EC2 R8g instances, which use AWS Graviton4 processors. These instances have been available in preview since November 2023 and are designed for memory-intensive workloads such as databases, in-memory caches, and real-time big data analytics.
-
Microsoft’s New Memory Optimized Ebsv5 VM Sizes in Preview Offer More Performance
Microsoft recently announced two additional Memory Optimized Virtual Machines (VM) sizes, E96bsv5 and E112ibsv5, to the Ebsv5 VM family developed with the NVMe protocol providing performance up to 260,000 IOPS and 8,000 MBps remote disk storage throughput.
-
Azure Virtual Machines with Ampere Altra Arm-Based Processors Now Generally Available
Microsoft recently announced the general availability (GA) of virtual machines (VMs) on Azure featuring the Ampere Altra, a processor based on the Arm architecture. In addition, the Arm-based virtual machines can be included in Kubernetes clusters managed using Azure Kubernetes Service (AKS).
-
ImageSharp 2.0.0: the Feature-Packed Release
ImageSharp, one of the most popular .NET image-processing libraries, released version 2 of their library. The release includes major features such as supporting WebP, TIFF and PBM as well adding XMP support with various performance improvements and enhancements for JPEG and PNG formats. This release drops support for .NET Standard 1.3. The update replaces version 1.0.4.
-
Measuring the Environmental Impact of Software and Cloud Services
Software has an influence on the limitation of the service life or the increased energy consumption. It’s possible to measure the environmental impacts that are caused by cloud services. The design of the software architecture determines how much hardware and electrical power is required. Software can be economical or wasteful with hardware resources.