InfoQ Homepage AI, ML & Data Engineering Content on InfoQ
-
Faster Container Startup in AWS Fargate with Seekable OCI
AWS recently announced support for Seekable Open Container Initiative (SOCI) with AWS Fargate. SOCI enables lazy-loading of images which reduces container start-up times on Amazon Elastic Container Service (ECS). In addition, AWS released two tools, SOCI Index Builder and SOCI Snapshotter, to help its customers take advantage of this new functionality.
-
AWS Launches General Availability of Amazon EC2 P5 Instances for AI/ML and HPC Workloads
AWS recently announced the general availability (GA) of Amazon EC2 P5 instances powered by the latest NVIDIA H100 Tensor Core GPUs suitable for users that require high performance and scalability in AI/ML and HPC workloads. The GA is a follow-up to the earlier announcement of the development of the infrastructure.
-
AWS Introduces a Generative AI-Powered Clinical Documentation Tool with HealthScribe in Preview
AWS recently announced a new HIPAA-eligible service called AWS HealthScribe in a preview that uses speech recognition and generative AI (powered by Amazon Bedrock) to generate clinical documentation.
-
Researchers Publish Attack Algorithm for ChatGPT and Other LLMs
Researchers from Carnegie Mellon University (CMU) have published LLM Attacks, an algorithm for constructing adversarial attacks on a wide range of large language models (LLMs), including ChatGPT, Claude, and Bard. The attacks are generated automatically and are successful 84% of the time on GPT-3.5 and GPT-4, and 66% of the time on PaLM-2.
-
Amazon Bedrock Unveils New Agents Feature
Amazon announced the release of agents for Amazon Bedrock, a new feature that allows developers to quickly create fully managed agents. By performing API calls to enterprise systems, agents for Amazon Bedrock speed up the release of generative AI applications that can manage and carry out activities.
-
Amazon Aurora PostgreSQL Adds pgvector to Support Embeddings from Generative AI
AWS recently announced that the PostgreSQL-compatible edition of Amazon Aurora now supports pgvector for vector storage and similarity search. Aurora is the latest managed PostgreSQL database supporting the open-source extension to store and search embeddings from machine learning models.
-
Meta Open Sources New AI Model Llama 2
Meta is open-sourcing its large language model, Llama 2. The model’s code and weights are being made available free of charge for both research and commercial use. Llama 2 is the result of the expanded partnership between Meta and Microsoft, with the latter being the preferred partner for the new model.
-
Grab Reduces Traffic Cost for Kafka Consumers on AWS to Zero
Grab took advantage of the ability of Apache Kafka consumers to connect to the broker node in the same availability zone (AZ) introduced in Kafka 2.3 and reduced the traffic cost on AWS to zero for reconfigured consumers. The change has substantially reduced overall infrastructure costs for running Apache Kafka on AWS.
-
LangChain - Working with Large Language Models, Made Easy
LangChain is a framework that simplifies working with large language models (LLMs) such as OpenAI GPT4 or Google PaLM by providing abstractions for common use cases. It supports both JavaScript and Python.
-
Pfizer Uses Serverless Architecture on AWS to Scale Processing of Digital Biomarkers
Pfizer upgraded the serverless architecture for processing digital biomarker data at scale to make it more flexible and configurable. They created a framework that uses a file processing pipeline built with AWS Step Functions and other serverless services, as well as a custom Python package for data ingestion and processing.
-
Meta's Voicebox Outperforms State-of-the-Art Models on Speech Synthesis
Meta recently announced Voicebox, a speech generation model that can perform text-to-speech (TTS) synthesis in six languages, as well as edit and remove noise from speech recordings. Voicebox is trained on over 50k hours of audio data and outperforms previous state-of-the-art models on several TTS benchmarks.
-
AI, ML, Data Engineering News Round up: Claude 2, Stable Doodle, CM3leon, Llama 2, Azure and xAI
The most recent update, covering developments from July 17th, 2023, showcases significant progress and announcements in the fields of data science, machine learning, and artificial intelligence. This week's focus centers on Anthropic, Stability AI, Microsoft, Meta and xAI.
-
Grammarly Replaces its in-House Data Lake with Databricks Platform Using Medallion Architecture
Grammarly adopted the medallion architecture while migrating from their in-house data lake, storing Parquet files in AWS S3, to the Delta Lake lakehouse. The company created a new event store for over 6000 event types from 40 internal and external clients and, in the process, improved data quality and reduced the data-delivery time by 94%.
-
GitHub Details Key Prompt Engineering Practices Used to Build Copilot
Prompt engineering is key to creating effective LLM-based applications and does not require to have a PhD in machine learning or generative AI, say GitHub engineers Albert Ziegler and John Berryman, who also shared the lessons they learned developing GitHub Copilot.
-
JetBrains Unveils AI Assistant for IntelliJ-Based IDEs and .NET Tools
JetBrains, the software development company known for creating the IntelliJ IDEA, has announced the introduction of a new AI Assistant in its Early Access Program (EAP) builds for all IntelliJ-based IDEs and .NET tools. This significant addition is aimed at transforming the landscape of software development tools by integrating generative AI and large language models into JetBrains' products.