InfoQ Homepage AI, ML & Data Engineering Content on InfoQ
-
Google Cloud Unveils AlloyDB AI: Transforming PostgreSQL with Advanced Vector Embeddings and AI
During the recent Google Cloud Next, Google announced AlloyDB AI in preview as an integral part of AlloyDB for PostgreSQL, allowing developers to build generative (gen) Artificial Intelligence (AI) applications leveraging large language models (LLMs) with their real-time operational data through built-in, end-to-end support for vector embeddings.
-
OpenAI Introduces ChatGPT Enterprise for Business-Focus
OpenAI is introducing ChatGPT Enterprise, which provides enterprise-grade protection and privacy, limitless higher-speed GPT-4 access, extended context windows for processing longer inputs, additional data analysis capabilities, customization possibilities, and much more.
-
6 Tracks Not to Miss at QCon San Francisco, October 2-6, 2023: ML, Architecture, Resilience & More!
At InfoQ’s international software development conference, QCon San Francisco (October 2-6) 2023, senior software practitioners driving innovation and change in software development will explore real-world architectures, technology, and techniques to help you solve such challenges.
-
Meta Open-Sources Code Generation LLM Code Llama
Meta recently open-sourced Code Llama, a code generation LLM which is based on the Llama 2 foundation model and carries the same community license. Code Llama was fine-tuned on 500B tokens of code and is available in three model sizes ranging up to 34B parameters. In evaluations on code-generation benchmarks, the model outperformed all other open-source models and is comparable to ChatGPT.
-
Stack Overflow Announced OverflowAI Initiative for Generative AI
Stack Overflow announced the roadmap for the integration of generative AI into their public platform, Stack Overflow for Teams, and brand new product areas, like an IDE integration that brings the vast knowledge of 58 million questions and answers from their community right into the area where developers find focus and get work done.
-
Google Expands Vertex AI Search and Conversation Capabilities
At its Google Cloud Next conference, Google officially introduced new capabilities for its enterprise AI platform, Vertex AI, which aim to enable more advanced user workflows, among other things.
-
Spring AI Provides Integration with OpenAI and Azure OpenAI
The experimental Spring AI project was introduced during the SpringOne conference and allows the creation of AI applications by using common concepts of Spring. Currently the project integrates Azure OpenAI and OpenAI as AI backends. Use cases like content generation, code generation, semantic search and summarization are supported by the project.
-
Stability AI Launches Open Source Chatbot Stable Chat
Stability AI, makers of the image generation AI Stable Diffusion, recently launched Stable Chat, a web-based chat interface for their open-access language model Stable Beluga. At the time of its release, Stable Beluga was the best-performing open large language model (LLM) on the HuggingFace leaderboard.
-
Google Launches Three New Cloud Storage Options: Cloud Storage FUSE, Parallelstore, NetApp Volumes
Google recently launched three new cloud storage options: Cloud Storage FUSE for Artificial Intelligence (AI) applications that require file system semantics, a parallel file system Parallelstore for demanding AI and HPC applications that use GPUs, and NetApp Volumes for enterprise applications running in the cloud.
-
GPT-3.5 Turbo Can Now Be Fine-Tuned for Improved Quality and Performance
OpenAI has made GPT-3.5 Turbo available to developers, with the added bonus of allowing them to customize the model to improve performance for their specific use cases. According to OpenAI, fine-tuning GPT-3.5 Turbo can even outperform base GPT-4 for certain tasks.
-
MetaGPT Leverages Human Collaboration Techniques for Multi-Agent-Based Software Engineering
Created by a team of researchers from Chinese and US universities, MetaGPT is a new LLM-based meta programming framework aiming to enable collaboration in multi-agent systems by leveraging human procedural knowledge to enhance robustness, reduce errors, and engineer software solutions for complex tasks.
-
LMSYS Org Releases Chatbot Arena and LLM Evaluation Datasets
Large Model Systems Organization (LMSYS Org) recently released Chatbot Arena, a comparison platform for large language models (LLMs), where users can pick the better response from a pair of chatbots. LMSYS also released a dataset containing conversations from the Arena as well as a dataset of human annotations of results from evaluating LLMs on the MT-Bench benchmark.
-
Vector Engine for Amazon Opensearch Serverless Now in Preview
AWS announced the preview release of vector storage and search capability within Amazon OpenSearch Serverless. The capability is intended to support machine learning augmented search experiences and generative AI applications.
-
Mountpoint for Amazon S3 Now GA to Access Bucket Like Local File System
During the latest AWS Storage Day event, Amazon announced the general availability of Mountpoint for Amazon S3. The new open-source file client provides through a file interface the elastic storage and throughput of Amazon S3, supporting data transfer at up to 100 Gb/second between each EC2 instance and the object storage.
-
Running Apache Flink Applications on AWS KDA: Lessons Learnt at Deliveroo
Deliveroo introduced Apache Flink into its technology stack for enriching and merging events consumed from Apache Kafka or Kinesis Streams. The company opted to use AWS Kinesis Data Analytics (KDA) service to manage Apache Flink clusters on AWS and shared its experiences from running Flink applications on KDA.