InfoQ Homepage Large language models Content on InfoQ
-
NVIDIA Introduces Metropolis Microservices for Jetson to Run AI Apps at the Edge
NVIDIA has expanded its Nvidia Metropolis Microservices Cloud-based AI solution to run on the NVIDIA Jetson IoT embedded platform, including support for video streaming and AI-based perception.
-
Meta Releases Code Generation Model Code Llama 70B, Nearing GPT-3.5 Performance
Code Llama 70B is Meta's new code generation AI model. Thanks to its 70 billion parameters, it is "the largest and best-performing model in the Code Llama family", Meta says.
-
Stability AI Releases 1.6 Billion Parameter Language Model Stable LM 2
Stability AI released two sets of pre-trained model weights for Stable LM 2, a 1.6B parameter language model. Stable LM 2 is trained on 2 trillion tokens of text data from seven languages and can be run on common laptop computers.
-
LeftoverLocals May Leak LLM Responses on Apple, Qualcomm, and AMD GPUs
Security firm Trail of Bits disclosed a vulnerability allowing malicious actors to recover data from GPU local memory on Apple, Qualcomm, AMD, and Imagination GPUs. Dubbed LeftoverLocals, the vulnerability affects any application using the GPU, including Large Language Models (LLMs) and machine learning (ML) models.
-
Mistral AI's Open-Source Mixtral 8x7B Outperforms GPT-3.5
Mistral AI recently released Mixtral 8x7B, a sparse mixture of experts (SMoE) large language model (LLM). The model contains 46.7B total parameters, but performs inference at the same speed and cost as models one-third that size. On several LLM benchmarks, it outperformed both Llama 2 70B and GPT-3.5, the model powering ChatGPT.
-
LLMs May Learn Deceptive Behavior and Act as Persistent Sleeper Agents
AI researchers at OpenAI competitor Anthropic trained proof-of-concept LLMs showing deceptive behavior triggered by specific hints in the prompts. Furthermore, they say, once deceptive behavior was trained into the model, there was no way to circumvent it using standard techniques.
-
Google Announces Video Generation LLM VideoPoet
Google Research recently published their work on VideoPoet, a large language model (LLM) that can generate video. VideoPoet was trained on 2 trillion tokens of text, audio, image, and video data, and in evaluations by human judges its output was preferred over that of other models.
-
OpenAI GPT Store is a Nascent Marketplace for Custom ChatGPTs
OpenAI has started rolling out its new GPT Store, announced a few months ago along with GPTs, to provide a mechanism for ChatGPT Plus, Team and Enterprise users to share custom ChatGPT-based chatbots they create.
-
OpenAI Adopts Preparedness Framework for AI Safety
OpenAI recently published a beta version of their Preparedness Framework for mitigating AI risks. The framework lists four risk categories and definitions of risk levels for each, as well as defining OpenAI's safety governance procedures.
-
Griffin 2.0: Instacart Revamps Its Machine Learning Platform
Instacart created the next-generation platform based on experiences using the original Griffin machine-learning platform. The company wanted to improve user experience and help manage all ML workloads. The revamped platform leverages the latest developments in MLOps and introduces new capabilities for current and future applications.
-
Quarkus LangChain4J Extension Allows Developers to Integrate LLMs in Their Quarkus Applications
Inspired by the presentation “Java Meets AI” at Devoxx BE 2023, the Quarkus team started working on an extension based on the LangChain4J library, the Java re-implementation of the langchain library. This would allow developers to integrate LLMs Quarkus applications. The current is version, 0.5. The extension was built using Quarkus' usual declarative style, resembling the REST client.
-
OpenAI Publishes GPT Prompt Engineering Guide
OpenAI recently published a guide to Prompt Engineering. The guide lists six strategies for eliciting better responses from their GPT models, with a particular focus on examples for their latest version, GPT-4.
-
Microsoft Announces Small Language Model Phi-2
Microsoft Research announced Phi-2, a 2.7 billion-parameter Transformer-based language model. Phi-2 is trained on 1.4T tokens of synthetic data generated by GPT-3.5 and outperforms larger models on a variety of benchmarks.
-
Microsoft Announces Copilot for Azure, an AI Assistant for IT Professionals
Microsoft has introduced Copilot for Azure, an AI-based tool designed to enhance the management and operation of cloud infrastructure and services. It leverages the capabilities of large language models (LLMs) with Azure's Resource Model to provide a comprehensive understanding and handling of Azure's functionalities, spanning from cloud services to edge technology.
-
JetBrains Launches AI Assistant Integrated in its 2023.3 Release IDEs
JetBrains refreshes all of its IDEs in the last release of the year and promotes its integrated AI Assistant out of preview into general availability for paying customers. Besides its strong integration with the IDEs, JetBrains AI Assistant tries to stand out thanks to its support for multiple LLMs.