InfoQ Homepage Large language models Content on InfoQ
-
OpenAI GPT Store is a Nascent Marketplace for Custom ChatGPTs
OpenAI has started rolling out its new GPT Store, announced a few months ago along with GPTs, to provide a mechanism for ChatGPT Plus, Team and Enterprise users to share custom ChatGPT-based chatbots they create.
-
OpenAI Adopts Preparedness Framework for AI Safety
OpenAI recently published a beta version of their Preparedness Framework for mitigating AI risks. The framework lists four risk categories and definitions of risk levels for each, as well as defining OpenAI's safety governance procedures.
-
Griffin 2.0: Instacart Revamps Its Machine Learning Platform
Instacart created the next-generation platform based on experiences using the original Griffin machine-learning platform. The company wanted to improve user experience and help manage all ML workloads. The revamped platform leverages the latest developments in MLOps and introduces new capabilities for current and future applications.
-
Quarkus LangChain4J Extension Allows Developers to Integrate LLMs in Their Quarkus Applications
Inspired by the presentation “Java Meets AI” at Devoxx BE 2023, the Quarkus team started working on an extension based on the LangChain4J library, the Java re-implementation of the langchain library. This would allow developers to integrate LLMs Quarkus applications. The current is version, 0.5. The extension was built using Quarkus' usual declarative style, resembling the REST client.
-
OpenAI Publishes GPT Prompt Engineering Guide
OpenAI recently published a guide to Prompt Engineering. The guide lists six strategies for eliciting better responses from their GPT models, with a particular focus on examples for their latest version, GPT-4.
-
Microsoft Announces Small Language Model Phi-2
Microsoft Research announced Phi-2, a 2.7 billion-parameter Transformer-based language model. Phi-2 is trained on 1.4T tokens of synthetic data generated by GPT-3.5 and outperforms larger models on a variety of benchmarks.
-
Microsoft Announces Copilot for Azure, an AI Assistant for IT Professionals
Microsoft has introduced Copilot for Azure, an AI-based tool designed to enhance the management and operation of cloud infrastructure and services. It leverages the capabilities of large language models (LLMs) with Azure's Resource Model to provide a comprehensive understanding and handling of Azure's functionalities, spanning from cloud services to edge technology.
-
JetBrains Launches AI Assistant Integrated in its 2023.3 Release IDEs
JetBrains refreshes all of its IDEs in the last release of the year and promotes its integrated AI Assistant out of preview into general availability for paying customers. Besides its strong integration with the IDEs, JetBrains AI Assistant tries to stand out thanks to its support for multiple LLMs.
-
Microsoft's Orca 2 LLM Outperforms Models That Are 10x Larger
Microsoft Research released its Orca 2 LLM, a fine-tuned version of Llama 2 that performs as well as or better than models that contain 10x the number of parameters. Orca 2 uses a synthetic training dataset and a new technique called Prompt Erasure to achieve this performance.
-
Anthropic Announces Claude 2.1 LLM with Wider Context Window and Support for AI Tools
According to Anthropic, the newest version of Claude delivers many “advancements in key capabilities for enterprises—including an industry-leading 200K token context window, significant reductions in rates of model hallucination, system prompts and our new beta feature: tool use.” Anthropic also announced reduced pricing to improve cost efficiency for our customers across models.
-
xAI Introduces Large Language Model Grok
xAI, the AI company founded by Elon Musk, recently announced Grok, a large language model. Grok can access current knowledge of the world via the X platform and outperforms other LLMs of comparable size, including GPT-3.5, on several benchmarks.
-
AI Researchers Improve LLM-Based Reasoning by Mimicking Learning from Mistakes
Researchers from Microsoft, Peking University, and Xi’an Jiaotong University claim to have developed a technique to improve large language models' (LLMs) ability to solve math problems by replicating how humans learn from their own mistakes.
-
Jina AI's Open-Source Embedding Model Outperforms OpenAI's Ada
Multimodal AI company Jina AI recently released jina-embeddings-v2, a sentence embedding model. The model supports context lengths up to 8192 tokens and outperforms OpenAI's text-embedding-ada-002 on several embedding benchmarks.
-
Google Open-Sources AI Fine-Tuning Method Distilling Step-by-Step
A team from the University of Washington and Google Research recently open-sourced Distilling Step-by-Step, a technique for fine-tuning smaller language models. Distilling Step-by-Step requires less training data than standard fine-tuning and results in smaller models that can outperform few-shot prompted large language models (LLMs) that have 700x the parameters.
-
Google DeepMind Announces LLM-Based Robot Controller RT-2
Google DeepMind recently announced Robotics Transformer 2 (RT-2), a vision-language-action (VLA) AI model for controlling robots. RT-2 uses a fine-tuned LLM to output motion control commands. It can perform tasks not explicitly included in its training data and improves on baseline models by up to 3x on emergent skill evaluations.