InfoQ Homepage Large language models Content on InfoQ
-
Google Announces State-of-the-Art PaLM 2 Language Model Powering Bard
Google DeepMind recently announced PaLM 2, a large language model (LLM) powering Bard and over 25 other product features. PaLM 2 significantly outperforms the previous version of PaLM on a wide range of benchmarks, while being smaller and cheaper to run.
-
Minecraft Welcomes Its First LLM-Powered Agent
Researchers from Caltech, Stanford, the University of Texas, and NVIDIA have collaboratively developed and released Voyager, an LLM power agent that utilizes GPT-4 to engage in Minecraft gameplay. Voyager demonstrates remarkable capabilities by learning, retaining knowledge, and showcasing exceptional expertise in Minecraft.
-
InfraCopilot, a Conversational Infrastructure-as-Code Editor
Klotho announced InfraCopilot, an infrastructure as a Code (IaC) editor with natural language processing capabilities. The user can chat with InfraCopilot describing their infrastructure needs and it translates these ideas into a low-level architecture. Users can then iterate with incremental high-level and low-level architecture changes.
-
Microsoft Open-Sources 13 Billion Parameter Language and Vision Chatbot LLaVA
Researchers from Microsoft, the University of Wisconsin–Madison, and Columbia University have open-sourced Large Language and Vision Assistant (LLaVA). LLaVA is based on a CLIP image encoder and a LLaMA language decoder, is fine-tuned on a synthetic instruction-following dataset, and achieved state-of-the-art accuracy on the ScienceQA benchmark.
-
Google Previews Studio Bot, a Coding Bot for Android Development
At Google I/O 2023, Google has previewed Studio Bot, an AI-powered coding bot integrated in Android Studio latest version, codenamed Hedgehog. Studio Bot aims to help developers generate code, unit tests, and fix errors.
-
Running Large Language Models Natively on Mobile and Laptops
MLC LLM is a new open source project aimed to enable deploying large language models on a variety of hardware platforms and applications. It additionally includes a framework to optimize model performance for each specific use case.
-
Efficiently Applying LLMs to Transform Semi-Structured Data
LLMs can be an effective way to generate structured data from semi-structured data, although an expensive one. A team of Stanford and Cornell researchers claim to have found a technique to reduce inference costs by 110x while improving inference quality.
-
Stability AI Open-Sources 7B Parameter Language Model StableLM
Stability AI released two sets of pre-trained model weights for StableLM, a suite of large language models (LLM). The models are trained on 1.5 trillion text tokens and are licensed for commercial use under CC BY-SA-4.0.
-
Hugging Face Presents HuggingChat, Open Source Alternative to ChatGPT
HuggingChat is a new AI-powered chatbot available for testing on Hugging Face. HuggingChat is able to carry through many tasks that have made ChatGPT attract lot of interest recently, including drafting articles, solving coding problems, or answering questions.
-
Meta's Toolformer Uses APIs to Outperform GPT-3 on Zero-Shot NLP Tasks
Meta AI Research announced Toolformer, a language model that learns to call APIs to help solve natural language processing (NLP) tasks. Toolformer automatically annotates a training dataset which is used to fine-tune the model and can outperform the much larger GPT-3 model on several zero-shot NLP tasks.
-
HuggingGPT: Leveraging LLMs to Solve Complex AI Tasks with Hugging Face Models
A recent paper by researchers at Zhejiang University and Microsoft Research Asia explores the use of large language models (LLMs) as a controller to manage existing AI models available in communities like Hugging Face.
-
AI, ML & Data News Roundup: OpenAI’s GPT-4, Microsoft’s Semantic Kernel, Meta SAM and BloombergGPT
This week's roundup for April 3rd, 2023, includes the most recent news and information from the fields of data science, machine learning, and artificial intelligence.
-
Meta AI’s Large Language Model with 10x Fewer Parameters
Meta AI recently released a new large language model called Language Large Models Meta AI (LLaMA) that outperforms foundational models such as GPT-3 and is competitive with PaLM, despite having 10 times fewer parameters. LLaMA has better performance in language tasks such as natural questions, common-sense reasoning and mathematical reasoning.
-
Stanford Researchers Present AI Framework to Implement and Validate Complex Algorithms
Parsel, an AI framework created by a group of researchers at Stanford, uses large language model (LLM) reasoning to transform hierarchical functions descriptions in natural language into an implementation in code. Additionally, the researchers maintain, Parsel can be used for robot planning and theorem proving.