InfoQ Homepage Large language models Content on InfoQ
-
Alibaba Releases Two Open-Weight Language Models for Math and Voice Chat
Alibaba released two open-weight language model families: Qwen2-Math, a series of LLMs tuned for solving mathematical problems; and Qwen2-Audio, a family of multi-modal LLMs that can accept voice or text input. Both families are based on Alibaba's Qwen2 LLM series, and all but the largest version of Qwen2-Math are available under the Apache 2.0 license.
-
Grok-2 Beta Version Released on X Platform
The Grok-2 language model has been released in beta on the X platform, introduced alongside Grok-2 mini. The model, tested under the designation "sus-column-r" on the LMSYS leaderboard, has achieved a higher Elo Score compared to Claude 3.5 Sonnet and GPT-4-Turbo. Grok-2 mini, a smaller variant, is also part of the beta release, designed to offer a balance between speed and performance.
-
Microsoft Launches Open-Source Phi-3.5 Models for Advanced AI Development
Microsoft launched three new open-source AI models in its Phi-3.5 series: Phi-3.5-mini-instruct, Phi-3.5-MoE-instruct, and Phi-3.5-vision-instruct. Available under a permissive MIT license, these models offer developers powerful tools for various tasks, including reasoning, multilingual processing, and image and video analysis.
-
Microsoft Releases Prompty: New VS Code Extension for Integrating LLMs into .NET Development
Microsoft has released a new Visual Studio Code extension called Prompty, designed to integrate Large Language Models (LLMs) like GPT-4o directly into .NET development workflows. This free tool aims to simplify the process of adding AI-driven capabilities to applications. The official release post includes a practical example demonstrating how Prompty can be used in real-world scenarios.
-
Apple Unveils Apple Foundation Models Powering Apple Intelligence
Apple published the details of their new Apple Foundation Models (AFM), a family of large language models (LLM) that power several features in their Apple Intelligence suite. AFM comes in two sizes: a 3B parameter on-device version and a larger cloud-based version.
-
LLMs and Agents as Team Enablers
Eric Naiburg and Birgitta Böckeler published articles on the benefits and challenges of using AI as a multiplier in dev teams. We report on their insights for scenarios such as simplifying the germane cognitive load of a domain, automating code migrations, and coaching scrum masters on team facilitation. We also cover Böckeler's experiments with using LLMs to onboard onto a complex project.
-
Meta's Research SuperCluster for Real-Time Voice Translation AI Systems
A recent article from Engineering at Meta reveals how the company is building Research SuperCluster (RSC) infrastructure that is used for advancements in real-time voice translations, language processing, computer vision, and augmented reality (AR).
-
NVIDIA NIM Now Available on Hugging Face with Inference-as-a-Service
Hugging Face has announced the launch of an inference-as-a-service capability powered by NVIDIA NIM. This new service will provide developers easy access to NVIDIA-accelerated inference for popular AI models.
-
GitHub Models Brings New AI Playground and Tight Integration with Other GitHub Tools
GitHub has launched GitHub Models, a free capability aimed at letting developers explore various AI models from within the GitHub tool ecosystem and make it easier to deploy AI-based services using Azure AI. GitHub Models includes both private and public models and is currently in closed preview.
-
Mistral AI Releases Three Open-Weight Language Models
Mistral AI released three open-weight language models: Mistral NeMo, a 12B parameter general-purpose LLM; Codestral Mamba, a 7B parameter code-generation model; and Mathstral, a 7B parameter model fine-tuned for math and reasoning. All three models are available under the Apache 2.0 license.
-
Meta Releases Llama 3.1 405B, Largest Open-Source Model to Date
Meta recently unveiled its latest language model, Llama 3.1 405B. This AI model is the largest of the new Llama models, which also include 8B and 70B versions. With 405 billion parameters, 15 trillion tokens, and 16,000 GPUs, Llama 3.1 405B offers a range of impressive features.
-
Gen AI Increases Workloads and Decreases Productivity, Upwork Study Finds
A controversial survey by Upwork Research Institute found that while 96% of C-suite leaders expect the use of generative AI tools to increase overall productivity levels, 77% of surveyed employees say they have actually decreased their productivity. In fact, the survey contradicts previous research showing a positive correlation.
-
Google Open Sources 27B Parameter Gemma 2 Language Model
Google DeepMind recently open-sourced Gemma 2, the next generation of their family of small language models. Google made several improvements to the Gemma architecture and used knowledge distillation to give the models state-of-the-art performance: Gemma 2 outperforms other models of comparable size and is competitive with models 2x larger.
-
Amazon Q Apps Aim to Simplify the Creation of Generative AI Apps for the Enterprise
Part of its Amazon Q Business offering, Amazon Q Apps enable the creation of generative AI-powered apps integrating enterprise data that can be shared securely within an organization. Along with their general availability, Amazon announced new APIs for Amazon Q Apps and more granular data source definition.
-
Amazon SageMaker Now Offers Managed MLflow Capability for Enhanced Experiment Tracking
AWS has announced the general availability of MLflow capability in Amazon SageMaker. MLflow is an open-source tool commonly used for managing ML experiments. Users can now compare model performance, parameters, and metrics across experiments in the MLflow UI, keep track of their best models in the MLflow Model Registry, and automatically register them as a SageMaker model.