InfoQ Homepage Large language models Content on InfoQ
-
Synthetic Data Generator Simplifies Dataset Creation with Large Language Models
Hugging Face has introduced the Synthetic Data Generator, a new tool leveraging Large Language Models (LLMs), that offers a streamlined, no-code approach to creating custom datasets. The tool facilitates the creation of text classification and chat datasets through a clear and accessible process, making it usable for both non-technical users and experienced AI practitioners.
-
OpenAI Presents Research on Inference-Time Compute to Better AI Security
OpenAI presented Trading Inference-Time Compute for Adversarial Robustness, a research paper that investigates the relationship between inference-time compute and the robustness of AI models against adversarial attacks.
-
Microsoft Phi-4 is a Small Language Model Specialized for Complex Math Reasoning
Phi-4 is 14B parameter model from Microsoft Research that aims to improve the state of the art for math reasoning. Previously available on Azure AI Foundry, Phi-4 has recently become available on Hugging Face under the MIT license.
-
Amazon Bedrock Introduces Multi-Agent Systems (MAS) with Open Source Framework Integration
Amazon Web Services has released a multi-agent collaboration capability for Amazon Bedrock, introducing a framework for deploying and managing multiple AI agents that collaborate on complex tasks. The system enables specialized agents to work together under a supervisor agent's coordination, addressing challenges developers face with agent orchestration in distributed AI systems.
-
Microsoft Research Unveils rStar-Math: Advancing Mathematical Reasoning in Small Language Models
Microsoft Research unveiled rStar-Math, a framework that demonstrates the ability of small language models (SLMs) to achieve mathematical reasoning capabilities comparable to, and in some cases exceeding, larger models like OpenAI's o1-mini. This is accomplished without the need for more advanced models, representing a novel approach to enhancing the inference capabilities of AI.
-
Nvidia Ingest Aims to Make it Easier to Extract Structured Information from Documents
Nvidia Ingest is a new microservice aimed at processing document content and extracting metadata into a well-defined JSON schema. Ingest is able to process PDFs, Word, and PowerPoint documents and extract structured information from tables, charts, images, and text using optical character recognition.
-
Microsoft Research AI Frontiers Lab Launches AutoGen v0.4 Library
Microsoft Research’s AI Frontiers Lab has announced the release of AutoGen version 0.4, an open-source framework designed to build advanced AI agent systems. This latest version as stated marks the complete redesign of the AutoGen library, focusing on enhancing code quality, robustness, usability, and the scalability of agent workflows.
-
DeepSeek Open-Sources DeepSeek-V3, a 671B Parameter Mixture of Experts LLM
DeepSeek open-sourced DeepSeek-V3, a Mixture-of-Experts (MoE) LLM containing 671B parameters. It was pre-trained on 14.8T tokens using 2.788M GPU hours and outperforms other open-source models on a range of LLM benchmarks, including MMLU, MMLU-Pro, and GPQA.
-
Google Releases Experimental AI Reasoning Model
Google has introduced Gemini 2.0 Flash Thinking Experimental, an AI reasoning model available in its AI Studio platform.
-
Google Vertex AI Provides RAG Engine for Large Language Model Grounding
Vertex AI RAG Engine is a managed orchestration service aimed to make it easier to connect large language models (LLMs) to external data sources to be more up-to-date, generate more relevant responses, and hallucinate less.
-
Android Studio Ladybug Update Adds Gemini Support, New Debugging Features, and More
In its recent update to Android Studio Ladybug (2024.2.2), Google has added new Gemini Code Transforms to modify, refactor, or create code, debugging and testing tools, and developer experience improvements. Additionally, the IDE adopts the latest IntelliJ 2024.2 platform release.
-
Major LLMs Have the Capability to Pursue Hidden Goals, Researchers Find
Researchers at AI safety firm Apollo Research found that AI agents may covertly pursue misaligned goals and hide their true objectives. Known as in-context scheming, this behavior does not seem to be accidental as LLMs explicitly reason about deceptive strategies and consider them a viable strategy.
-
HuatuoGPT-o1: Advancing Complex Medical Reasoning with AI
Researchers from The Chinese University of Hong Kong, Shenzhen, and the Shenzhen Research Institute of Big Data have introduced HuatuoGPT-o1, a medical large language model (LLM) designed to improve reasoning in complex healthcare scenarios.
-
Google Releases PaliGemma 2 Vision-Language Model Family
Google DeepMind released PaliGemma 2, a family of vision-language models (VLM). PaliGemma 2 is available in three different sizes and three input image resolutions and achieves state-of-the-art performance on several vision-language benchmarks.
-
Nvidia Announces Arm-Powered Project Digits, Its First Personal AI Computer
Capable of running 200B-parameter models, Nvidia Project Digits packs the new Nvidia GB10 Grace Blackwell chip to allow developers to fine-tune and run AI models on their local machines. Starting at $3,000, Project Digits targets AI researchers, data scientists, and students to allow them to create their models using a desktop system and then deploy them on cloud or data center infrastructure.