InfoQ Homepage Large language models Content on InfoQ
-
How a Software Architect Uses Artificial Intelligence in His Daily Work
Software architects and system architects will not be replaced anytime soon by generative artificial intelligence (AI) or large language models (LLMs), Avraham Poupko said. They will be replaced by software architects who know how to leverage generative AI and LLMs, and just as importantly, know how NOT to use generative AI.
-
Latin America Launches Latam-GPT to Improve AI Cultural Relevance
Latin America is advancing in the development of artificial intelligence with the creation of Latam-GPT, a language model designed to better represent the history, culture, and linguistic diversity of the region.
-
Meta Introduces LLM-Powered Tool for Software Testing
Meta has unveiled the Automated Compliance Hardening (ACH) tool, a mutation-guided, LLM-based test generation system. Designed to enhance software reliability and security, ACH generates faults in source code and subsequently creates tests to detect and address these issues.
-
UC Berkeley's Sky Computing Lab Introduces Model to Reduce AI Language Model Inference Costs
UC Berkeley's Sky Computing Lab has released Sky-T1-32B-Flash, an updated reasoning language model that addresses the common issue of AI overthinking. The model, developed through the NovaSky (Next-generation Open Vision and AI) initiative, "slashes inference costs on challenging questions by up to 57%" while maintaining accuracy across mathematics, coding, science, and general knowledge domains.
-
Gemini 2.0 Family Expands with Cost-Efficient Flash-Lite and Pro-Experimental Models
Announced last December, the Gemini 2.0 family of models now has a new member, Gemini 2.0 Flash-Lite, which Google says is cost-optimized for large scale text output use cases and is now available in preview. Along with Flash-Lite, Google also announced Gemini 2.0 Pro.
-
OpenAI Releases Reasoning Model o3-mini, Faster and More Accurate Than o1
OpenAI released OpenAI o3-mini, their latest reasoning LLM. o3-mini is optimized for STEM applications and outperforms the full o1 model on science, math, and coding benchmarks, with lower response latency than o1-mini.
-
Micronaut Framework 4.7.0 Provides Integration with LangChain4j and Graal Languages
The Micronaut Foundation has released Micronaut Framework 4.7.0 in December 2024, four months after the release of version 4.6.0. This version provides LangChain4J support to integrate LLMs into Java applications. Micronaut Graal Languages provides integration with Graal-based dynamic languages such as the Micronaut GraalPy feature to interact with Python.
-
OpenEuroLLM: Europe’s New Initiative for Open-Source AI Development
A consortium of 20 European research institutions, companies, and EuroHPC centers has launched OpenEuroLLM, an initiative to develop open-source, multilingual large language models (LLMs). Coordinated by Jan Hajič and co-led by Peter Sarlin, the project aims to provide transparent and compliant AI models for commercial and public sector applications.
-
OpenAI Launches Deep Research: Advancing AI-Assisted Investigation
OpenAI has launched Deep Research, a new agent within ChatGPT designed to conduct in-depth, multi-step investigations across the web. Initially available to Pro users, with plans to expand access to Plus and Team users, Deep Research automates time-consuming research by retrieving, analyzing, and synthesizing online information.
-
DeepSeek Database Leaking Sensitive Information Highlights AI Security Risks
Cloud security firm Wiz uncovered unprotected DeepSeek database giving full control over database operations and access to internal data including millions of lines of chat logs. While the vulnerability has been quickly fixed, the incident shows the need for the AI industry to enforce higher security standards, says the company.
-
DeepSeek Open-Sources DeepSeek-R1 LLM with Performance Comparable to OpenAI's o1 Model
DeepSeek open-sourced DeepSeek-R1, an LLM fine-tuned with reinforcement learning )RL) to improve reasoning capability. DeepSeek-R1 achieves results on par with OpenAI's o1 model on several benchmarks, including MATH-500 and SWE-bench.
-
Hugging Face Expands Serverless Inference Options with New Provider Integrations
Hugging Face has launched the integration of four serverless inference providers Fal, Replicate, SambaNova, and Together AI, directly into its model pages. These providers are also integrated into Hugging Face's client SDKs for JavaScript and Python, allowing users to run inference on various models with minimal setup.
-
JetBrains AI Coding Agent Junie Provides Tight Integration with JetBrains IDEs
JetBrains has announced Junie, its new AI coding agent, in closed preview. Junie, says the company, is able to carry through the coding tasks you assign it and leverage the knowledge about your project context as available in the IDE.
-
AMD and Johns Hopkins Researchers Develop AI Agent Framework to Automate Scientific Research Process
Researchers from AMD and Johns Hopkins University have developed Agent Laboratory, an artificial intelligence framework that automates core aspects of the scientific research process. The system uses large language models to handle literature reviews, experimentation, and report writing, producing both code repositories and research documentation.
-
DeepSeek Release Another Open-Source AI Model, Janus Pro
DeepSeek has released Janus-Pro, an updated version of its multimodal model, Janus. The new model improves training strategies, data scaling, and model size, enhancing multimodal understanding and text-to-image generation.