InfoQ Homepage Large language models Content on InfoQ
-
Meta Releases Llama 3 Open-Source LLM
Meta AI released Llama 3, the latest generation of their open-source large language model (LLM) family. The model is available in 8B and 70B parameter sizes, each with a base and instruction-tuned variant. Llama3 outperforms other LLMs of the same parameter size on standard LLM benchmarks.
-
The Devoxx Genie IntelliJ Plugin Provides Access to Local or Cloud Based LLM Models
Devoxx Genie, a 100% Java based JetBrains IntelliJ IDEA Plugin, uses local- or cloud-based Large Language Models (LLMs) for generating unit tests and explaining, reviewing and improving source code.
-
Ines Montani at QCon London: Economies of Scale Can’t Monopolise the AI Revolution
During her presentation at QCon London, Ines Montani, co-founder and CEO of explosion.ai (the maker of spaCy), stated that economies of scale are not enough to create monopolies in the AI space and that open-source techniques and models will allow everybody to keep up with the “Gen AI revolution”.
-
Enhancing Developer Experience for Creating Artificial Intelligence Applications
For one company, large language models created a breakthrough in artificial intelligence (AI) by shifting to crafting prompts and utilizing APIs without a need for AI science expertise. To enhance developer experience and craft applications and tools, they defined and established principles around simplicity, immediate accessibility, security and quality, and cost efficiency.
-
Google Text Embedding Model Gecko Distills Large Language Models for Improved Performance
Gecko is a text embedding model that Google created by distilling knowledge from large language models into a general-purpose model. Gecko is trained using a novel approach on a variety of tasks including document retrieval, semantic similarity, and classification, and aims to be as general-purpose as it goes as well as highly performant.
-
Large Language Models for Code by Loubna Ben Allal at QCon London
At QCon London, Loubna Ben Allal discussed Large Language Models (LLMs) for code. She discussed the lifecycle of code completion models, which consists of pre-training on vast codebases and finetuning and continuous adaptation. She specifically discussed open-source models, which are powered by platforms like Hugging Face.
-
Navigating LLM Deployment: Tips, Tricks and Techniques by Meryem Arik at QCon London
At QCon London, Meryem Arik discussed deploying Large Language Models (LLMs). While initial proofs of concept benefit from hosted solutions, scaling demands self-hosting to cut costs, enhance performance with tailored models, and meet privacy and security requirements. She emphasized understanding deployment limits, quantization for efficiency, and optimizing inference to fully use GPU resources.
-
Databrix Announces DBRX, an Open Source General Purpose LLM
Databricks launched DBRX, a new open-source large language model (LLM) that aims to redefine the standards of open models and outperform well-known competitors on industry benchmarks.
-
Apple Researchers Detail Method to Combine Different LLMs to Achieve State-of-the-Art Performance
Many large language models (LLMs) have become available recently, both closed and open source further leading to the creation of combined models known as Multimodal LLMs (MLLMs). Yet, few or none of them unveil what design choices were made to create them, say Apple researchers who distilled principles and lessons to design state-of-the-art (SOTA) Multimodal LLMs.
-
Researchers Open-Source LLM Jailbreak Defense Algorithm SafeDecoding
Researchers from the University of Washington, the Pennsylvania State University, and Allen Institute for AI have open-sourced SafeDecoding, a technique for protecting large language models (LLMs) against jailbreak attacks. SafeDecoding outperforms baseline jailbreak defenses without incurring significant computational overhead.
-
eBay’s Lessons Learned about Generative AI in Software Development Productivity
Recently eBay disclosed the lessons learned about the application of generative AI in the development process. eBay's AI endeavors have uncovered three pivotal avenues toward enhancing developer productivity: through the integration of commercial offerings, fine-tuning existing Large Language Models (LLMs), and harnessing an internal knowledge network.
-
Eric Evans Encourages DDD Practitioners to Experiment with LLMs
In his keynote presentation at Explore DDD 2024 in Denver, Colorado, Eric Evans, author of Domain-Driven Design, argued that software designers need to look for innovative ways to incorporate large language models. He encouraged conference attendees to start learning about LLMs and conducting experiments now, and sharing those results with the community.
-
Azure OpenAI's “Use Your Data” Feature Now Generally Available
Microsoft has officially made On Your Data generally available in Azure OpenAI Service. This feature enables users to harness the full power of OpenAI models, including GPT-4, and seamlessly integrate the advanced features of the RAG (Retrieval Augmented Generation) model with their data. According to the company, all this is backed by enterprise-grade security on Azure.
-
RWKV Project Open-Sources LLM Eagle 7B
The RWKV Project recently open-sourced Eagle 7B, a 7.52B parameter large language model (LLM). Eagle 7B is trained on 1.1 trillion tokens of text in over 100 languages and outperforms other similarly-sized models on multilingual benchmarks.
-
Enhanced Protection for Large Language Models (LLMs) against Cyber Threats with Cloudflare for AI
Cloudflare recently announced a new capability called Firewall for AI in its Web Application Firewall (WAF) offering. The capability adds a new layer of protection that will identify abuse and attacks before they reach and tamper with Large Language Models (LLMs).