InfoQ Homepage AI, ML & Data Engineering Content on InfoQ
-
QCon New York: Five Tracks to Level-up on the Latest Software Development Practices
The 2023 edition of the QCon New York (June 13-15) software development conference, hosted by InfoQ, is set to bring together over 800 senior software developers. The three-day conference will feature over 80 innovative senior software practitioners from early adopter companies sharing how they are solving current challenges, providing new ideas and perspectives across multiple domains.
-
AWS Enters the Generative AI Race with Bedrock and Titan Foundation Models
AWS announced their entry into the generative AI race with the launch of Amazon Bedrock and Titan foundation models. Amazon aims to democratize access to generative AI technology, catering to customers across various industries and use cases. This groundbreaking development positions Amazon as a formidable competitor in the rapidly growing AI market.
-
Microsoft Open-Sources Multimodal Chatbot Visual ChatGPT
Microsoft Research recently open-sourced Visual ChatGPT, a chatbot system that can generate and manipulate images in response to human textual prompts. The system combines OpenAI's ChatGPT with 22 different visual foundation models (VFM) to support multi-modal interactions.
-
HuggingGPT: Leveraging LLMs to Solve Complex AI Tasks with Hugging Face Models
A recent paper by researchers at Zhejiang University and Microsoft Research Asia explores the use of large language models (LLMs) as a controller to manage existing AI models available in communities like Hugging Face.
-
Meta AI Introduces the Segment Anything Model, a Game-Changing Model for Object Segmentation
Meta AI has introduced the Segment Anything Model (SAM), aiming to democratize image segmentation by introducing a new task, dataset, and model. The project features the Segment Anything Model (SAM) and the Segment Anything 1-Billion mask dataset (SA-1B), which is the most extensive segmentation dataset to date.
-
Twitter Open-Sources Recommendation Algorithm
Twitter recently open-sourced several components of their system for recommending tweets for a user's Twitter timeline. The release includes the code for several of the services and jobs that run the algorithm, as well as code for training machine learning models for embedding and ranking tweets.
-
Bloomberg Unveils a GPT Finance-Focused AI Model
Bloomberg has released BloombergGPT, a new large language model (LLM) that has been trained on enormous amounts of financial data and can help with a range of natural language processing (NLP) activities for the financial sector.
-
Azure Previews ND H100 V5 Virtual Machines to Accelerate Generative AI
Azure recently announced the preview of the ND H100 v5, virtual machines that integrate the latest Nvidia H100 Tensor Core GPUs and support Quantum-2 InfiniBand networking. According to Microsoft, the new option will offer AI developers improved performance and scaling across thousands of GPUs.
-
AI, ML & Data News Roundup: OpenAI’s GPT-4, Microsoft’s Semantic Kernel, Meta SAM and BloombergGPT
This week's roundup for April 3rd, 2023, includes the most recent news and information from the fields of data science, machine learning, and artificial intelligence.
-
OpenAI Announces GPT-4, Their Next Generation GPT Model
OpenAI recently announced GPT-4, the next generation of their GPT family of large language models (LLM). GPT-4 can accept both text and image inputs and outperforms state-of-the-art systems on several natural language processing (NLP) benchmarks. The model also scored in the 90th percentile on a simulated bar exam.
-
Strategies and Principles to Scale and Evolve MLOps - at QCon London
At the QCon London conference, Hien Luu, senior engineering manager for the Machine Learning Platform at DoorDash, discussed strategies and principles for scaling and evolving MLOps. With 85% of ML projects failing, understanding MLOps at an engineering level is crucial. Luu shared three core principles: "Dream Big, Start Small," "1% Better Every Day," and "Customer Obsession."
-
Responsible AI: from Principle to Practice at QCon London
At the QCon London conference, Microsoft's Mehrnoosh Sameki discussed Responsible AI principles and tools. She emphasized fairness, reliability, safety, privacy, inclusiveness, transparency, and accountability. Tools such as Fairlearn, InterpretML, and the Responsible AI dashboard help implement these principles.
-
AWS Data on EKS Provides Opinionated Data Workload Blueprints
AWS has released Data on EKS (DoEKS), an open-source project providing templates, guidance, and best practices for deploying data workloads on Amazon Elastic Kubernetes Service (EKS). While the main focus is on running Apache Spark on Amazon EKS, blueprints also exist for other data workloads such as Ray, Apache Airflow, Argo Workflows, and Kubeflow.
-
Microsoft Semantic Kernel Enables LLM Integration with Conventional Programs
Microsoft has open sourced Semantic Kernel (SK), a lightweight SDK enabling the integration of large language models (LLMs) with conventional programs which can leverage prompt templating, vectorized memory, intelligent planning, and other capabilities.
-
PyTorch 2.0 Compiler Improves Model Training Speed
The PyTorch Foundation recently released PyTorch version 2.0, a 100% backward compatible update. The main API contribution of the release is a compile function for deep learning models, which speeds up training. Internal benchmarks on 163 open-source AI projects showed that the models ran on average 43% faster during training.