InfoQ Homepage AI, ML & Data Engineering Content on InfoQ
-
Azure Previews ND H100 V5 Virtual Machines to Accelerate Generative AI
Azure recently announced the preview of the ND H100 v5, virtual machines that integrate the latest Nvidia H100 Tensor Core GPUs and support Quantum-2 InfiniBand networking. According to Microsoft, the new option will offer AI developers improved performance and scaling across thousands of GPUs.
-
AI, ML & Data News Roundup: OpenAI’s GPT-4, Microsoft’s Semantic Kernel, Meta SAM and BloombergGPT
This week's roundup for April 3rd, 2023, includes the most recent news and information from the fields of data science, machine learning, and artificial intelligence.
-
OpenAI Announces GPT-4, Their Next Generation GPT Model
OpenAI recently announced GPT-4, the next generation of their GPT family of large language models (LLM). GPT-4 can accept both text and image inputs and outperforms state-of-the-art systems on several natural language processing (NLP) benchmarks. The model also scored in the 90th percentile on a simulated bar exam.
-
Strategies and Principles to Scale and Evolve MLOps - at QCon London
At the QCon London conference, Hien Luu, senior engineering manager for the Machine Learning Platform at DoorDash, discussed strategies and principles for scaling and evolving MLOps. With 85% of ML projects failing, understanding MLOps at an engineering level is crucial. Luu shared three core principles: "Dream Big, Start Small," "1% Better Every Day," and "Customer Obsession."
-
Responsible AI: from Principle to Practice at QCon London
At the QCon London conference, Microsoft's Mehrnoosh Sameki discussed Responsible AI principles and tools. She emphasized fairness, reliability, safety, privacy, inclusiveness, transparency, and accountability. Tools such as Fairlearn, InterpretML, and the Responsible AI dashboard help implement these principles.
-
AWS Data on EKS Provides Opinionated Data Workload Blueprints
AWS has released Data on EKS (DoEKS), an open-source project providing templates, guidance, and best practices for deploying data workloads on Amazon Elastic Kubernetes Service (EKS). While the main focus is on running Apache Spark on Amazon EKS, blueprints also exist for other data workloads such as Ray, Apache Airflow, Argo Workflows, and Kubeflow.
-
Microsoft Semantic Kernel Enables LLM Integration with Conventional Programs
Microsoft has open sourced Semantic Kernel (SK), a lightweight SDK enabling the integration of large language models (LLMs) with conventional programs which can leverage prompt templating, vectorized memory, intelligent planning, and other capabilities.
-
PyTorch 2.0 Compiler Improves Model Training Speed
The PyTorch Foundation recently released PyTorch version 2.0, a 100% backward compatible update. The main API contribution of the release is a compile function for deep learning models, which speeds up training. Internal benchmarks on 163 open-source AI projects showed that the models ran on average 43% faster during training.
-
The AI Revolution Is Just Getting Started: Leslie Miley Bids Us to Act Now against Its Bias and CO2
At his inaugural keynote of the QCON London conference, Leslie Miley, technical advisor for the CTO at Microsoft, spoke about AI Bias and Sustainability, and how the march towards transformative technologies, like large-scale AI and even crypto, has an inherent cost in the increased CO2 that comes with deployment at scale. More than just context and impact, he suggests mitigation techniques.
-
AWS and NVIDIA to Collaborate on Next-Gen EC2 P5 Instances for Accelerating Generative AI
AWS and NVIDIA announced the development of a highly scalable, on-demand AI infrastructure that is specifically designed for training large language models and creating advanced generative AI applications. The collaboration aims to create the most optimized and efficient system of its kind, capable of meeting the demands of increasingly complex AI tasks.
-
Google Uses AutoML to Discover More Efficient AI Training Algorithm
Researchers at Google have open-sourced EvoLved sIgn mOmeNtum (Lion), an optimization algorithm for training neural networks, which was discovered using an automated machine learning (AutoML) evolutionary algorithm. Models trained with Lion can achieve better accuracy on several benchmarks than models trained with other optimizers, while requiring fewer compute cycles to converge.
-
The Sentence Similarity Scenario in ML.NET Model Builder
Microsoft recently published information about adding the Sentence Similarity scenario in Model Builder. This scenario allows the training of custom sentence similarity models. Together with the addition of this scenario to the Model Builder, it is no longer necessary to install the Model Builder GPU extension.
-
Meta AI’s Large Language Model with 10x Fewer Parameters
Meta AI recently released a new large language model called Language Large Models Meta AI (LLaMA) that outperforms foundational models such as GPT-3 and is competitive with PaLM, despite having 10 times fewer parameters. LLaMA has better performance in language tasks such as natural questions, common-sense reasoning and mathematical reasoning.
-
Google AI Updates Universal Speech Model to Scale Automatic Speech Recognition beyond 100 Languages
Google AI has recently unveiled a new update for their Universal Speech Model (USM), to support the 1,000 Languages Initiative. The new model performs better than OpenAI Whisper for all segments of automation speech recognition.
-
ChatGPT Now Available in Preview on Microsoft’s Azure OpenAI Service
ChatGPT is now available in preview on Microsoft’s Azure OpenAI service allowing developers to integrate ChatGPT directly into a host of different enterprise and end-user applications using a token-based pricing system.