InfoQ Homepage GPU Content on InfoQ
-
NVIDIA Announces Next Generation AI Hardware H100 GPU and Grace CPU Superchip
At the recent GTC conference, NVIDIA announced their next generation processors for AI computing, the H100 GPU and the Grace CPU Superchip. Based on NVIDIA's Hopper architecture, the H100 includes a Transformer engine for faster training of AI models. The Grace CPU Superchip features 144 Arm cores and outperforms NVIDIA's current dual-CPU offering on the SPECrate 2017_int_base benchmark.
-
Ten Lessons from Three Generations of Tensor Processing Units
A recent report published by Google’s TPU group highlights ten takeaways from developing three generations of tensor processing units. The authors also discuss how their previous experience will affect the development of future tensor processing units.
-
Microsoft Introduces NVads A10 V5 Azure VMs in Preview for Graphics-Heavy Workloads
Microsoft recently announced the NVads A10 v5 series in preview. These virtual machines (VMs) are powered by NVIDIA A10 GPUs and AMD EPYC 74F3V(Milan) CPUs with a base frequency of 3.2 GHz and an all-core peak frequency of 4.0 GHz.
-
AMD Introduces Its Deep-Learning Accelerator Instinct MI200 Series GPUs
In its recent Accelerated Data Center Premiere Keynote, AMD unveiled its MI200 accelerator series Instinct MI250x and slightly lower-end Instinct MI250 GPUs. Designed with CDNA-2 architecture and TSMC’s 6nm FinFET lithography, the high-end MI250X provides 47.9 TFLOPs peak double precision performance and memory that will allow training larger deep networks by minimizing model sharding.
-
AWS Announces the Availability of EC2 Instances (G5) with NVIDIA A10G Tensor Core GPUs
Recently AWS announced the availability of new G5 instances, which feature up to eight NVIDIA A10G Tensor Core GPUs. These instances are powered by second-generation AMD EPYC processors.
-
Amazon Releases DL1 Instances Powered by Gaudi Accelerators
Amazon recently announced the general availability of the EC2 DL1 instances powered by Gaudi accelerators from Habana Labs. The new instances promise better price performances in training deep learning models for use cases such as computer vision, natural language processing, autonomous vehicle perception and recommendation engines.
-
OpenAI Releases Triton, Python-Based Programming Language for AI Workload Optimization
OpenAI released their newest language, Triton, an open-source programming language that enables researchers to write highly efficient GPU code for AI workloads. Triton is Python-compatible and allows new users to achieve expert-quality results in only 25 lines of code. The code is written in Python using Triton’s libraries, which are then JIT-compiled to run on the GPU.
-
Microsoft Announces the General Availability of Azure ND A100 V4 Cloud GPU Instances
Recently Microsoft announced the general availability of the Azure ND A100 v4 Cloud GPU instances—powered by NVIDIA A100 Tensor Core GPUs. These Virtual Machines (VMs) are targeted at customers with high performance and demanding workloads like Artificial Intelligence (AI) and Machine Learning (ML) workloads.
-
Deno 1.8 Ships with WebGPU Support, Dynamic Permissions, and More
Deno 1.8 recently shipped with plenty of new features, including WebGPU support, internationalization APIs, stabilized import maps, support for fetching private modules, and more. Deno permissions, links, and symlinks are now stable. Deno 1.8 additionally ships with TypeScript 4.2.
-
Is Julia Production Ready? Q&A with Bogumił Kamiński
On the heels of JuliaCon 2020, SGH Warsaw School of Economics professor and DataFrames.jl maintainer Bogumił Kamiński summarized the status of the language and its ecosystem and stated that Julia is finally production-ready. InfoQ has taken the chance to speak with professor Kamiński.
-
Concurnas: the New Language on the JVM for Concurrent and GPU Computing
Concurnas is a new open source JVM programming language designed for building concurrent and distributed systems. Concurnas is a statically typed language with object oriented, functional, and reactive programming constructs. With native support for GPU computing and vectorization, Concurnas allows for building machine learning applications and high performance parallel applications.
-
TornadoVM: Running Java on GPUs and FPGAs with Dr Juan Fumero at QCon London
Dr Juan Fumero presented at QCon London on TornadoVM, a plug-in to OpenJDK and GraalVM that runs Java on heterogeneous hardware including Graphical Processing Units (GPUs) and Field Programmable Gate Arrays (FPGAs). Demos during the presentation showed code being sped up by hundreds of times when running on a GPU vs a CPU.
-
Boosting Apache Spark with GPUs and the RAPIDS Library
At the 2019 Spark AI Summit Europe conference, NVIDIA software engineers Thomas Graves and Miguel Martinez hosted a session on Accelerating Apache Spark by Several Orders of Magnitude with GPUs and RAPIDS Library. InfoQ recently talked with Jim Scott, head of developer relations at NVIDIA, to learn more about accelerating Apache Spark with GPUs and the RAPIDS library.
-
PyTorch 1.1 Release Improves Performance, Adds New APIs and Tools
Facebook AI Research announced the release of PyTorch 1.1. The latest version of the open-source deep learning framework includes improved performance via distributed training, new APIs, and new visualization tools including native support for TensorBoard.
-
Uber Introduces AresDB: GPU-Powered, Open-Source, Real-Time Analytics Engine
Uber recently introduced AresDB, an open-source real-time analytics engine leveraging an unconventional power source - graphics processing units (GPUs) - for meeting the growing demands of analysis at scale and at the same time unifying, simplifying and improving Uber’s existing solutions.