AMD is hoping to make AI more accessible to developers and researchers by adding support for PyTorch to its Radeon RX 7900 XTX and Radeon Pro W7900 graphics cards.
Based on its RDNA 3 GPU architecture, these are some of the best GPUs out there, and they can now let users establish a private and cost-effective workflow for machine learning training and inference. Previously, users may have needed to rely on cloud access to compatible GPUs for such AI workloads.
“We are excited to offer the AI community new support for machine learning development using PyTorch built on the AMD Radeon RX 7900 XTX and Radeon Pro W7900 GPUs and the ROCm open software platform,” said Dan Wood, VP of Radeon product management. “This is our first RDNA 3 architecture-based implementation, and we are looking forward to partnering with the community.”
Making the most of ROCm
With the PyTorch machine learning framework now supported on its most powerful graphics cards, AMD is hoping to crack open access to AI workloads for users who don’t have the means or infrastructure that you’d otherwise need.
Anybody hoping to take advantage of PyTorch can also use the Radeon Open Compute (ROCm) software stack for GPUs – which spans general-purpose computing, high-performance computing (HPC), as well as heterogeneous computing.
With AMD ROCm 5.7, users of machines with RDNA 3-based GPUs, as well as CDNA GPU and AMD Instinct MI series accelerators, can also use PyTorch.
Because ROCm is open source, developers may wish to take in all kinds of different directions and add support for their own particular AI processing needs. There is, for example, a huge amount of appetite out there to get Stable Diffusion running on AMD accelerated processing units (APUs).
One user, for example, turned a 4600G APU into a 16GB VRAM GPU that could run AI workloads – including on Stable Diffusion – without too much of a hitch, according to a video they posted on YouTube.
More from TechRadar Pro
This post originally appeared on TechToday.