Jump to content
  • Sign Up
×
×
  • Create New...

AMD’s ROCm documentation now includes a dedicated ROCm on Radeonsection


Recommended Posts

  • Diamond Member

This is the hidden content, please

AMD’s ROCm documentation now includes a dedicated ROCm on Radeonsection

data:image/gif;base64,R0lGODlhAQABAIAAAAAAAP///ywAAAAAAQABAAACAUwAOw==

AMD has expanded support for machine learning (ML) development on its RDNA 3 GPUs with Radeon Software for Linux 24.10.3 and ROCm 6.1.3. This enables developers using frameworks like PyTorch, ONNX Runtime, or TensorFlow to perform ML tasks on a local workstation featuring Radeon RX 7000 and Radeon Pro W7000-series GPUs and avoid the usage of expensive cloud-based solutions.   

Normally, AMD only supported AI/ML frameworks on Instinct accelerators based on the CDNA architecture for AI and HPC applications, but recently it changed its mind and enabled AI acceleration on its RDNA 3-based GPUs. To make it easier for developers to use Radeon graphics processors for AI development, starting from ROCm 6.1.3 and Radeon Software for Linux 24.10.3 software, 

This is the hidden content, please
 to simplify installation and reveal compatibility details. 

With ROCm 6.1.3, users can now use these GPUs for advanced AI tasks using frameworks like PyTorch, TensorFlow, and ONNX Runtime. Other highlights of the ROCm 6.1.3 are enhanced inference capabilities for ONNX Runtime, enabling optimized processing with INT8 data format using MIGraphX. 

AMD’s ROCm 6.1.3 software stack turns a desktop setup into a cost-effective alternative to servers or cloud-based ML platforms, bringing parallel computing capabilities directly to local machines. Also, since AMD’s ROCm is compatible with AMD’s CDNA architecture, it makes it easy for developers to transition their desktop-developed applications to datacenter environments without changing frameworks. 

While it is evident that AMD is doing a lot to make its Radeon GPUs more attractive to AI developers, there are still a few things it could add. AMD’s ROCm 6.1.3 was released in June. So far, AMD has released its 

This is the hidden content, please
 for Instinct APUs and accelerators, which has more features and capabilities, such as broader FP8 support. Perhaps more importantly, ROCm 6.2 brings vLLM, which addresses significant inferencing challenges, such as efficient multi-GPU computation, reduced memory usage, and minimized computational bottlenecks, according to AMD. As a result, AMD’s Instinct accelerators and processors still have numerous advantages over consumer GPUs, particularly when it comes to the usage of such GPUs in the datacenter.



This is the hidden content, please

#AMDs #ROCm #documentation #includes #dedicated #ROCm #Radeonsection

This is the hidden content, please

This is the hidden content, please

For verified travel tips and real support, visit: https://hopzone.eu/

Join the conversation

You can post now and register later. If you have an account, sign in now to post with your account.

Guest
Unfortunately, your content contains terms that we do not allow. Please edit your content to remove the highlighted words below.
Reply to this topic...

×   Pasted as rich text.   Paste as plain text instead

  Only 75 emoji are allowed.

×   Your link has been automatically embedded.   Display as a link instead

×   Your previous content has been restored.   Clear editor

×   You cannot paste images directly. Upload or insert images from URL.

  • Vote for the server

    To vote for this server you must login.

    Jim Carrey Flirting GIF

  • Recently Browsing   0 members

    • No registered users viewing this page.

Important Information

Privacy Notice: We utilize cookies to optimize your browsing experience and analyze website traffic. By consenting, you acknowledge and agree to our Cookie Policy, ensuring your privacy preferences are respected.