Jump to content
  • Sign Up
×
×
  • Create New...

Recommended Posts

  • Diamond Member

AMD Instinct MI300 AI accelerator takes aim at Nvidia GPUs

Both AMD and

This is the hidden content, please
released AI accelerators today: AMD Instinct MI300 and
This is the hidden content, please
TPU v5e. Both are data center-grade processors that speed AI tasks, such as training large language models.

AMD is playing catch-up to Nvidia, which has parleyed its gaming tech expertise into an AI processing superpower. AI typically runs on chips adjacent to CPUs; AMD’s accelerator is a GPU, while

This is the hidden content, please
’s is a proprietary tensor processing unit (TPU) that powers AI in the
This is the hidden content, please
Cloud.

What do the 153 billion transistors in AMD’s MI300 accelerator — and its claimed 17TB/second bandwidth — get enterprise IT buyers? The Instinct MI300 chips run AI operations much faster, AMD CEO Lisa Su said at a launch event.

AMD customers and partners there, including Dell, HPE,

This is the hidden content, please
, Meta, Oracle, Databricks and others, said they had the chips either running in their products and services, are testing them, or plan to use them soon. Not only are the chips faster than their predecessors, but they can be combined to further improve performance.

“Generative AI is the most demanding data center workload ever,” Su said. “It requires tens of thousands of accelerators to train and refine models with billions of parameters. And that same infrastructure is also needed to answer the millions of queries from everyone around the world.

AMD Instinct MI300 GPU Accelerator.

“It’s very simple: The more compute you have, the more capable the model, the faster the answers are generated. And the GPU is at the center of this generative AI world,” she said.

The hardware upon which AI accelerators run has become a key feature of AI accelerators, said Daniel Newman, Futurum Research founder. It’s not just speeds and feeds anymore but open source platforms that let developers build software and connect their large language models to the hardware.

“Today is all about AMD entering with valid, competitive capabilities and products using open source in the era of an incredibly strong or even dominant Nvidia in the AI training [chip] and overall AI chip,” said Daniel Newman, Futurum Research founder. “It isn’t just about performance. It is also about availability, viability, capability, and the world understanding that open-source collaborative ecosystems for AI are important.”

Enterprise AI buyers, take note

Many companies still field their own GPUs in their data centers or colocations — even in the cloud-first era — Gartner analyst Chirag Dekate said. Data privacy regulations or the need for intellectual property protection force companies to take a hybrid approach that mixes their own data centers and public clouds such as

This is the hidden content, please
, AWS and
This is the hidden content, please
.

In some cases, an enterprise might run its proprietary LLM in its own data center to keep it off a public cloud.

The AMD GPU accelerators will be adopted not only by large public clouds but also by individual enterprise customers, Dekate predicted. The combination of hardware, software and partnerships will help those customers set up their AI operations faster.

“What AMD is announcing today is not just a GPU that can be deployed in the data center,” Dekate said. “They’re also announcing cloud partnerships. They’re announcing platforms and software stacks. [Together they will] enable enterprises to hit the ground running with an AMD-native strategy.”

This is the hidden content, please
delivers new AI accelerators

Amid its Gemini general AI model release and unveiling of plans to be the first manufacturer to put generative AI on smartphones,

This is the hidden content, please
also released the
This is the hidden content, please
, its latest AI accelerator. TPUs power
This is the hidden content, please
’s own AI in apps such as Maps,
This is the hidden content, please
and Gmail, and it hopes
This is the hidden content, please
Cloud Platform customers will follow suit.

In the future, it’s likely that enterprise cloud services buyers will have different AI services powered by different manufacturers’ chips, Dekate said. Some enterprise applications and operations will work best — or cheapest — on one chipmaker’s array compared to the others. It will depend on the scale and bandwidth required for a job, such as training a large enterprise language model.

Competition will be the key to keeping AI chips viable and to keep advancements moving in the AI hardware race as each manufacturer tries to outdo the others, Newman said.

“Ultimately we need a highly competitive marketplace for AI infrastructure, chipsets, software, and more,” Newman said. “[Generative AI represents] the biggest transformation our world has seen technologically, and a healthy, vibrant, competitive ecosystem is critical.”

Don Fluckinger covers digital experience management, end-user computing, CPUs and assorted other topics for TechTarget Editorial. Got a tip? Email him here.



This is the hidden content, please

#AMD #Instinct #MI300 #accelerator #takes #aim #Nvidia #GPUs

This is the hidden content, please

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now
  • Vote for the server

    To vote for this server you must login.

    Jim Carrey Flirting GIF

  • Recently Browsing   0 members

    • No registered users viewing this page.

Important Information

Privacy Notice: We utilize cookies to optimize your browsing experience and analyze website traffic. By consenting, you acknowledge and agree to our Cookie Policy, ensuring your privacy preferences are respected.