Jump to content
  • Sign Up
×
×
  • Create New...

‘Enhanced’ Nvidia A100 GPUs appear in China’s second-hand market — new cards surpass sanctioned counterparts with 7,936 CUDA cores and 96GB HBM2 memory


Recommended Posts

  • Diamond Member



‘Enhanced’ Nvidia A100 GPUs appear in China’s second-hand market — new cards surpass sanctioned counterparts with 7,936 CUDA cores and 96GB HBM2 memory

Nvidia’s Ampere A100 was previously one of the top AI accelerators, before being dethroned by the newer Hopper H100 — not to mention the H200 and upcoming Blackwell GB200. It looks like the chipmaker may have experimented with an enhanced version that never hit the market, or perhaps companies have clandestinely modified the A100 to make it even faster in the wake of U.S. sanctions against China. X user

This is the hidden content, please
recently discovered various A100 prototypes in the ******** second-hand market that flaunt substantially higher specifications than Nvidia’s ‘regular’ A100.

Despite the beefed-up attributes, the A100 7936SP (unofficial name, based on its having 7936 shader processors) shares the same GA100 Ampere **** as the regular A100. However, the former has 124 enabled SMs (Streaming Multiprocessors) out of the possible 128 on the GA100 silicon. While it’s not the maximum configuration, the A100 7936SP has 15% more CUDA cores than the standard A100, representing a significant performance uplift.

Tensor core counts likewise increase in proportion to the number of SMs. Having more enabled SMs thus means that the A100 7936SP also possesses more Tensor cores. Based on specs alone, the 15% increase in SM, CUDA, and Tensor core counts could similarly boost AI performance by 15%.

Nvidia offers the A100 in 40GB and 80GB configurations. The A100 7936SP likewise comes in two variants. The A100 7936SP 40GB model flaunts a 59% higher base clock than the A100 80GB while maintaining the same 1,410 MHz boost clock. On the other hand, the A100 7936SP 96GB shows an 18% faster base clock compared to the regular A100, and it also enables the sixth HBM2 stack to get to 96GB of total memory. Sadly, ******** sellers have censored the boost clock speed from the GPU-Z screenshot.

Nvidia A100 7936SP Specifications

Graphics Card A100 7936SP 96GB A100 80GB A100 7936SP 40GB A100 40GB
Architecture GA100 GA100 GA100 GA100
Process Technology TSMC 7N TSMC 7N TSMC 7N TSMC 7N
Transistors (Billion) 54.2 54.2 54.2 54.2
**** size (mm^2) 826 826 826 54.2
SMs 124 108 124 108
CUDA Cores 7,936 6,912 7,936 6,912
Tensor / AI Cores 496 432 496 432
Ray Tracing Cores N/A N/A N/A N/A
Base Clock (MHz) 1,260 1,065 1,215 765
Boost Clock (MHz) ? 1,410 1,410 1,410
TFLOPS (FP16) >320 312 358 312
VRAM Speed (Gbps) 2.8 3 2.4 2.4
VRAM (GB) 96 80 40 40
VRAM Bus Width (Bit) 6,144 5,120 5,120 5120
L2 (MB) ? 80 ? 40
Render Output Units 192 160 160 160
Texture Mapping Units 496 432 432 432
Bandwidth (TB/s) 2.16 1.94 1.56 1.56
TDP (watts) ? 300 ? 250

The A100 7936SP 40GB memory subsystem is identical to the A100 40GB. The 40GB of HBM2 memory runs at 2.4 Gbps across a 5120-bit memory interface using five HBM2 stacks. The design contributes to a maximum memory bandwidth of up to 1.56 TB/s. The A100 7936SP 96GB model, however, is the centerfold here. The graphics card has 20% more HBM2 memory than what Nvidia offers thanks to the sixth enabled HBM2 stack. Training very large language models can be memory intensive, so the added capacity would certainly come in handy for AI work.

The A100 7936SP 96GB appears to sport a revamped memory subsystem compared to the A100 80GB — the HBM2 memory checks in at 2.8 Gbps instead of 3 Gbps but resides on a wider 6144-bit memory bus to help make up the difference. This results in the A100 7936SP 96GB having approximately 11% more memory bandwidth than the A100 80GB.

The A100 40GB and 80GB have TDPs of 250W and 300W, respectively. Given the faster specifications, the A100 7936SP could have a higher TDP. However, the value isn’t available from the shared GPU-Z screenshots. The engineering PCB has three 8-pin PCIe power connectors instead of the vanilla A100’s single 8-pin PCIe power connector. Being an engineering prototype, the A100 7936SP may not use all three power connectors, but it should draw somewhat more power than the standard A100 due to the extra CUDA cores and HBM2 memory.

Many ******** sellers are selling the A100 7936SP on

This is the hidden content, please
. The 96GB model ranges between
This is the hidden content, please
and
This is the hidden content, please
. It’s unknown if the accelerators are engineering samples that escaped Nvidia’s lab, or if they’re customized models that the chipmaker developed for a specific client. In any event, it isn’t legal to pick one up while the A100 may be subject to the latest U.S. export sanctions, that doesn’t affect cards already within China.

Of course, there’s no warranty or official driver support. While the A100 7936SP offers better performance than the A100 at the same or potentially lower price, purchasing a retail product or renting a GPU for all your AI needs is safer. But for the ******** market, which can no longer import A100 GPUs, the added memory and compute are apparently worth considering.





This is the hidden content, please

#Enhanced #Nvidia #A100 #GPUs #Chinas #secondhand #market #cards #surpass #sanctioned #counterparts #CUDA #cores #96GB #HBM2 #memory

This is the hidden content, please

Create an account or sign in to comment

You need to be a member in order to leave a comment

Create an account

Sign up for a new account in our community. It's easy!

Register a new account

Sign in

Already have an account? Sign in here.

Sign In Now
  • Vote for the server

    To vote for this server you must login.

    Jim Carrey Flirting GIF

  • Recently Browsing   0 members

    • No registered users viewing this page.

Important Information

Privacy Notice: We utilize cookies to optimize your browsing experience and analyze website traffic. By consenting, you acknowledge and agree to our Cookie Policy, ensuring your privacy preferences are respected.