- Joined
- Aug 20, 2006
- Messages
- 13,000
According to Wccftech, AMD was not being entirely honest when comparing its new 7nm Radeon Instinct M160 GPU with NVIDIA’s Tesla V100. The company’s RESNET-50 benchmark suggested that the parts were comparable, but that was only because the Tensor cores in the Tesla were disabled: it was only running at a third of its potential. “The performance of the V100 is just over three times that of the Radeon Instinct MI60.”
The company had claimed comparable inference performance as compared to NVIDIA’s Tesla V100 flagship GPU. I remembered seeing ResNet 50 performance before and could distinctly remember it being in the 1000s, so I looked through the footnotes and found the cause: the test was conducted in FP32 mode. The Tesla V100 contains Tensor cores and significantly more die space (the GCN architecture is hard-limited to 4096 stream processors) and those can be used to accelerate inference and learning performance by multiple factors.
The company had claimed comparable inference performance as compared to NVIDIA’s Tesla V100 flagship GPU. I remembered seeing ResNet 50 performance before and could distinctly remember it being in the 1000s, so I looked through the footnotes and found the cause: the test was conducted in FP32 mode. The Tesla V100 contains Tensor cores and significantly more die space (the GCN architecture is hard-limited to 4096 stream processors) and those can be used to accelerate inference and learning performance by multiple factors.