this post was submitted on 23 Nov 2023
1 points (100.0% liked)
Hardware
48 readers
5 users here now
A place for quality hardware news, reviews, and intelligent discussion.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Surely it cant be due to be AI performance?
The 7900xt is 103 tlops of fp16, 7900xtx is 122.
the 4070 is at 117 of fp16 (234 using sparsity) on a smaller chip and thats not banned.
Apparently the AMD significantly outperforms Nvidia in specific calculations used for nuclear weapons simulation software.
Amd is better at fp32 and FP64
During 2017 ish Nvidia and Amd focused on different parts with data centre cards.
Amd went in on Compute with fp32 and fp64.
Nvidia went full in on AI with Tensor cores and fp16 performance.
Amd got faster than Nvidia in some tasks. But Nvidia's bet on AI is the clear winner.
Not FP32, MI300 has 48 TFLOPS, H100 has 60TFLOPs
https://www.topcpu.net/en/cpu/radeon-instinct-mi300
https://www.nvidia.com/en-us/data-center/h100/#:~:text=H100%20triples%20the%20floating%2Dpoint,of%20FP64%20computing%20for%20HPC.
AMD FP64 still gaps Nvidia who in turn gap FP16
Nobody knows the actual flops of the mi300
The mi250x had 95.7 tflops of fp32 due the matrix cores
https://www.amd.com/en/products/server-accelerators/instinct-mi250x
That's more than the H100 even