Instalar Steam
iniciar sesión
|
idioma
简体中文 (chino simplificado)
繁體中文 (chino tradicional)
日本語 (japonés)
한국어 (coreano)
ไทย (tailandés)
Български (búlgaro)
Čeština (checo)
Dansk (danés)
Deutsch (alemán)
English (inglés)
Español de Hispanoamérica
Ελληνικά (griego)
Français (francés)
Italiano
Bahasa Indonesia (indonesio)
Magyar (húngaro)
Nederlands (holandés)
Norsk (noruego)
Polski (polaco)
Português (Portugués de Portugal)
Português-Brasil (portugués de Brasil)
Română (rumano)
Русский (ruso)
Suomi (finés)
Svenska (sueco)
Türkçe (turco)
Tiếng Việt (vietnamita)
Українська (ucraniano)
Comunicar un error de traducción
So you're dodging the question then?
Let that sink in... It's faster, yet has slower theoretical performance... Hmmm. Why is that though? Could it be because most run mixed, since FP16 is lower accuracy and limited due to being half precision. This means a maximum of 65504 bits. And while super fast, mixed is generally used more. Why though?
Well, let's see... Maybe because FP16 suffers from poor weight updating, gradient underflow, and god forbid activation underflow/loss. How do we fix it? Using FP32 to do the calculation to prevent over/underflow, FP32 handles the scaling factor so weights are made more precise, then sent back to FP16.
So, now that you've been educated, tell me how it's gimped. Since even being "Gimped" as you claim, it's out performing the TitanV, RTX Titan, and 3090TI in such areas. Barely losing to the H100 due to memory limitations.
I really don't see how this is gimped, but by all means, show me how its FP16 performance gimps gaming, since it's sold as a gaming and content creator GPU.
And again...
https://youtu.be/K8_QPx-IN-o
It's a beast at what they advertising it does.
By gimped, I mean it's a driver limitation forcing only 1:1 conversions, so they literally limit it.
So for AI elements the FP16 is actually severely limiting, and you're not actually better off using FP32 than you'd think due to with FP16 being gimped, in the TC operations it limits the FP32 accumulation.
FP16 isn't really the greatest for mixed precision training, but for raw number crunching it's actually okay. But for image processing and neural networks, FP16 is great.
I literally said how it's gimped. They force 1:1 ratio in the drivers.
And yet...
Wow shocker FP and TF not the same thing.
This just in, the sky is blue.
https://github.com/lambdal/deeplearning-benchmark/blob/22.09-py3/pytorch/pytorch-train-throughput-fp32.csv
Wanna keep going?
You're showing you don't know things again. And the card was never advertised to do deep learning, ya showing how dumb you are.
And well, look at that... It's still the most affordable for even something it's not meant to do compared to how well it performs. SHOCKER!
Wanna keep going? Notice how on tacotron and waveglow they're practically identical?
That's the TC operation limit that the driver is forcing.
But hey have if you have core sparsity on regular tasks it does a 4:2. Go ♥♥♥♥♥♥♥ figure
That's literally my whole point this entire time. Even on the Hopper GPU's Nvidia decided a 1:1 FP16 on non sparse matrices was a good idea.
Guess how many times in real world cases do you have a sparse matrice? If you guessed virtually none, then you'd be correct.
I think we can all see now, that the 4090 is a fine card for what it offers and is designed for.
People in this thread just broke as a joke, as it were, and can't afford high end hardware as it's released. Instead they try and claim said hardware is crap, and performs poorly, even when proven wrong consistently.
May I recommend for those people, to buy used, or wait for next generation to release and then upgrade when current gen becomes not so current, this way you'll save money.
Time to let the thread die and the two cry babies to wallow in their own misery.
The true 2:1 explains why it more than doubles on a lot of the tests.
And then tacotron and waveform were designed to run on the V100 DGX. which conveniently wasn't on that list. (no i'm not actually saying that sarcastically, why wasn't it on the list?)
And then we finally agree on something. I'd call the 4090 a "fine" card at best. but not a GREAT card.