GPUs & CPUs & Enthusiast hardware: Questions, Discussion and fanboy slap-fights - Nvidia & AMD & Intel - Separe but Equal. Intel rides in the back of the bus.

  • 🐕 I am attempting to get the site runnning as fast as possible. If you are experiencing slow page load times, please report it.
I always wondered why they switched. I remember seeing gcn/vega reviews and they were going toe to toe with the Titan class cards in productivity when they came out. It was a genuine strong point of AMD at the time, but then they switched to a separate gaming and compute division for rdna/cdna just really odd choices.
RDNA absolutely sucked for general compute. Don't ask me why, I don't know the details. I just know that it did and couldn't really be made to work. I am not an EE, but my Dad works at Nintendo.
 
I need someone to correct me if I'm wrong, but I've had an observation when doing GPU undervolting.

Normally, when I stress the GPU with something like 3DMark at the stock curve, I notice that it aggressively boosts to ~1800MHz, then back to ~1600MHz, while the card's nominal frequency is 1710MHz.
When I undervolt it, I essentially lock it at 1710MHz, the GPU holds that clock steady. If I boost it a bit, then it'll only clock down by 15-30MHz on occasion, and in the end I will get a higher benchmark score than on stock.

Is it that by default, when GPU's constantly whiplash between boosting to the highest clock possible, then overcorrect it to a safer one, they end up being less performant than if you were to force them at a given nominal frequency and not let them exceed it? Is that what I'm noticing?
 
I wonder what the catches will be with UDNA. They'll have to nerf the consumer cards in some ways to avoid undercutting the pro cards (FP64 being an obvious one).
Probably just different vram amounts. I don't see them massively differentiating them again as that would kill the goal of unifying compute and gaming.
Considering the Sony partnership pays off udna might actually be what everybody wants in terms of all around performance and features. Its good to see them now putting similar focus on the gpu division and giving it the same spotlight the cpus get.

RDNA absolutely sucked for general compute
They probably completely stripped anything related to compute out.
 
Last edited:
They probably completely stepped anything related to compute out.
Shaders wouldn't work if they did. All I know is that when it came to ingesting large arrays and executing arbitrary code, it was woefully inefficient and couldn't really be fixed. I don't understand nearly enough about electronics design to explain why.

Some differences I can find online:
  • CDNA has higher compute density, i.e. a greater % of the die area is for doing math
  • CDNA has smaller warps (16 vs 32). A lot of HPC & AI code has branching and other exceptional behavior that cause warps to stall, so smaller warps execute more efficiently.
  • RDNA is designed for low latency, while CDNA is designed for high bandwidth. Latency doesn't matter much for HPC or AI.
  • CDNA cache was more optimized for bandwidth than capacity

GPU architectures in general tend not to be good at it. That's why they're GPUs and not CPUs.

NVIDIA GPUs are very good at it, and by "general compute," I mean generic computations on large arrays of data, or what used to be called GPGPU.
 
Shaders wouldn't work if they did. All I know is that when it came to ingesting large arrays and executing arbitrary code, it was woefully inefficient and couldn't really be fixed. I don't understand nearly enough about electronics design to explain why.

Some differences I can find online:
  • CDNA has higher compute density, i.e. a greater % of the die area is for doing math
  • CDNA has smaller warps (16 vs 32). A lot of HPC & AI code has branching and other exceptional behavior that cause warps to stall, so smaller warps execute more efficiently.
  • RDNA is designed for low latency, while CDNA is designed for high bandwidth. Latency doesn't matter much for HPC or AI.
  • CDNA cache was more optimized for bandwidth than capacity



NVIDIA GPUs are very good at it, and by "general compute," I mean generic computations on large arrays of data, or what used to be called GPGPU.
From some quick reading it seems like CDNA is like a really fast cpu(I am most likely wrong), while RDNA functions more like a typical gpu.
 
Is it that by default, when GPU's constantly whiplash between boosting to the highest clock possible, then overcorrect it to a safer one, they end up being less performant than if you were to force them at a given nominal frequency and not let them exceed it? Is that what I'm noticing?
Yep. I think the idea is games are so badly optimized, boosting high during a bottleneck and lower during light utilization gives better a framerate. But in practice it doesn't work out.

There's also the fact they design them with stupid high boosts so overclockers can get extra performance for the cost of a ridiculous TPD that they could never justify out of the box.
 
Is it that by default, when GPU's constantly whiplash between boosting to the highest clock possible, then overcorrect it to a safer one, they end up being less performant than if you were to force them at a given nominal frequency and not let them exceed it? Is that what I'm noticing?

Close, but it's not an overcorrection. They boost to the highest clock possible, which gets the GPU absolutely blazing hot, which then requires throttling it back so it can cool down. This results in the average speed being less than if you just held it at a higher, but thermally safe setting. CPUs do the same thing with even more extreme clock variation.
 
  • Like
Reactions: indomitable snowman
Close, but it's not an overcorrection. They boost to the highest clock possible, which gets the GPU absolutely blazing hot, which then requires throttling it back so it can cool down. This results in the average speed being less than if you just held it at a higher, but thermally safe setting. CPUs do the same thing with even more extreme clock variation.
prob why undervolting is all the rage in the overclocking scene. helps cull thermals and power draw a little.
Doesnt help that nigvidya has been pushing for the housefire connector while jacking the hell out of the power draw each gen
 
prob why undervolting is all the rage in the overclocking scene. helps cull thermals and power draw a little.
Doesnt help that nigvidya has been pushing for the housefire connector while jacking the hell out of the power draw each gen
even just dropping down the power limit from 100% to 80% can save a ton of power while barely hitting the framerate, a lot of cards are totally juiced out of the box
 
even just dropping down the power limit from 100% to 80% can save a ton of power while barely hitting the framerate, a lot of cards are totally juiced out of the box
Yeah, i dropped power limit to 450W since im a bit paranoid when it comes to the power connector. Card also isnt as much of a furnace compared to stock
 
even at 450w the 5090 still smokes every other card on the market
Listen, when it comes to getting the coveted Soyface Award, you can't be too careful. One day, you're running your card at a wattage that doesn't melt connecters, and the next, your competitor releases a card juiced so hot that all the YT thumbnails with your card are getting the dreaded scowling head smack.
 
Listen, when it comes to getting the coveted Soyface Award, you can't be too careful. One day, you're running your card at a wattage that doesn't melt connecters, and the next, your competitor releases a card juiced so hot that all the YT thumbnails with your card are getting the dreaded scowling head smack.
awww man! I wish i had done a shunt mod to my $2000+ graphics card to get an extra 5 fps in UE5 slop. Clearly that cobson lookalike called frame chasers has the right idea
 
TSMC Arizona’s Male Technician “Patted On The Buttocks” While Colleague Found Rubber Chicken Hanging Over His Desk, Alleges Lawsuit (archive) (Discuss at A&N!)
A lawsuit against the Taiwan Semiconductor Manufacturing Company (TSMC), which was initially filed last year, has been expanded to include 17 plaintiffs who allege a host of discriminatory and unsafe practices at the company. The suit surrounds TSMC's much-hyped Arizona plant, with the plaintiffs, all of whom are American citizens, alleging that the firm discriminated against them during the time they spent at TSMC Arizona. The allegations include a preference for Mandarin or Chinese during the hiring process, routine bias against non-Taiwanese employees and unsafe working conditions.
TSMC-TAIWAN-US-ENGINEERS-BABIES-NOVEMBER-2022-728x362.webp
A snippet from Taiwanese social media platform PTT taken in November 2022 shows some thoughts about TSMC's American engineers shared by locals.

Uppity American swine trying to derail Taiwan's miracle factory. Sad.
 
Last edited:
Back