I'm adapting a post I made in another thread, since the topic of Maxwell's efficiency greatly interests me and pisses me off at the same time (you'll see why later).
As you may know, Maxwell's efficiency comes from its aggressive dynamic micro throttling, which adjusts the GPU core voltage based on load. This micro adjustment happens so rapidly that on a macro (second long) timescale, we only see the "averaged out" value that's reported in Afterburner. Case in point:
![]()
See how the power consumption constantly fluctuates, going from over 250W (!) to less than 100W in the timespan of 1 second? Yeah that's what I mean by micro adjustment. This constant, rapid adjustment of voltage is the source of Maxwell's efficiency, but unfortunately is also its Achilles heel. Let me explain why with the help of Maxwell BIOS Tweaker.
I'm going to show a few pictures of the boost and voltage tables of my Gigabyte 970's stock vbios. Let's go step by step and start with the boost table:
![]()
This one is fairly straightforward and doesn't really need much explanation. Each value in the table represents a particular boost/clock state. The only thing to note here is clock states #35 through #74 - highlighted in yellow - belong to the P0 (full load) state of the GPU.
Now let's look at some voltage tables:
![]()
![]()
![]()
![]()
![]()
Let me go a bit slower here so everyone will follow. Each entry in this voltage table corresponds exactly to the particular clock state in the boost table. So for example CLK 40 corresponds to a boost clock of 1076 MHz by the boost table. Now you'll see that each CLK state has a fairly wide voltage range, starting from about 106mV (CLK 35), all the way up to a ridiculous 219mV (CLK 60).
For my particular 970s, running each of them solo, one boosts to 1380 (CLK 63), and the other 1405 (CLK 65).
Now let's look at the corresponding voltage entries for CLK 63 and CLK 65.
![]()
You can see both states have a defined upper limit of 1.281V, while the lower limit is 1.075V for CLK 63 (1380 boost), and 1.081V for CLK 65 (1405 boost).
As you can imagine, trying to push almost 1400MHz on the GPU core with a measly 1.081V is simply going to end in tears. Now what I don't know is how exactly, through what algorithm, the vbios picks the exact voltage to use for each clock state. Actually on second thought, the algorithm is most likely programmed into the driver, and the vbios simply delineates what voltages are "allowed" for what boost clocks. In any case I'm going to wager a guess that the driver adjusts the voltage dynamically based on load, the "micro adjustment" I mentioned in the opening paragraph. I'm also going to presume the voltage range is set so wide for each clock state to maximize efficiency, by giving the core as little power as possible based on what nVidia engineers have somehow determined to be the minimum stable working voltage for each clock state. (pure speculation on my part here)
The very unfortunate thing here is, because the voltage range for each clock state is set so wide, sometimes the voltage simply gets stuck at the lower limit of that clock state, and doesn't ramp up fast enough to keep up with the GPU core under load, which results in crashing without any forewarnings (artifacts, glitches etc).
From my own experience, this is especially prone to happen right after a non-demanding cutscene, where the core is basically chilling out, and then is immediately thrown back into action when the cutscene ends. What typically happens - as I've observed from Afterburner's OSD - is that the boost clock shoots right to where it should be due to the suddenly increased load, but the voltage is either stuck at the lower limit of that particular clock state, or worse, stuck in the voltage range of a lower clock state. This is what is referred to as a "boost/voltage table crossover" in the desktop community, and is CRAZY ANNOYING.
So suffice to say the only fix to this dynamic micro throttling garbage, is to clamp both the lower and upper voltage limits to the same value for each clock state #35 through #74. This way it pretty much guarantees that no matter where the GPU is at in the boost table, it will always be delivered a constant voltage (I set mine at 1.25V), ensuring 100% stability regardless of load and what not.
But of course in doing so efficiency goes out the window, and the TDP numbers don't look as nice anymore. Not that I give a damn, but does make me suspect this is yet another trick nVidia pulled with Maxwell to make it look more impressive from a performance per watt standpoint at the expense of 100% stability. And DO NOT get me started on the 970 3.5GB vram + 500MB G-Cache™ issue.
Hopefully all that made sense to you guys. So what do you think, have I lost my marbles, or is this really what it is?
Hey look I wrote a book. @D2 Ultima: You're not the only one writing books ha. (although you might still be the only one writing books at 2am XD)
-
-
-
Not surprising. Intel and AMD does this too. They are all doing crud for numbers on paper but sacrificing actual use. Look at AMD, their hybrid parallel CPU-GPU APU seem great on paper, but suck in practice. Intel's switchable graphics is great on paper, implementation and actual results? Rubbish. Don't see why it would be any different with NVidia.
Efficiency at these levels of performance I always assume some gimmick is being used. I don't think any genuinely real improvements and advances in technology have happened in graphics technology in 10 years, it's just one gimmick after another. Thankfully performance improvements are real, but the battery saving without sacrificing performance claims are BS all around.
You are right. Whether I'm using dGPU or Switchable, the fluctuations between min-max-avg is out of control for years now. The min are really low, and highs really high and averages are meh at best because the min are so low. Demanding games like Crysis 3 suffer, Max of 90? Wow... but the lows of 30s-40s keep my avg framerate under 60 with an overclocked 980M.Last edited: Mar 17, 2015Vasudev likes this. -
Then again, this also brings to light the whole DX12 troubles. If DX12 can pull all from a CPU, what's going to happen to mobile maxwell users on small power bricks? P650SG might as well be running a 20W brick when that happens.
Vasudev likes this. -
-
Anyway, I went with 330W just in case of power issues for the 4790K. -
Worst case scenario Clevo will have to refund all of us owners for selling us faulty hardware. We'll still be covered under warranty. Or Microsoft will have to foot the bill for providing a dangerous, faulty update. -
D2 Ultima likes this.
-
-
-
You're basing your assertion off that Star Swarm graph. What it didn't show was that under DX11, the GPUs were being bottlenecked badly, so usage was much lower. Remove the CPU bottleneck in DX12/Mantle, GPU usage shoots up to 99%, and system power consumption goes up as a result.Cakefish likes this. -
-
-
-
-
-
Microsoft will also face severe backlash if Windows 10 literally kills PCs. Even desktops don't be safe from increased power draw. Many PSUs could fail and it will be the biggest scandal in the PC industry in years. Windows 10 will go down in history as an even bigger disaster than Windows 8. Their reputation will be irreversibly damaged.
So... basically it'll spell doom for Clevo, other OEMs that use small PSUs, desktop PCs, Microsoft, NVIDIA and AMD. Considering what's at stake I can't see this coming to fruition.
I'm sure my 980M will keep on chugging along quite nicely under DX12 - it has to for the sake of PC gaming -
-
-
And the affected laptop OEMs will certainly suffer the consequences regardless. How could consumers trust them ever again if suddenly our PSUs (and potentially whole systems) end up being bricked en mass just because of a simple OS update? I personally would be absolutely furious.
However, I doubt they would let such a scandal ensue. Microsoft, NVIDIA and AMD must have done extensive testing of the API. Any issues should be resolved before release.
It'd be dark days indeed for a simple API update to brick laptops. -
Last edited: Mar 17, 2015
-
Much ado about nothing
-
If you can run Furmark/Kombuster you're probably safe. GPU usage will be much more efficient, but the game still has to use the CPU for countless other tasks.
-
I'm using dramatic language, but it really would be a huge deal in my opinion. -
In one year's time I will read this thread and laugh as I play DX12 games on my overclocked P650SG with 240W brick
-
in 1 year's time you'll likely have forgotten about this XD
-
Exactly, because all this scaremongering and then it turns out to be a non-issue.
BTW 3DMark14 low overhead API (DX12, Mantle) test inbound. Let's put these power consumption fears to rest once and for all.
-
so people are worried about overloading PSU when the API got MORE efficient...? this shouldn't be happening....
-
Well NVIDIA have most certainly advertised TDPs of their desktop cards so DX12 can't increase TDP of GPUs or else they will be sued for false advertising (again).
You can't judge power draw from Star Swarm - which is the DX12 equivalent of Furmark. It's designed to push the system way beyond reasonable workloadsLast edited: Mar 18, 2015 -
Oh wait, look:
Crysis 3 and FurMark on the exact same test bench use more power. -
Ah OK then, I stand corrected!
Still, I really doubt all parties involved would be so foolish so as to intentionally brick countless laptops and desktops. The backlash would be catastrophic. We'll be fine -
It's easy to smell FUD from a mile away
-
I just learnt a new acronym!
B-b-but DirectX 12 is supposed to be the messiah of PC gaming, not the harbinger of its doom -
-
^TIL
-
CptXabaras Overclocked, Overvolted, Liquid Cooled
That's the reason why i modded my 980's bios and from clock bin 74 to 34 i've set min/max at 1.250V / 1.250V.
No fluctuation at all.
Under heavy benchmark, i've seen almost 675 Watt pulled from the power outlet, with the config in my sign. Not that different form when i was running crossfired 7970 Ghz edition cards. -
Damn and the 7970 GHz was a 275W card
-
CptXabaras Overclocked, Overvolted, Liquid Cooled
Any way i don't really care, i went with 980's without thinking about power efficency, but performances. Watercooled them and hardly pass 42 degrees celcius, under full load. (happy panda!!)
On the subject of Maxwell GPUs' efficiency
Discussion in 'Gaming (Software and Graphics Cards)' started by n=1, Mar 16, 2015.