I've got a quick question...
When I connect my DM4 with a 5450 to my external monitor via HDMI the clocks obviously jump. If the powerplay setting is at maximum battery, it's pegged at 300/300mhz (gpu/mem). If I change it to maximum performance, the minimum it runs at is 400/800mhz and it can jump all the way to 675/800mhz which is it's max. Looking at the temperature difference, there only seems to be a 1-3 C difference at most. For the long-term health of the laptop/gpu, would running it at maximum batter (300/300) be better? Or due to the little difference temperature wise, there really wouldn't be that much benefit. To be honest due to the 100/500mhz difference, I thought there would be a more significant temperature difference, but there isn't. Keep in mind that I will be running it via HDMI for a significant period of time, basically all day. Any thoughts would be appreciated. Thanks.
Oh, and I won't be gaming. Just going through slides and using Word/iTunes.
-
For the DM4, there should be no significant difference. The GPU still is well within it's thermal and power draw specs. The laptop's mobo is more likely to fail of a bad cap instead of the relatively durable silicon substrate GPU dies from TSMC (and most of the world) use.
TSMC and UMC being AMD's GPU division manuf. partner. -
What's the thermal limit on the GPU? I know for my i5 it's 105C. Having the GPU run at 47-50C for extended periods of time is fine, right? Thanks man.
-
100C, if it follows the Desktop variant (HD5500-HD5600 series/"Redwood" core)
-
Panther214
ATI powerplay Q
Discussion in 'Gaming (Software and Graphics Cards)' started by nu_D, Jan 2, 2011.