Yes. With Task Scheduler you have it load up based on user. So you'll have to set it up for each user. Not a big deal really.
-
I have the colons in there, but not the start in. When I had that, it told me incorrect directory (in task manager). I'll play around with that some more. Thanks HT.
-
How do I update the graphics driver? Am I supposed to use the one from HP or AMD? Instructions please? 6620g only
-
You should probably google up the 12.1/12.1a drivers and install them directly over what's on the system.
-
In Task Scheduler where you tell it start up directory, it should point to the directory of the K10stat.exe that you're using.
Install HP drivers. Then download and install the latest AMD drivers. -
HP Pavilion dv6z Vs. dv6t Quad Edition, Which is Best?
Cute article, have a laugh
(I guess too much computing trash their math)
-
Stupid me. I was pointing it to the right place, but without the C:\. The screen shot in the tutorial didn't show it, and I wasn't thinking properly! Got it going fine now. Thanks again!
-
Wow, I finally managed to read the last 100 posts, that gathered around christmas ^^.
I used the holidays to upgrade my screen to the 1080p one and to renew the thermal compound.
The latter didn't prove to be much helpful, since the cpu is reaching the same temps as before(just a few mins later), and although the gpu runs a little bit cooler, it doesn't reach higher clocks :-(. But as someone here allready stated, the factory pasting job for the european models ought to be not that bad...
But the FHD screen was the best upgrade choice there is! The screen with it's 4x(+) contrast is so much more crisp and sharp. And I finally could activate the EDID which used to make the 768 screen colors way to much green. It's like lifting a gray veil from the screen ^^.
Although... 2 things did make me curious.
1. If watching pictures like this in fullscreen, I can see some interferences in the dark areas, like when using an external FHD screen with an analog vga port.
Is that the same for everyone? I don't really see this in everyday gaming, just when looking really carefully at the screen. Maybe I pushed the FHD cable to hard somewhere...
2. I thought of doing some benchmarks the measure the performance decrease of 1366x768 vs 1920x1080, and 768 with antialiasing vs 1080p without. But the problem was that the benchmarks with the new 1080p screen were all somehow higher!
I used different scenes from anno2070 and skyrim and the fps increased between 10-15% although the resolution doubled! Of course the 6750m has fast gddr5 at hand, but there can't be an increase!
Strange thing is, when benchmarking skyrim in 136 0x768 with the new screen now, I get almoast 50% more fps than with the old panel's 136 6x768.
That means I eighter just got a somehow borked benchmark run with the old screen (most probably), or the catalyst somehow can't display the odd 136 6x768 resolution in skyrim that well.
But you definitely get better visual quality with 1920x1080 an no aa, vs 1366x768 with 4xaa.
So nevertheless, the 1080p screen is worth every penny
.
And to two other postings that were made since I last had time to post:
There are a lot of different sites right now, that post different specs. The true specs will probably be released at the CES. But as far as it looks with the Envy 15"'s benchmarks, it's just a rebranding (with same 40nm process & VLIW5 design, and even the same speeds), like with the desktop cards ranging from HD 70xx to 76xx. (amd source here: OEM Solutions)
Seems like the thread is more and more becoming cluttered/overloaded, because i think this was asked before.
Like here:
Although I don't remember how that discussion ended
.
But afaik, the A75M (aka Hudson M3) have one SATA-III port for the internal drive, and one SATA-II for the optical one.Attached Files:
-
-
well about k10stat I always keep all state fixed to 2400mhzbaiii said: ↑Sound like the CPU throttle on you when the CPU+GPU load go high.
Try either of following:
1. Use k10 and keep all your pstat the same
2. Turn off CnQ via advanced bios (if you flashed that one)
I currently edited my registry to lessen the throttle but It still throttle on me .( But I like the way window clock control works)Click to expand...
temps is about 82° on full load cpu 2700mhz and gpu oclocked to 800/930
so I tried turn off CnQ via advanced bios mod... this broke my crossfire again... but no more spikes even on large maps via 6750m o-clocked to 800/930... still some FPS drop, but I guess it's because BF3 is intended for desktops and gaming laptops... I'm still thinking about dv6t and core i7... (last day i can return my dv7z for a dv6t)...
what's the bios are you using in you dv6z?
and how to lessen the throttle in the registry? -
strange... my crossfire is partially working (some fps drop on bf3, but when cf is on i have 8+fps, no more shuttering) but only when the 6620 in the CCC is set to high performance... I also upgraded the ram to 8gb Corsair vengeance 1600mhz CL9, maybe this helps...HTWingNut said: ↑Yep, did all that, still stutters. I just turn crossfire off. It just doesn't work well the way it was intended. Every driver release I try again and still fubar.Click to expand...
I also noticed that if I set the game to high quality with CF on the framerate on top render 2 different value when fps drops.. that's probably because of AFR .... seems like the igpu bottleneck the dgpu because of it's weak core clock
we have to manage how to edit bios to overclock the igpu.... the 6620 is too weak to be useful for CF... -
From Passmark:
PassMark - G3D Mark - High End Videocards - Updated 7th of January 2012
6620G – 464
6750M – 1130
Ratio of performance is then:
1130 : 464 = 2.4 : 1
Again, from HardOCP’s real-world testing:
“There is a limit as to what you can run in a Dual Graphics situation...Once the discrete GPU starts outstripping the on-die APU by about 1.75 frames to 1, CrossFireX basically stops working at that point."
So, the performance ratio at which CrossFireX stops working is:
1.75 : 1
…so what this means is that Hybrid CrossFireX is only useful when two cards’ performance are within this margin, which, by Passmark’s testing clearly shows they are not. (Conclusion: Hybrid WORKS when it’s used as designed with two closely matched cards.)
What this doesn’t tell you is that, outside this ratio, Hybrid CrossFireX actually hurts performance because every time the 6750M says, “Hey, I could use a little help here,” the 6620G simply isn’t up to the task of filling in a frame at the same speed the 6750M is capable of. The result? High frames rates crippled by stutter, the main complaint noticed by many users and reported by professional reviews.
The problem is that AFR (Alternate Frame Rendering, used in CrossFireX and SLI) doesn’t work so well with overly-mismatched cards. A better rendering scheme would be for each card to do their part with each and every frame, in the proportion of their relative performance. Consequently, about 70.5% of a frame should be allocated to the 6750M and 29.5% to the 6620g. At this point, neither CrossFireX’s or SLI’s drivers are written that way. After all, both were originally written to take advantage of matched cards.
So, as is, you’re going to get the best combination of performance (frame rate and smoothness) by using the 6750M alone until drivers are rewritten and profiles are written to take advantage of a non-AFR driver scheme. This requires individualized game settings (profiles) that give you 30FPS or above, in the combination users find most pleasing.
If you're going to use Hybrid CrossfireX in it's current incarnation, for best performance you need to OC the 6620G and leave the 6750M alone, in order to close the ratio to reduce stuttering. (Lowering the 6750M to close the ratio would cripple performance, so this approach wouldn't really optimize performance.) The problem with this approach is that you would probably get even better performance by turning CrossFireX off, bypassing the 6620G, and OC'ing the 6750M.
(For you gaming enthusiasts, this might be a worthwhile experiment! The trick might be to properly synchronize clock speeds between the iGPU and the dGPU in order to get the optimal mix, instead of maxing out both clock speeds.)
This doesn’t make the 6620G useless, however. For all non-intensive graphics performance, it’ll still save you electricity! Remember that this was actually the original intention of the dual graphics setup for the AMD A series notebooks!
Bottom line: You bought a Volkswagen for the price of a Volkswagen. They're fun and inexpensive to "trick out" to optimize performance. But even at peak performance, don't complain that it it isn't a Porsche (especially while you're passing all the Chevys and Fords on the road.). If you want a Porsche, go buy one, but bear in mind that you'll pay for one! -
I'm still not sure what you mean about setting the 6620 in the CCC to high performance?davidgreen said: ↑strange... my crossfire is partially working (some fps drop on bf3, but when cf is on i have 8+fps, no more shuttering) but only when the 6620 in the CCC is set to high performance... I also upgraded the ram to 8gb Corsair vengeance 1600mhz CL9, maybe this helps...
I also noticed that if I set the game to high quality with CF on the framerate on top render 2 different value when fps drops.. that's probably because of AFR .... seems like the igpu bottleneck the dgpu because of it's weak core clock
we have to manage how to edit bios to overclock the igpu.... the 6620 is too weak to be useful for CF...Click to expand...
Also, it's not really that the 6620 is too weak, they need to make it so it's proper ratio of performance to processing. It would be great if this was default at 6620g and 6750m clocks, but user adjustable for overclocking. In the meantime I'm sticking with just the 6750m overclocked. Runs smooth and no headaches. Just wish I could get the little extra boost from CrossfireX that is *should* provide.
edit: Do you mean this screen and what setting?
or this one, because this one below has nothing to do with CrossfireX:
-
first screenshot.. on power saving graphics tab... set all tabs to max performance.. leave smoothvision settings application controlled, and select the bfbc2 preset under crossfirex profiles tab....HTWingNut said: ↑I'm still not sure what you mean about setting the 6620 in the CCC to high performance?
Also, it's not really that the 6620 is too weak, they need to make it so it's proper ratio of performance to processing. It would be great if this was default at 6620g and 6750m clocks, but user adjustable for overclocking.
edit: Do you mean this screen and what setting?
or this one, because this one below has nothing to do with CrossfireX:
Click to expand... -
Any one of you guys owned the dv6 QE AND the dv6z QE?
Which would you recommend for a friend? The dv6z seems cheaper but I've always bought laptops with Intel cpu's and I have no experience with AMD....lastly this computer will eventually pass on to a younger brother once Ivy Bridge makes it ways into the Envy line up later this year.....
Any input would be greatly appreciated -
Thanks for the input. Tried all that. 45 fps feels like 15 fps with crossfire. I'll take 30fps with just the 6750m any day. So much more smooth. I'd really like 1920x1080 even at all low, but alas, I'm acutally playing at 1600x900 all low with just the 6750m at decent framerates.davidgreen said: ↑first screenshot.. on power saving graphics tab... set all tabs to max performance.. leave smoothvision settings application controlled, and select the bfbc2 preset under crossfirex profiles tab....Click to expand...
The nice thing is with 12.1 drivers you can overclock the GPU again using tools like MSI Afterburner.
If you're comfortable with tuning and overclocking, then the DV6z is a good choice and save you a couple hundred bucks. But if you don't want to spend the time to tune your CPU then I'd say go for the DV6t QE. The Intel Quads are definitely more powerful, and just run without need to overclock/undervolt, etc. But on the other hand the DV6z can also get better battery life since you can significantly undervolt and run at slower clocks on demand.TheAtreidesHawk said: ↑Any one of you guys owned the dv6 QE AND the dv6z QE?
Which would you recommend for a friend? The dv6z seems cheaper but I've always bought laptops with Intel cpu's and I have no experience with AMD....lastly this computer will eventually pass on to a younger brother once Ivy Bridge makes it ways into the Envy line up later this year.....
Any input would be greatly appreciatedClick to expand...
As far as GPU, many have been able to get the 6750m close to stock 6770m speeds. Although the 6770m should overclock a bit as well. -
DV6 with AMD + Tunning = i5 equivalent with xfire graphics (warning, for some games at 1080p like battlefield tends to act not so nicely)TheAtreidesHawk said: ↑Any one of you guys owned the dv6 QE AND the dv6z QE?
Which would you recommend for a friend? The dv6z seems cheaper but I've always bought laptops with Intel cpu's and I have no experience with AMD....lastly this computer will eventually pass on to a younger brother once Ivy Bridge makes it ways into the Envy line up later this year.....
Any input would be greatly appreciatedClick to expand...
DV6 with intel quad = more raw processing power, slightly less graphics.
At the end the difference is in price and also how much do you like to tune laptops. -
so I should be able to overclock to 2.6 upwards to 2.8 while still undervolting from the stock defaults and still be about the same battery life?
Just did a few minutes of BF3.. auto settings are pretty good.. mostly medium with ultra on textures runs at a good frame rate, though this at the beginning of the sp. only miss my 5.1 logitech system on my desktop lol.
Left cpuid open during that time, apu seen up to 89c temp -
I personally have many profiles, a low one ranging from 800 mhz up to 2.2 ghz with all the values undervolted, when I want battery I use that one.hypertek said: ↑so I should be able to overclock to 2.6 upwards to 2.8 while still undervolting from the stock defaults and still be about the same battery life?
Just did a few minutes of BF3.. auto settings are pretty good.. mostly medium with ultra on textures runs at a good frame rate, though this at the beginning of the sp. only miss my 5.1 logitech system on my desktop lol.
Left cpuid open during that time, apu seen up to 89c tempClick to expand...
Of course more profiles from moderate, high, very high an extreme change the first and the last p state, up to 2.9 Ghz. -
Good luck with anything above 2.4GHz. Temps start to skyrocket and/or throttling engages causing severe performance issues.
-
I do 2.5p0 and 2.7b0 for daily desktop use. Lock 2.8 with k10 if need some serious CPU power (no mod but ambient temp here is like 40F)
The worst thing about this CPU is that it doesnt throttle base on current temp, It throttle when cpu+gpu load is high which is kinda ...
and b0 is kind of pointless cause it never switch itself when load get high. (It get boost up when load is low? I dun get the logic behind that)
I probably try to set p0 2.4 and see if that helps. -
im going to go ahead and cut case holes, ill keep you guys posted
-
I am not sure you're right here. My understanding is that it will throttle based on how much heat the CPU is producing versus how much it's getting rid of. If you're running at anything over 2.4GHz the CPU will start making more heat than the heat sink can get rid of and as the junction temps rise it'll auto throttle. It just does it so fast that you don't see the very short temp peak that made it throttle.baiii said: ↑I do 2.5p0 and 2.7b0 for daily desktop use. Lock 2.8 with k10 if need some serious CPU power (no mod but ambient temp here is like 40F)
The worst thing about this CPU is that it doesnt throttle base on current temp, It throttle when cpu+gpu load is high which is kinda ...
and b0 is kind of pointless cause it never switch itself when load get high. (It get boost up when load is low? I dun get the logic behind that)
I probably try to set p0 2.4 and see if that helps.Click to expand...
That said, a 2.4 or 2.5GHz OC is PLENTY of CPU power for most games. -
If that were true, a burntest would show this. But it does not, even with an OC. It's based on utilization of the CPU. P-states work behind load states. The throttling happens a lot if you don't set your utilization update interval. All cores will run at P1 if core load = 35% for example. If the throttling was managed by only temps, then why does the temperature keep rising with loaded cores if it's suppose to throttle down? It doesn't because the cores load is still high %.
The boost feature really only jolts the lowest loaded core (duration) into the top for a few seconds to achieve it's highest state before falling back into the rest of the running cores current P-states. Reason it doesn't hold it there long enough is because AMD knew it would overheat if it ran in BOOST longer than at STOCK. Same goes with any of the models. They can't sell their highest model if it doesn't meet the requirements, so they sell it as a lower model. Reason we all inherit the same problem. Anyway, that's why they called it "boost." Because it does boost lowest loaded core up depending on all cores %. It's sad, in terms of ability but in verdict it does what it sells for.
Massive amount of cooling is necessary to get into the big shiny 3 at 100% load. The reason most people see so much throttling happening, is because games... programs... don't stay locked into the same core % load, they fluctuate! from 49-62-80% depending on what's happening on screen. And that's why you notice your OC clock going down into the lower 2Ghz area for a brief second to come back into the OC clock again. Lock all cores or set all P-states the same, either way you will notice that it won't throttle at all. Just make sure you have adequate cooling for your OC. Heat kills. I'm at 3Ghz at over 75 Watts. I'm over the TDP of 45.
Now the iGPU load working with CPU load, this is entirely different! (APU) This is what determines the P-state at which the CPU is allowed in. If the iGPU has a high load, then the CPU won't be able to get into it's highest P-State or hold it for very long. This is another part of the problem (stutter) with CrossfireX in these Llano chips. It's why these chips make for great battery life because most of the time, the CPU isn't even reaching it's max clock of stock 1.9Ghz (because of the 6620G load.) It's running below or around longer than it's peak, this coupled with Cool'n'Quiet. The iGPU plays a huge role in this is all I'm saying. Again, all controlled by utilization values and running applications.
Disable the dual-graphics, run discrete and OC! -
Not true. I have done quite a bit of abusive testing with my DV6z and a K53TA. It will throttle despite setting all P-states the same. It's implicit in the CPU itself. It WILL throttle down to like P3 state on all four cores for several seconds until temps get under control. It exhibits this when taxing it with Prime95 where all cores are at 100%. It has more to do with rate of temperature increase along with the actual temperature.dreamonic said: ↑If that were true, a burntest would show this. But it does not, even with an OC. It's based on utilization of the CPU. P-states work behind load states. The throttling happens a lot if you don't set your utilization update interval. All cores will run at P1 if core load = 35% for example. If the throttling was managed by only temps, then why does the temperature keep rising with loaded cores if it's suppose to throttle down? It doesn't because the cores load is still high %.
The boost feature really only jolts the lowest loaded core (duration) into the top for a few seconds to achieve it's highest state before falling back into the rest of the running cores current P-states. Reason it doesn't hold it there long enough is because AMD knew it would overheat if it ran in BOOST longer than at STOCK. Same goes with any of the models. They can't sell their highest model if it doesn't meet the requirements, so they sell it as a lower model. Reason we all inherit the same problem. Anyway, that's why they called it "boost." Because it does boost lowest loaded core up depending on all cores %. It's sad, in terms of ability but in verdict it does what it sells for.
Massive amount of cooling is necessary to get into the big shiny 3 at 100% load. The reason most people see so much throttling happening, is because games... programs... don't stay locked into the same core % load, they fluctuate! from 49-62-80% depending on what's happening on screen. And that's why you notice your OC clock going down into the lower 2Ghz area for a brief second to come back into the OC clock again. Lock all cores or set all P-states the same, either way you will notice that it won't throttle at all. Just make sure you have adequate cooling for your OC. Heat kills.Click to expand...
Normal core operation it will bounce around a bit, but that's not throttling. That's just core load.
The iGPU has priority over one core only. You can see the correlation by running something like MSI Kombustor at the same time as Prime95. Three of the cores will be at P0, where one will bounce around. That one bouncing around is headroom for the iGPU, the lower the P-state (or higher number really), the more resources the iGPU is consuming.Now the iGPU load working with CPU load, this is entirely different! (APU) This is what determines the P-state at which the CPU is allowed in. If the iGPU has a high load, then the CPU won't be able to get into it's highest P-State or hold it for very long. This is another part of the problem (stutter) with CrossfireX in these Llano chips. It's why these chips make for great battery life because most of the time, the CPU isn't even reaching it's max clock of stock 1.9Ghz. It's running below or around longer than it's peak, this coupled with Cool'n'Quiet. The iGPU plays a huge role in this is all I'm saying. Again, all controlled by utilization values.
Disable the dual-graphics, run discrete and OC!Click to expand... -
I've had my burntest on max running at default 5 times. I watched my cores, not once did I see it throttle down a state. It more like throttled at 50-80mhz lower than my OC mark - this what you are referring to?
-
OK, need some more help, this time with Kombuster. I've chosen to run the DX11 one from the start menu, but it still only shows the 6620. How can I get it to stress the 6750? It has the icon for 3 gpu's but only tags gpu 2 & 3. It labels gpu 2 as the 6620, and gpu 3 has the 6620's clock speeds. All it says about gpu 1 is that it is a 6000 series. When you toggle the graphs, anly 2 & 3 show up. What's the deal? It does the same whether I run it from Afterburner, or choose any option from the start menu. Help!
edit: Also, which option do you choose in the Kombuster window? Gpu burn-in, Kmark normal or extreme, or something else entirely? -
Use Prime95. Intel Burntest will pass when Prime95 Large FFT will fail, and also results in real usage failure too.dreamonic said: ↑I've had my burntest on max running at default 5 times. I watched my cores, not once did I see it throttle down a state. It more like throttled at 50-80mhz lower than my OC mark - this what you are referring to?Click to expand...
Make sure Kombustor is set to "High Performance" in switchable graphics. You need to either "prime" Afterburner with an app set to "high performance" and then start Afterburner or disable ULPS. I have registry entries here: http://forum.notebookreview.com/hp-...xxx-series-owners-lounge-227.html#post7861634goresnet said: ↑OK, need some more help, this time with Kombuster. I've chosen to run the DX11 one from the start menu, but it still only shows the 6620. How can I get it to stress the 6750? It has the icon for 3 gpu's but only tags gpu 2 & 3. It labels gpu 2 as the 6620, and gpu 3 has the 6620's clock speeds. All it says about gpu 1 is that it is a 6000 series. When you toggle the graphs, anly 2 & 3 show up. What's the deal? It does the same whether I run it from Afterburner, or choose any option from the start menu. Help!Click to expand... -
Alright, Prime95 it is.
-
Not sure about disabling ULPS, but I suppose I could give that a try. Kombuster is set to High Performance. What do you recommend to prime it with? So far the only game I have on here after the new install is Civilization I! Used to have Civ V, but will have to reinstall that.HTWingNut said: ↑Make sure Kombustor is set to "High Performance" in switchable graphics. You need to either "prime" Afterburner with an app set to "high performance" and then start Afterburner or disable ULPS. I have registry entries here: http://forum.notebookreview.com/hp-...xxx-series-owners-lounge-227.html#post7861634Click to expand...
-
Then if all P-states are the same, say, 2.8Ghz. How exactly does it throttle down? Are you saying it overrides the user made clock changes to shipped clocks.. "throttling" inherit within the CPU itself? I don't think so. Mine still shows steady in Prime95. I'm using FusionTweaker for the clocks. I noticed when K10Stat was in control of it, it was all over the place.HTWingNut said: ↑It WILL throttle down to like P3 state on all four cores for several seconds until temps get under control. It exhibits this when taxing it with Prime95 where all cores are at 100%. It has more to do with rate of temperature increase along with the actual temperature.Click to expand...
Mine will not do this. Given core load, given P-state clock. Higher P-states, mean higher core loads. Higher Temperature, means all of the above. Mine does not throttle down on any of these tests... because they are at 100% core load, so it's using the highest frequency I've set. It'll reach it's max temp if I purposely put it into that scenario and then will it BSOD as it cannot auto-throttle down any P-states if they are all the same for example. I only see it throttle down P-states on core load percent, not temperature itself, which coexist with higher (lower number) P-state values. If it was temperature, then why can I run 3Ghz at 90 degrees and not throttle down due to high temperature? It's because the cores are set to 100%!
Even if I have a lower P-state it will still stay at it's max P-state for 100% core load, temperature doesn't matter until....
It's not within junction temp - which if reached, BSOD or auto-throttle (depending on DEFINED P-states values only.)
Now I know I'm missing the latency, field, timer, instructions, macro instructions, (ms) time for each request to be used on each core and how they can work with each separately based on the P-state they are in, for how long, when they got there, where the other cores are instructed to next, changing given values to another request and so forth. But based on load factors, which pertain to load amount, request time, field size and instruction, it'll either reach P1 sooner and have to wait till it can lower itself to P2. This all depends on the handles each core has too. That is why if you load a browser in P7 state, you get P1 state because a given core(s) were able to accept a new request given the timed amount it was in idle mode in that state. Then as you go beyond a certain P-state, it becomes core load based.
No?
Not trying to come across sarcastic, just trying to understand what we both see here. -
grill provided by a mini desk sorter from walmart for $2.88
something like this but mine is a different type (cant find the exact one on their website) Walmart.com: Rolodex Mini Sorter, 3 Sections: Office -
Hi.I have a 8-3510mx,trotting on 77*c
This is norm?
-
You need to use gpu burn in, run it windowed, and than run amd system monitor and you will tell from there which gpu it is using. Thats how I run it.goresnet said: ↑OK, need some more help, this time with Kombuster. I've chosen to run the DX11 one from the start menu, but it still only shows the 6620. How can I get it to stress the 6750? It has the icon for 3 gpu's but only tags gpu 2 & 3. It labels gpu 2 as the 6620, and gpu 3 has the 6620's clock speeds. All it says about gpu 1 is that it is a 6000 series. When you toggle the graphs, anly 2 & 3 show up. What's the deal? It does the same whether I run it from Afterburner, or choose any option from the start menu. Help!
edit: Also, which option do you choose in the Kombuster window? Gpu burn-in, Kmark normal or extreme, or something else entirely?Click to expand... -
OK, I like that tool (AMD System Monitor). I can still only seem to get Kombuster to run the 6620. Launching the DX11 one and with ULPS disabled per HT's registry app. I guess I'll have to prime it. Do you think Civ V will do it? I guess I should run it (Civ V), then alt-tab out and run Kombuster? Or does it stay active for a bit, so I could exit Civ V then go straight to Kombuster? Also, do you run at the default 1280x720, no AA?hypertek said: ↑You need to use gpu burn in, run it windowed, and than run amd system monitor and you will tell from there which gpu it is using. Thats how I run it.Click to expand...
And thanks everyone! This laptop (6135dx) is a BEAST! -
Great job! Very clean man! I like the grill used too.hypertek said: ↑
grill provided by a mini desk sorter from walmart for $2.88
something like this but mine is a different type (cant find the exact one on their website) Walmart.com: Rolodex Mini Sorter, 3 Sections: OfficeClick to expand...
To make good use of the openings, requires a modified cooler.
I just modded mine from a 5v dual slim 60mm (2300rpm) fans USB powered cooler to a AC Adapter 12v dual 80mm (1800rpm) LED fans. Night and Day difference in temperature rate drop.
My grill is raised out of the hole. Appeals to me, might not be to some. Using expanded metal. I'll be painting it black if I decide to keep it.
-
Just use GPU-z. No need to load up a game or anything.goresnet said: ↑Not sure about disabling ULPS, but I suppose I could give that a try. Kombuster is set to High Performance. What do you recommend to prime it with? So far the only game I have on here after the new install is Civilization I! Used to have Civ V, but will have to reinstall that.Click to expand...
If I have time, I'll try to make a video of the different throttles because I can reproduce it fairly easily (at least I used to be able to). With throttling, all four cores will drop from 2.8GHz down to like 1.3GHz (or whatever P5 or P4 state is) for a second or two until temps drop then it will spike right back to 2.8GHz.dreamonic said: ↑Then if all P-states are the same, say, 2.8Ghz. How exactly does it throttle down? Are you saying it overrides the user made clock changes to shipped clocks.. "throttling" inherit within the CPU itself? I don't think so. Mine still shows steady in Prime95. I'm using FusionTweaker for the clocks. I noticed when K10Stat was in control of it, it was all over the place.Click to expand...
You can control the clocks with K10Stat but you have to set it to ganged to get them to move together. Unganged will move independently. If you tell set it to no control the voltage still works, but then Windows takes control.
Like I said, I don't think it's necessarily due to absolute temperature, but rate of temperature increase. I can get mine to stay rock solid all the way to 95C if I cool it right and it increases progressively slow (not sure of defined interval). But then it WILL throttle beyond that.Mine will not do this. Given core load, given P-state clock. Higher P-states, mean higher core loads. Higher Temperature, means all of the above. Mine does not throttle down on any of these tests... because they are at 100% core load, so it's using the highest frequency I've set. It'll reach it's max temp if I purposely put it into that scenario and then will it BSOD as it cannot auto-throttle down any P-states if they are all the same for example. I only see it throttle down P-states on core load percent, not temperature itself, which coexist with higher (lower number) P-state values. If it was temperature, then why can I run 3Ghz at 90 degrees and not throttle down due to high temperature? It's because the cores are set to 100%!
Even if I have a lower P-state it will still stay at it's max P-state for 100% core load, temperature doesn't matter until....
It's not within junction temp - which if reached, BSOD or auto-throttle (depending on DEFINED P-states values only.)Click to expand...
What you are saying is correct. But you can control that in K10stat with the Up and Down parameters, and with unganged (independent cores), ganged on highest load core, ganged to lowest load core. I leave the up/down alone though, after messing with them default seems to work for me. K10stat and FusionTweaker do the exact same thing. Only FusionTweaker allows for the (forget the term) jump to P0 state without going to all states in between.Now I know I'm missing the latency, field, timer, instructions, macro instructions, (ms) time for each request to be used on each core and how they can work with each separately based on the P-state they are in, for how long, when they got there, where the other cores are instructed to next, changing given values to another request and so forth. But based on load factors, which pertain to load amount, request time, field size and instruction, it'll either reach P1 sooner and have to wait till it can lower itself to P2. This all depends on the handles each core has too. That is why if you load a browser in P7 state, you get P1 state because a given core(s) were able to accept a new request given the timed amount it was in idle mode in that state. Then as you go beyond a certain P-state, it becomes core load based.
No?
Not trying to come across sarcastic, just trying to understand what we both see here.Click to expand...
Maybe we're saying the same things, just differently.
-
hypertek said: ↑grill provided by a mini desk sorter from walmart for $2.88
something like this but mine is a different type (cant find the exact one on their website) Walmart.com: Rolodex Mini Sorter, 3 Sections: OfficeClick to expand...Great work guys. Better than my hack and slash job I did, lol. I had limited time to do it, so I just dremeled away with tools I had, should have bought the proper bits and got a hole saw. Oh well, end result is the same.dreamonic said: ↑Great job! Very clean man! I like the grill used too.
To make good use of the openings, requires a modified cooler.
I just modded mine from a 5v dual slim 60mm (2300rpm) fans USB powered cooler to a AC Adapter 12v dual 80mm (1800rpm) LED fans. Night and Day difference in temperature rate drop.
My grill is raised out of the hole. Appeals to me, might not be to some. Using expanded metal. I'll be painting it black if I decide to keep it.Click to expand...
One thing though that I did that helped quite a bit though is adding the stick on heatsinks. Requires sanding of the heatsink over the CPU and GPU to get the black paint off, and need to trim the heatsinks down to fit under the bottom panel, but dropped temps a solid 5C at load. If you haven't seen: http://forum.notebookreview.com/hp-...z-cooling-mod-thread-warning-lots-photos.html -
HTWingNut said: ↑Great work guys. Better than my hack and slash job I did, lol. I had limited time to do it, so I just dremeled away with tools I had, should have bought the proper bits and got a hole saw. Oh well, end result is the same.
One thing though that I did that helped quite a bit though is adding the stick on heatsinks. Requires sanding of the heatsink over the CPU and GPU to get the black paint off, and need to trim the heatsinks down to fit under the bottom panel, but dropped temps a solid 5C at load. If you haven't seen: http://forum.notebookreview.com/hp-...z-cooling-mod-thread-warning-lots-photos.htmlClick to expand...
That is my next mod.
I was thinking of doing what you did but with a smaller GPU heatsink (instead of vram sinks.) As you can see in hypertek's pics, the GPU and CPU heat pipe is exposed through the fan grill mods in PIC 1. If you were to fill that space under the grill with the same sized heatsink (or cut to fit) and solder them to the heatpipes, you could theoretically reduce temps by a huge margin if a cooler is blowing at it from there too. Because the bigger the heatsink and fins, the more heat is drawn away from the chips and less heat traveling down the rest of the heatpipes as well. Hmmm, sounds like more overclock potential..
Using something like this just for the sink and cut it to size and profile. This is the Akasa Cool Blue ChipSet Cooler and the sink measures: 38 x 38 x 13mm, which will require 1 width cut only. $14 for 2 of these. I may just do it.
That would make at least a 10 degree drop.
Definitely need vram sinks too. I'd imagine those suckers are cooking at 950.
Anyone here ever try the fan speed mod? You cut into the white wire and attach one side to a switch terminal and the other half of the wire on the other side of the switch terminal which is sent back to the fan. When the switch is toggled, it will break the connection and the fan runs 100%. -
I think the throttling is related to the TDP of the processors that we have, for example, a a8-3500 will throttle more easily than a a6-3410mx or a a8-3530mx because the former has a rated TDP of 35W.
When I was working with the 3500 I noted that even the voltage range was "limited", I mean, more limited than mine. My max voltage available is 1.3250 vs the 1.2750 of the 3500.
I don't have throttle, in fact my actual temps are not 89° as my signature says, that's the max temp of my processor because when the 4th fan speed kicks in it reduces the temp around 83°, which is far more than acceptable. Nowadays in games I don't even hear the 4th speed but instead that the 3th and the 2nd speed are in constant change.
Also, I don't know why but even using kombustor my temps for the GPU at 770 and 970 doesn't reach more than 72° with stock cooling, no need to improve cooling in there for some reason, please note that my xfire does actually works, except for battlefield 3 at 1080p but other than that I don't have problems or extreme stuttering (I have it but is barely visible). 12.1a drivers.
Also, something curious is that prime 95 didn't made mi computer crash after 30 min at 2.6 GHz when I removed the old mod an just put an small heatsink... but Skyrim did since that game is an extremely CPU bounded one, since then I both test Prime 95 and Skyrim to test every frequency, no more crashes after that.
P.D. Nice mods in there!
-
I think people here are confusing TDP with their CoolSense setting or forgot it was on coolest mode. While control on the CPU stepping presets or unless defined, make sure you are not running on "coolest mode" This WILL auto-throttle down to reduce temperatures. I believe some people here are confusing this. If you run on "performance optimized mode" no throttle will be done to the cores under CoolSense. If you do not have CoolSense installed, coolest mode is selected by default under control of the bios.
I'm over my TDP of 45 to nearly 80 Watts. If it was about thermal cooling ability to core throttling than my chip should be up and down all over the place at my Overclock. If you run on Performance Optimized Mode in CoolSense, at a high Overclock, no throttling is done when core load is high. Temperatures will continue to increase until it reaches junction temperature and BSOD or built into the chip to auto-throttle down at the lowest P-state. Also account for temperature rise to fan speed states. This is all relevant to what you define in K10Stat, FusionTweaker, Windows Power Management, CoolSense.. etc
As most should know, the temperature sensor for the GPU that is following the temperature of the CPU is the iGPU temp sensor (active- HWMonitor), not dGPU. So yes, the dGPU does run a lot cooler, but it's better to make it run cooler than for it to slowly increase with case/room temperature, that's why a heatsink will be better than an additional fan mod there.
What is your avg. FPS in games you play and with what settings? -
In all games I have vsync disabled, for skyrim I deactivate the multisampling and the antialiasing because I use a FXAA injector, everything on high except for the distance drawing which is in low, for battlefield everything was in low but I got very nice fps, deus ex I just activated the fxaa but deactivated everything else (I didn't have the mod back then)dreamonic said: ↑Not in all cases, especially when you are beyond the TDP (which most are.) I'm almost double mine, nearly 80watts. You are right about the voltage and this is probable to TDP & Cooling. See, it won't throttle down for me because it still thinks it's below initial TDP since the temperatures are within normal operating specs. of 35/45watts TDP's - even with increase of voltages and clocks. See how that works? All based on the CPU stepping presets, unless defined.
As most should know, the temperature sensor for the GPU that is following the temperature of the CPU is the iGPU temp sensor (active- HWMonitor), not dGPU. So yes, the dGPU does run a lot cooler, but it's better to make it run cooler than for it to slowly increase with case/room temperature, that's why a heatsink will be better than an additional fan mod there.
What is your avg. FPS in games you play and with what settings?Click to expand...
All my fps have been above 30, with some spikes of 45. -
dreamonic said: ↑I think people here are confusing TDP with their CoolSense setting or forgot it was on coolest mode. While control on the CPU stepping presets or unless defined, make sure you are not running on "coolest mode" This WILL auto-throttle down to reduce temperatures. I believe some people here are confusing this. If you run on "performance optimized mode" no throttle will be done to the cores under CoolSense. If you do not have CoolSense installed, coolest mode is selected by default under control of the bios.
I'm over my TDP of 45 to nearly 80 Watts. If it was about thermal cooling ability to core throttling than my chip should be up and down all over the place at my Overclock. If you run on Performance Optimized Mode in CoolSense, at a high Overclock, no throttling is done when core load is high. Temperatures will continue to increase until it reaches junction temperature and BSOD or built into the chip to auto-throttle down at the lowest P-state. Also account for temperature rise to fan speed states. This is all relevant to what you define in K10Stat, FusionTweaker, Windows Power Management, CoolSense.. etc
As most should know, the temperature sensor for the GPU that is following the temperature of the CPU is the iGPU temp sensor (active- HWMonitor), not dGPU. So yes, the dGPU does run a lot cooler, but it's better to make it run cooler than for it to slowly increase with case/room temperature, that's why a heatsink will be better than an additional fan mod there.Click to expand...
Those are good numbers. I found in DE:HR that if v-sync is on and triple buffering is enabled, it is much smoother -35+ feels like 50. -
Hmm, really? I should probably reinstall coolsense then...dreamonic said: ↑If you do not have CoolSense installed, coolest mode is selected by default under control of the bios.Click to expand...
-
What I meant with the tdp is perhaps because the processor expects a lower max temperature, for me throttling always starts at 95° but perhaps that's the "max" allowed for 45W TDP, for 35W may be much lower...dreamonic said: ↑Those are good numbers. I found in DE:HR that if v-sync is on and triple buffering is enabled, it is much smoother -35+ feels like 50.Click to expand...
-
...I put in an order for AMD stock. Why?
1. Although I initially thought the purchase of ATI by AMD was a huge mistake, I see now that it was a long-term strategy. AMD didn't stop competing with Intel, they just took a different direction in an already-proven tech: parallel computing. (Up until this point, serial has dominated because it's smaller and cheaper...not faster.)
2. AMD's llano/sabine line has been extraordinary popular, completely selling out AMD's limited stock. Strangely enough, this affected their already low stock price very little.
3. Referencing #2, part of the reason is that AMD has "gone green" in more ways than one: limited and realistic production runs, and reduced power consumption designs. (The latter has been proven by your abilities to undervolt successfully at a level that makes a statistical difference.) AMD has also restructured their company, trimming down considerably where Intel has not done so much.
4. Trinity promises (and, according to real-world testing, delivers) about a 30% increase in performance. At stock, that's about what you all are coaxing out of sabine now. I can only imagine that the approximate OC potential is about the same with Trinity, percentage-wise.
5. Unlike sabine, Trinity will be built on Bulldozer cores. Given that MS has recently acknowledged a flaw in Windows 7 that causes a lack of performance with BD, it promises to have a patch ready for Windows 7, and that this flaw won't exist in Windows 8 at all. This should put BD cores on a par with Intel performance.
6. Combining #5 with AMD's superior iGPU performance, I would expect AMD to stay far ahead of Intel's relative performance. Factor in cost, and AMD should kick .
7. Notebooks are now on a par with desktops in cost and performance. Trends over the last 5 years have shown notebooks outselling desktops, and especially when bargains exist...like the complete sellout of a "flawed" 6135dx stock originally from America's #1 computer retailer. (Which means, there were LOTS of these.)
8. The economy is pushing people into the "buy what you need that works for you" niche, and not the latest, greatest, and most expensive.
9. Relative ease of upgrading notebooks, as compared to only a few years ago.
10. Gaming dollars pushing development more than any other single reason.
11. Nvidia's relative "orphaning" by Intel and AMD. Just about the only cure for this would be for Intel to buy Nvidia, and that's not looking too well.
12. YOUR ENTHUSIASM over a product that was obvioulsy experimental in design product, using a surplus of cheaper old-gen cores and gpus.
13. Critical reviews that all agreed AMD was the way to go for the majority of users: "Buy the computer that best suits youor purpose for the least amount of money."
All indicators seem to point to AMD having a banner year in 2012. -
Many people said that AMD purchasing ATi was an error, and now we have Fusion chips all the way, extremely powerful, energy efficient and overclockable... in laptops. Back then I remember that overclock was exclusive to desktop and to very expensive laptops, our laptops, the very ones that we have been discussing here finally break that allowing the medium user to really get the most juice for the money.clarkkent57 said: ↑...I put in an order for AMD stock. Why?
1. Although I initially thought the purchase of ATI by AMD was a huge mistake, I see now that it was a long-term strategy. AMD didn't stop competing with Intel, they just took a different direction in an already-proven tech: parallel computing. (Up until this point, serial has dominated because it's smaller and cheaper...not faster.)
2. AMD's llano/sabine line has been extraordinary popular, completely selling out AMD's limited stock. Strangely enough, this affected their already low stock price very little.
3. Referencing #2, part of the reason is that AMD has "gone green" in more ways than one: limited and realistic production runs, and reduced power consumption designs. (The latter has been proven by your abilities to undervolt successfully at a level that makes a statistical difference.) AMD has also restructured their company, trimming down considerably where Intel has not done so much.
4. Trinity promises (and, according to real-world testing, delivers) about a 30% increase in performance. At stock, that's about what you all are coaxing out of sabine now. I can only imagine that the approximate OC potential is about the same with Trinity, percentage-wise.
5. Unlike sabine, Trinity will be built on Bulldozer cores. Given that MS has recently acknowledged a flaw in Windows 7 that causes a lack of performance with BD, it promises to have a patch ready for Windows 7, and that this flaw won't exist in Windows 8 at all. This should put BD cores on a par with Intel performance.
6. Combining #5 with AMD's superior iGPU performance, I would expect AMD to stay far ahead of Intel's relative performance. Factor in cost, and AMD should kick .
7. Notebooks are now on a par with desktops in cost and performance. Trends over the last 5 years have shown notebooks outselling desktops, and especially when bargains exist...like the complete sellout of a "flawed" 6135dx stock originally from America's #1 computer retailer. (Which means, there were LOTS of these.)
8. The economy is pushing people into the "buy what you need that works for you" niche, and not the latest, greatest, and most expensive.
9. Relative ease of upgrading notebooks, as compared to only a few years ago.
10. Gaming dollars pushing development more than any other single reason.
11. Nvidia's relative "orphaning" by Intel and AMD. Just about the only cure for this would be for Intel to buy Nvidia, and that's not looking too well.
12. YOUR ENTHUSIASM over a product that was obvioulsy experimental in design product, using a surplus of cheaper old-gen cores and gpus.
13. Critical reviews that all agreed AMD was the way to go for the majority of users: "Buy the computer that best suits youor purpose for the least amount of money."
All indicators seem to point to AMD having a banner year in 2012.Click to expand...
Even xfire! and that's something to say, even if it doesn't work on battlefield 3 I cannot complain about other games, is just simply wonderful to have 2 gpus rather than 1.
By the way, how much is amd stock right now? -
well, If AMD can persuade people that a i7 wont help you check email faster, then they probably will have a good year .
-
5.43...cheap when you consider Intel is 25.22, which also is an influencing factor...you can buy 5 times more shares of AMD for the same amount of money than Intel.
-
Which they can do by price alone.baiii said: ↑well, If AMD can persuade people that a i7 wont help you check email faster, then they probably will have a good year .Click to expand...
-
There have been Trinity benchmarks? Where?
*HP dv6z AMD Llano (6XXX series) Owners Lounge*
Discussion in 'HP' started by scy1192, Jun 22, 2011.
