Significantly more performant than the R9 290X using half the wattage. After looking at the R9 Nano kinda do want to ask the question, couldn't they make a cutdown version of it and place it in a laptop.
I am very curious as to the cost and actual performance of the R9 Nano.
-
-
Tbh I think the people that don't have money burning holes in their pockets and aren't in dire need of a computer should wait 1 more year for 2016 for both Pascal and R9 400 series. Simply because of the huge die shrink of 28-16/14 nm nodes, which I'm willing to bet will have massive accompanying performance upgrades. Not to mention HBM2 etc. I think that will be the prime time to upgrade and get your moneys worth. Seems like we've been in 28nm for ages..
ajc9988 likes this. -
Great to hear about the Nano TDP! AMD could obliterate NVIDIA in notebooks with their HBM, with it being so energy efficient and with 512GB/sec bandwidth. Now all they need to do is release it.
Sent from my Nexus 5 using TapatalkLast edited: Jun 17, 2015ajc9988 likes this. -
-
-
Cost isn't the biggest issue, heat is. At 125W the M295X is hard to cool even in an 18.4" Alienware. Remember the furnace that was 880M? Yeah that was a 120W card, and the M295X has 5W extra TDP.
-
-
Last edited: Jun 19, 2015
-
I'm thinking Apple overclocked their R9 m295x too much (+127 MHz on Core, + 112 MHz on Memory). Combined with Apple's poor cooling solution to let thermal throttling happen, then end result is... horrific on their end.
Maybe the R9 m390x will do better, but that has yet to be seen. -
enjoy!!!
-
Keep at it. It's just like the lack of knowledge regarding TDP regulation on haswell CPUs, that drive people crazy thinking many different things.
ajc9988 likes this. -
I'll be honest, for the longest time I believed the eye can "only" register 30FPS myth having been proved wrong. Personally I've found that my sensitivity between 30FPS and 60FPS isn't that bad. I can play at 30FPS and not rage about how it's not 60FPS. It really has to drop down into the teens before I start to get annoyed.
-
Now that I am gaming at 96hz, even 60fps seems like the bare minimum now haha. This doesn't mean I can't game 30fps. Some games just don't run well enough, specially on consoles and I still enjoy them. But if possible, I would always prefer higher fps, regardless of the game. -
-
I play them for exclusives and because it doesn't hurt to have a different system to relax and game on. But one of the compromises you get with a consoles, is that you can't determine your performance on your own, so you will have to make do with whatever is given to you. And lets face it, even with old consoles, we could finish games that often ran at 10 fps.... -
Anything above 25 FPS for instance seems to be fine with me (28 fps is basically the same like 30).
You might be able to notice some more 'smoothness' to fast pacing scenes at 60 fps (compared to 30), but to me, its hardly noticeable (actually, I don't think I can detect it to begin with).
Certainly not anything to get upset about or having a ridiculous demand that anything below 60 fps or even 35 fps is just 'inadequate' for playing. -
When it comes to fps in games there is no magic number above which you will get the best possible experience. There are games which will run just fine at lower fps while others will feel worse even though the average fps is higher. Generally minimum fps matters much more than the averages but this is still different from game to game. Some game may randomly drop fps from time to time which can be a bit annoying, another may run so consistent at 28 fps it may even feel smooth just because it doesn't fall below 26 and doesn't go above 30, and some games you may get average 45 fps but it will be so terribly optimized the frame rate would just drop to 15 and jump to 90 and end up a stuttery mess. It's obvious that while none of the three cases above are optimal, the third case is the worst by far while the second is arguably the best. Also a locked 30 fps may look smooth in some cases and not very smooth in others. For example fast paced games like Racing sims will feel smoother the higher the fps is and competitive shooters will benefit too because you turn the camera so much while RPGs and most MMOs are generally fine at lower fps if you can get consistent framerate, not all of them though, Action MMOs like Vindictus still need high framerates.
-
http://www.forbes.com/sites/jasonev...-new-fiji-graphics-card-beats-nvidias-980-ti/
An article showing AMD released benchmarks on game performance with settings. Granted reviews are two days out, but decided to share.
Also, got banned for trolling in Anandtech's forum for supposedly trolling trying to explain how the increased bandwidth allows for faster in and out of data from the HBM frame buffer allowing 4GB of HBM to function as well as 6-8GB of GDDR5 because they put the same amount of data through in the same amount of time... Pretty hilarious being called a troll for dispelling misinformation...Mr Najsman, TomJGX and triturbo like this. -
Surely that would mean that flagship GPUs would require less memory than low end GPUs? Because they have the faster memory. But that's not the case.
What am I missing?
Sent from my Nexus 5 using Tapatalk -
So how does this translate to HBM being able to perform the same function using less ram? Well, we have to look at what the industry has done to this point. As cards where being developed, the processing power required more bandwidth between the ram and the GPU. So, how can you increase that bandwidth if you are stuck with 32-bit bus width? You run them in parallel and slap more chips on the card!!! That's the solution, but has the cost of heat. It sped things up while costing less than developing a better solution. Why bandwidth? BECAUSE THAT WAS THE BOTTLENECK!!! After that solution was implemented, coders got lazy because they had so much ram to play with, there was no need to optimize the frame buffer.
But that still doesn't explain why HBM can perform the same function on less ram... But it actually does. By increasing the bandwidth by about 52% over the Titan X (336.5GBps vs 512GBps) along with caching improvements and reduced latency and a couple other tweaks, the data is transferred in and out of the frame buffer in HBM FASTER than GDDR5!!! So even if filled to capacity, it is functionally as if you had MORE ram if it was on the slower standard of DDR5! This also has to do with improvements to the ROP (which has only 64, far fewer than Titan X and 980 Ti) and an additional 800 shader cores to process the data quicker. Does this mean that it doesn't have limitations? NO! At 4K the performance is right on par. When 5K and 8K are used, due to larger frame buffer sizes, the speed advantage breaks down and capacity, once again, is king. It still MAJORLY ON SYNTHETIC BENCHMARKS outperforms the 980Ti, but gets owned by the Titan X! Why? Because now the larger capacity becomes necessary because of the resolution requiring larger size frames. This is where the 12GB of GDDR5 is able to buffer better and make the card shine. The 4GB on each standard IS NOT EQUAL and 4GB OF HBM HAS THE CAPABILITY OF PUTTING THROUGH AS MUCH DATA AS 6-8GB OF GDDR5 IN THE SAME AMOUNT OF TIME ALLOWING FOR EQUIVALENT PERFORMANCE! Even this is an oversimplification of all factors that take place.
Please excuse part of the above explanation, copied from what I posted in Anandtech...
As to why previous jumps in standards didn't yield the same results, it didn't have the bandwidth increase to the same degree, latency usually went up, not down, etc. It is a combination of the increased processing with the HBM that gives the same results on 4K. It is possible that 4GB is limiting the potential of the card to a degree, which is why I also had said:
"Now, as for the reasons for only 4GB instead of 8GB cards. It is a limitation of current interposer tech. Currently, for Gen1, interposers only allow a stack of 1GB of memory to rest on it. To do more, you must use a dual interposer which adds significant costs to production. It is not a lack of memory available, it is that the costs don't justify the return for the initial release!!! As this will only be needed for competing with the Titan X at 5K and 8K gaming, it can wait for the time period of releasing a dual Fiji card this fall without sacrificing sales."
So it is a mix of things to allow the utilization of the faster bandwidth, keeping frames in the frame buffer shorter time (but where it is, overall, giving the same amount of data put through the ram in the same amount of time), better caching, lower latency with much more powerful processing that makes it equal. That doesn't mean that a little more ram would not give any boost, but the gains may be minimal (compare the 980 Ti with 6GB vs Titan X with 12,while factoring in fewer ROPs and slightly fewer shader cores (300-400) giving approximately the same performance). In addition, there is more to be said on driver optimization and AMDs standard 3+ month optimization window that they have greatly increasing performance. As time moves on and the new APIs are introduced, the performance will grow...
If I made a mistake in here, please correct me... -
Thanks for that explanation.
So I take that it's due to reduced latency that HBM can do more with less? All GDDR5 modules have similar latencies so that's why the same doesn't apply there? Right?
Sent from my Nexus 5 using Tapatalk -
ajc9988 likes this.
-
-
Do you mean bandwidth? Because the bandwidth gap between low end GDDR5 GPUs and high end GDDR5 GPUs is greater than the difference between the top GDDR5 GPUs and the AMD Fury X's HBM 1 implementation?
Sent from my Nexus 5 using Tapatalk -
So my logic to get 4x4 gb ram is all in this too I knew i was not crazy in someways splitting data on smaller part with the highest bandwith is what Hbm is all about
not 10 chip with low bandwith
2-4 chip with very high bandwith and no latency -
bonami2 likes this.
-
The only limitation why it wont fit in MXM format is that the chip wont fit between the mounting holes, or it will be A REALLY tight fit (at 45º angle). Most likely wont fit. Take a look for yourself - http://www.bit-tech.net/hardware/graphics/2015/06/24/amd-radeon-r9-fury-x-review/2
EDIT: Just measured my Quadro 5000m's chip (the largest chip ever fitted to an MXM board (same as 480m), as far as I know) - it is 43x43mm and is quite literary over the clear zones of the mounting holes! If you rotate the chip 15~20º you could fit 45x45mm chip, but that's about it.Last edited: Jun 26, 2015 -
Edit: reason I ask is you could just move the holes. It would not be the first time an mxm heatsink was separately designed dependent upon whether you bought amd or Nvidia.... -
The hole spacing is there by standard, the design of the heat-sink does vary from GPU to GPU, but the hole spacing is the same. Don't know how much retooling would be involved if they decide to redesign the MXM module. The connector should be the same, I see no reason to change it. OK, maybe they can drop the LVDS, since most of the new displays are eDP anyway and replace it with two more DPs. That would be sad for those with older machines (me). Time will tell how it would play-out.
-
MXM is a loose standard... So they'd need new X brackets. That's really not a big deal. My concern is moving the heat fast enough not to make the HBM artifact and eventually fail. The common heat pipe design is just not going to work, its going to need active cooling old school fan sitting on heatsink Alienware style.
-
Edit: by modification I meant design mod to run the pipes directly over the vram and vrm on a card like that. -
I would love to be proven wrong but I'm tired of putting faith in AMD and so are the ODMs. -
Welp here I go
No, it doesn't. I'll elaborate a bit further soon.
This line in bold is incorrect. The 336.5GB/s is the mathematical bandwidth of the Titan X. I.E. a Titan Black would have this. But as we know, Maxwell has memory bandwidth optimizations that even showed up in that cuda benchmark around the 970 vRAM fiasco. I'd put the Titan X's bandwidth closer to 400GB/s; maybe somewhere around 380GB/s at MINIMUM. Assuming 380GB/s to be the effective bandwidth, this means only a ~35% improvement is gained over the Titan X/980Ti (132GB/s) at minimum. This also means that the Titan X/980Ti are 75% as fast as HBM on the Fury X. It's a large jump, sure, but it isn't a massive, end-all one allowing vRAM size to be stretched indefinitely. If you were comparing it to the 4GB on the 980, with a ~257GB/s effective bandwidth (224GB/s mathematical one) versus the 512, I would have wholeheartedly said that it would have allowed that 4GB to stretch further. But against something with even higher vRAM limits? It isn't happening. Even if the cache improvements are great and the latency is reduced.
I think this has less of an impact than you imagine. This would fall flat under a card's ability to render what's on-screen, and does not automagically mean that more cores would equal better performance with this. As far as I can see, the workload processing ability of the card as a whole is what matters. AMD happens to have a massive number of cores to get a certain processing power. One maxwell core is stronger than one Fiji core at the same clockspeeds. If someone wants to counter my statement on this, please do so.
There is no way that 4GB on a Fury X can put through 6-8GB of data that a 980Ti would in the same amount of time, going by bandwidth. And there is something else that is an issue: not all data is cached and swapped out. For example, a friend of mine with a 970 was explaining to me that while he sees his vRAM hitting the limits often, it does not trigger the 970 problem in every game. This is in line with behaviour that some games work just as fine with cards that have more vRAM as with cards that have less, except for the fact that the cards that have more vRAM see much heavier vRAM utilization. It's as if they just toss everything into cache for no reason. These kinds of games will not be hindered by the Fury X's 4GB limit. But the other kind; games that follow in the vein of Wolfenstein: The New Order where it cannot compress data? Those will falter. What happens with W:TNO is that HardOCP tried to force "ultra" textures on a 2GB GPU when it requires a 3GB GPU or better to turn them on. The game ran at a near-standstill and crashed with out-of-memory errors. This could happen for games that truly need (still dev unoptimization, but it could happen) over 4GB of vRAM at 1440p and higher resolutions.
Actually, we didn't have bandwidth improvements for quite a while. The GTX 280 and GTX 285 had huge amounts of memory bandwidth at the time (160GB/s on the 285; it had a 512-bit memory bus with GDDR3, which is EQUAL to a 256-bit memory bus on GDDR5 at the same memory clocks) and didn't actually get much faster until the GTX Titan/Titan Black/780Ti (192GB/s on GTX 580 and 680, 178GB/s on AMD 6970 etc). AMD actually sort of led the charge with the 7970; it having 288GB/s bandwidth, but its power would probably never be known because it took so long for its potential to shine with drivers. It's only lately that there's been a real big surge for them, with both AMD and nVidia trying to crack the 300GB/s barrier by default with their top tier cards, as well as large amounts of vRAM being a thing.
This is the real issue. nVidia is slacking lately, but we still get multiple drivers a month in most cases. AMD does not. Not even close, as far as I've seen. They need to polish things up here. I'm certain the Fury X can beat things. I've seen multiple benchmarks from multiple people using outdated drivers for nVidia, and when PCGamer produced their benchmarks with the absolute latest drivers (353.30 and whatever AMD's own is; it was a beta) the nVidia cards got significantly better scores in their game testing; something they even made note of in the tests as being a result of the recent drivers (with regards to the Titan X and 980Ti especially). So most synthetic benchmarks run by various websites including Tom's Hardware may need to be taken with a HUGE grain of salt. I decided I'm not accepting any benchmarks at all with drivers that are older (and perform worse than any existing newer driver) for nVidia cards, because then the results are skewed to AMD regardless of what the test produces, as the nVidia cards are not running as well as they could be at the time of the testing. And the same goes for the AMD cards, mind you. I want each card to be tested at the best they can be whenever the tests are done. But yes, 3 months down the line is going to be the best time. -
As for the pipes, they transfer the heat pretty well, not to say better than a fan blowing on top (take a look at all of the top-tier desktop air coolers). What is lacking is the "meat", so to say at the end of those pipes. You'll need a hell of a radiator (and fan) to dissipate the heat as fast as possible.
Ethrem likes this. -
-
I ust noticed the M395X @ notebookcheck and was amazed to see 2048 gcn shader cores in an amd mobile chip. When'd they launch this?
-
King of Interns Simply a laptop enthusiast
LOL sangemaru. More than 2 years ago haha! M395X is a rebrand of the M295X with a minor bump in clocks.
Hotter (or as hot) as the 980M with performance in the 970M ballpark but not as OCable due to the heat. Had it arrived alongside the 780M things might have been different... -
TomJGX likes this.
-
Meaker@Sager Company Representative
I assume this design of heatsink may be popular if the HBM is getting too hot:
TomJGX likes this.
AMD announces Radeon M300 series notebook video cards
Discussion in 'Gaming (Software and Graphics Cards)' started by octiceps, May 6, 2015.