The Notebook Review forums were hosted by TechTarget, who shut down them down on January 31, 2022. This static read-only archive was pulled by NBR forum users between January 20 and January 31, 2022, in an effort to make sure that the valuable technical information that had been posted on the forums is preserved. For current discussions, many NBR forum users moved over to NotebookTalk.net after the shutdown.
Problems? See this thread at archive.org.
← Previous page

    CPU isnt improving fast enough

    Discussion in 'Hardware Components and Aftermarket Upgrades' started by jsteng, Jan 26, 2010.

  1. IntelUser

    IntelUser Notebook Deity

    Reputations:
    364
    Messages:
    1,642
    Likes Received:
    75
    Trophy Points:
    66
    LOL, Daveperman indirectly citing superiority of his SSDs again(not saying you are wrong, just funny).

    Cache sizes ARE computation related on a CPU though. :)

    It's why GPUs and CPUs can NEVER replace each other. The workload of the two are fundamentally different. Core 2 increased performance by increasing things that weren't computation related like the ALUs, but memory like caches, memory disambiguation and prefetchers.

    Sandy Bridge will focus majorly on cache and memory functions and will be the next big improvement. Media enhancements like AVX are really the icing on the cake, not the main feature.
     
  2. davepermen

    davepermen Notebook Nobel Laureate

    Reputations:
    2,972
    Messages:
    7,788
    Likes Received:
    0
    Trophy Points:
    205
    no, gpus and cpus can never replace each other as long as they have a different instruction set. larrabee could theoretically replace a cpu completely (not being that efficient at it, though).

    and no, caches are not computational workload. they are memory-workload.

    if i want to sum two registers 10 billion times, all i can do is call add reg0, reg1 10 billion times. no cache or so is used for this, but a computational part of the cpu: the adder unit.

    and yes, i'm nitpicking. but i want accuracy from someone called intel user :)
     
  3. thinkpad knows best

    thinkpad knows best Notebook Deity

    Reputations:
    108
    Messages:
    1,140
    Likes Received:
    0
    Trophy Points:
    55
    Ahem, wasn't ARM already dominating that market?
     
  4. davepermen

    davepermen Notebook Nobel Laureate

    Reputations:
    2,972
    Messages:
    7,788
    Likes Received:
    0
    Trophy Points:
    205
    and? that's actually a reason FOR trying to build atom.

    arm isn't intel. atom is. they want a bit of the growing market.
     
  5. Pitabred

    Pitabred Linux geek con rat flail!

    Reputations:
    3,300
    Messages:
    7,115
    Likes Received:
    3
    Trophy Points:
    206
    A CPU can ALWAYS replace a GPU. That's the whole idea... a CPU is a completely general-purpose processing device. It can do ANYTHING. The reason we have GPUs is that for any given task, purpose-built hardware will always be faster than general-purpose. The GPU is designed to make certain operations very fast, and it will always be faster at those than the CPU. But that's at the expense of generality... it only does a few things.
     
  6. thinkpad knows best

    thinkpad knows best Notebook Deity

    Reputations:
    108
    Messages:
    1,140
    Likes Received:
    0
    Trophy Points:
    55
    I know ARM isn't Intel for blank sakes, but the Atom still puts out too much heat to be inside a SmartPhone, 11nm maybe it might be. The GPU is just a high performance processor that exclusively handles graphics oriented operations. I wonder when we'll start having 128-bit CPU's :p
     
  7. IntelUser

    IntelUser Notebook Deity

    Reputations:
    364
    Messages:
    1,642
    Likes Received:
    75
    Trophy Points:
    66
    That's why I said its related. True, it never really does computation. But in modern CPUs, adding an extra computation unit like an extra ALU does nothing. Those ALUs are limited by the memory ops, and better caches opens them up. So its sorta true, and sorta not. :)

    It won't in the sense that GPU will always perform better than CPU at graphics, but you already knew that.

    The Pineview/Diamondville/Silverthorne Atom in current Netbooks/UMPCs/MIDs might put too much heat. The Lincroft Atom coming in Moorestown Atom will be low enough power to fit in larger smartphones. The Z500 CPU running at 800MHz is already at only 0.6W.

    In a smartphone, CPU, memory controller, and the I/O controller aren't the only big contributors to power consumption. The display and the type of interface, operating systems, BIOSes and components on the motherboard also use hefty amount of power. If you can reduce that significantly you'll get that much closer to smartphones.

    Moorestown will focus on not just integrating the graphics and memory controller like pineview and be done with it. Rest of the components I have mentioned above will be significantly improved.

    First 2 Moorestown (smart)phones:
    http://www.youtube.com/watch?v=5m79buEJQQY
    http://www.youtube.com/watch?v=WfkzpdB97fg

    The LG device is rumored to have 5 hours browsing time on 3G with its 1850mAH battery. It is 50% larger than the one in the iPhone 3GS, but considering the bigger screen and much better performance, its no way that far off. 45nm and Moorestown is enough to make Atom relevant for smartphones, 32nm with Medfield will get power parity.
     
  8. davepermen

    davepermen Notebook Nobel Laureate

    Reputations:
    2,972
    Messages:
    7,788
    Likes Received:
    0
    Trophy Points:
    205
    hm no. give me another one, and my app runs faster.

    only the moment you have to wait for memory data (and no other hyperthread has something to do that it waited for), memory starts to be relevant.

    that is, by default, quite often, yes. but it doesn't change the fact that numerical power on it's own is 100% defined without any caches in mind. and can be used as those.

    but in most real world apps, one has to consider the data part as well, not just the computing part. and there, memory bandwith and latency gets important, of course.

    that's why gpu's by now are sort of suuuperhyperthreaded (they call it differently). they mostly have tons of jobs pending for their data, and still have another one in the actual core to compute while the others wait for the data. that way, they can hide the latencies in their main job (doing billions of times the very same again and again)
     
  9. Meaker@Sager

    Meaker@Sager Company Representative

    Reputations:
    9,431
    Messages:
    58,189
    Likes Received:
    17,900
    Trophy Points:
    931
    You HAVE to include more cores, it has several benefits:

    1. Larger contact area for heatsinks
    2. Having 1156 pins requires a certain area around a die to actually make the connections, thats the minimum die area you can have. If you only have one core you have wasted space anyway.

    Oh and if memory was such a bottleneck, why does increasing the clock speed/ decreasing the latencies have such a small effect on real world apps?
     
  10. davepermen

    davepermen Notebook Nobel Laureate

    Reputations:
    2,972
    Messages:
    7,788
    Likes Received:
    0
    Trophy Points:
    205
    it WAS a huge bottleneck for the first quadcores.

    determinable by being able to scale about linear from 1 to 2 to 3 cores, but the 4th core normally never gained the same. there, increasing the memory performance helped (by f.e. reducing the amount of needed memory to access, etc)

    but nowadays, not so much anymore. still obviously dependent on the workload.

    for big scenes one renders (as an example), programmer have to take much care to prepare the data in a way that all cores get feed with data. it's not "just working well". programmers have to manage memory manually quite a bit. so there is a bottleneck, which so far can be fixed by programmers.

    but that's work. work that could be spent better.
     
  11. IntelUser

    IntelUser Notebook Deity

    Reputations:
    364
    Messages:
    1,642
    Likes Received:
    75
    Trophy Points:
    66
    It's that or basically nothing. Double caches giving you 5-7% increase or lower latencies on Nehalem might have helped only 10%, but double your ALUs or your issue width and you gain 2-3%.

    BTW, Athlon 64 gained 20% by the lower latency memory controller alone, on PC apps, and way more on server apps.

    Pentium M gained 30-40% over Tualatin with no better FPU or more ALUs.
     
  12. Undertaxxx

    Undertaxxx Notebook Consultant

    Reputations:
    33
    Messages:
    264
    Likes Received:
    0
    Trophy Points:
    30
    Can't agree more
    +1
     
  13. thinkpad knows best

    thinkpad knows best Notebook Deity

    Reputations:
    108
    Messages:
    1,140
    Likes Received:
    0
    Trophy Points:
    55
    Yeah, they increase everything within the time period of lets say 10 years, miniscule power improvements with each release, (few exceptions) but then you look back and there's 40-50% performance gain with processors, it's also marketting influence. Most of us would be fine with advanced Core 2's for the next 5 years, then Intel would release the i series, improving performance in the double digit percentages, but they released it now because they wouldn't make as much money right now, instead of releasing the technology when it would have been truly a need among power users. It's sort of like GTA 4, Rockstar probably should have released the PC version in August 09, or even December 09, and pushed the console version up a little too.
     
← Previous page