The Notebook Review forums were hosted by TechTarget, who shut down them down on January 31, 2022. This static read-only archive was pulled by NBR forum users between January 20 and January 31, 2022, in an effort to make sure that the valuable technical information that had been posted on the forums is preserved. For current discussions, many NBR forum users moved over to NotebookTalk.net after the shutdown.
Problems? See this thread at archive.org.

    ADL is here! (Alder Lake 12th Gen)

    Discussion in 'Hardware Components and Aftermarket Upgrades' started by Tech Junky, Jan 4, 2022.

  1. Tech Junky

    Tech Junky Notebook Deity

    Reputations:
    341
    Messages:
    1,489
    Likes Received:
    601
    Trophy Points:
    131
    Looks like the quickest releases are Dell / Alienware / Acer / Razer / HP

    upload_2022-1-4_13-7-56.png

    Looks like Intel is using the KISS approach with 8 SKU's being released and they look well balanced. As the day of CES progresses there should be more OEM's pushing out new products for sale.

    https://www.theverge.com/2022/1/4/2...-ces-2022-h-series-laptops-gaming-performance

    In addition to the performance gains, the new 12th Gen chips also come with a variety of other improvements, including Wi-Fi 6E, Thunderbolt 4, PCIe 4.0, and support for the latest in memory standards (with support for DDR5 memory at up to 4,800MHz and LPDDR5 at up to 5,200MHz both offered). Notably missing, however, is integrated support for the latest in HDMI standards (Intel’s chip tops out at HDMI 2.0b) and PCIe 5.0 lanes (the latter of which Intel does offer on its desktop Alder Lake chips).
     
  2. saturnotaku

    saturnotaku Notebook Nobel Laureate

    Reputations:
    4,879
    Messages:
    8,926
    Likes Received:
    4,701
    Trophy Points:
    431
    I feel the same here as I do about AMD's announcements today - Meh.
     
    Papusan likes this.
  3. Tech Junky

    Tech Junky Notebook Deity

    Reputations:
    341
    Messages:
    1,489
    Likes Received:
    601
    Trophy Points:
    131
  4. saturnotaku

    saturnotaku Notebook Nobel Laureate

    Reputations:
    4,879
    Messages:
    8,926
    Likes Received:
    4,701
    Trophy Points:
    431
    LMAO @ Razer

    [​IMG]
     
    Papusan and etern4l like this.
  5. etern4l

    etern4l Notebook Virtuoso

    Reputations:
    2,931
    Messages:
    3,533
    Likes Received:
    3,492
    Trophy Points:
    331
    Haha. Bitcoin and qualifying organs accepted.
     
    Papusan likes this.
  6. Shark00n

    Shark00n Notebook Deity

    Reputations:
    110
    Messages:
    901
    Likes Received:
    414
    Trophy Points:
    76
    Are there TDP figures for the CPUs and 3070/3080Tis? On those Razers

    If you think those are expensive :cool: Just wait for MSI's new ADL line
     
  7. Tech Junky

    Tech Junky Notebook Deity

    Reputations:
    341
    Messages:
    1,489
    Likes Received:
    601
    Trophy Points:
    131
    I didn't dig into them as I'm more geared towards more of a barebone but beefy CPU setup myself. I know some of the hype around the desktop versions didn't really play out when it comes to consumption or heat. The server I rebuilt with a 12700K / air cooler sits under 80F most of the time and might get up to 130F under load. When I was planning it the thought came to mind of water cooling due to the hype surrounding the heat these things are supposed to be putting out but, I'm not seeing it. I don't have a GPU in it though to complicate the thermals with higher output of heat. When I was playing around with dual GPU's & 8700K in a setup though that pushed the temps up considerably to the point where LC would have been a benefit.

    As to the laptop side of things it's TBD to see what kind of numbers these new chips put out. There are some hints as to the power of the new Xe iGPU though indicating they support OpenCL means they could have ample power for mining as OCL is used for hashing. That makes them a bit more interesting to play with. Pairing them with a dGPU as most OEM's will do anyway means they should be able to blow even recent systems out of the water with significant boosts in performance.

    Of course this is all to be taken as a grain of salt until production units ship and get benched. I'm a bit disappointed they don't come with Gen 5 under the hood though Gen 4 works fine in the meantime since there's no Gen 5 devices out to put into use at the moment. It would just be nice to have both ADL systems with the same playing field.

    I did happen to spot something that looks appealing with ADL laptop supporting 10GE NIC's which would be a huge upgrade compared to having to spend ~$300 on a TB3 dongle to get that sort of speed on a laptop.
     
  8. Tech Junky

    Tech Junky Notebook Deity

    Reputations:
    341
    Messages:
    1,489
    Likes Received:
    601
    Trophy Points:
    131
  9. Shark00n

    Shark00n Notebook Deity

    Reputations:
    110
    Messages:
    901
    Likes Received:
    414
    Trophy Points:
    76
    AORUS seem to have really low TDPs... Like all below 130W. And thats max TGP, so, with Dynamic Boost. Real TGP should be around the 115W mark. And it seems like it's the same for 3080Ti/3070Ti.
    Or 3080 Thai like the nvidia reps like to say

    Makes me think manufacturers are up to something fishy...

    Plus I don't think we're gonna see a new Legion 7.
    Legion laptops leaker like a week or two ago and the 7 wasn't on the leaked models either.

    Well, see ya next year :cool:
     
    Last edited: Jan 6, 2022
  10. StormJumper

    StormJumper Notebook Virtuoso

    Reputations:
    579
    Messages:
    3,537
    Likes Received:
    488
    Trophy Points:
    151
    For me if someone wants a difference in CPU it has to be 2x the one your comparing to for it to make a substantial change. Otherwise your through good money.
     
  11. Shark00n

    Shark00n Notebook Deity

    Reputations:
    110
    Messages:
    901
    Likes Received:
    414
    Trophy Points:
    76
    I agree, but Moore's Law has been pretty much dead for a while tho.
    These are an awesome update if you're coming from 6th or 7th gen. Not that crazy if you're coming from 10th or 11th gen.
     
  12. Tech Junky

    Tech Junky Notebook Deity

    Reputations:
    341
    Messages:
    1,489
    Likes Received:
    601
    Trophy Points:
    131
    10/11 gen CPU's to me were more of a stopgap leading up to ADL. At least on the desktop side. On the laptop side there's a decent boost in tech but, yeah, you'd need to be 9th gen or earlier to realize

    2X what?

    On the desktop side there was the intro of a few more things such as DDR / PCI Gen 5 options. Laptop side is a little less significant as Intel didn't push Gen 5 anything from what I've seen so far. Sure they opened the door for DDR 4 / 5 options like they did on the desktop SKU's but probably the most noticeable might be the iGPU hitting 96EU's for someone not looking to spend on a RTX model.
     
  13. Casowen

    Casowen Notebook Evangelist

    Reputations:
    64
    Messages:
    399
    Likes Received:
    108
    Trophy Points:
    56
    I am just curious, but what is intel doing that actually boost gaming performance each generation? Is it just singled thread work? I rarely see my cpu utilization over 50% and thats with a 4 core. Or is the advertised performance boost mainly just a gimick?
     
  14. Tech Junky

    Tech Junky Notebook Deity

    Reputations:
    341
    Messages:
    1,489
    Likes Received:
    601
    Trophy Points:
    131
    Well, they're pushing the iGPU towards being more powerful. The recent Iris Xe chips they're embedding just jumped from 32 EU's to 96EU's putting them closer to older dGPU implementations. How the chips run a game depends more on how they're coded to use the HW than any specific part. It's a puzzle of pieces that come together to make your gaming experience smoother.

    Loading times are reduced with each generation but, the experience still mainly depends on which GPU they get paired with since the dGPU is what handles the heavy work when a game is open and running. Even though my current GTX1650 has 4GB of RAM it shares the 32GB of RAM I have installed for additional processing room as needed. The 2 GPU's share the system RAM @ ~16GB each total to provide faster processing of images.

    Due to the confined space of CPU's though they'll never be as powerful as the add on of a true GPU but, they're making more progress towards that sort of scenario. If they took the space consumes by both the CPU / GPU in a laptop and made a large layout combined chip there cold be a serious change in the performance. The BGA size of these right now are about the size of your thumb nail and fit into a space together about the size of a matchbook yet don't come close to filling it at about ~50%.

    The desktop size of the CPU is about the same size as a matchbook on its own. There are some laptops that use the desktop version of CPU in their builds so, there's room to make this happen if the designers wanted to develop something for use. Tying things together would open some doors to serious competition to AMD / NVIDIA laptop GPU's when it comes to performance.
     
    etern4l likes this.