The Notebook Review forums were hosted by TechTarget, who shut down them down on January 31, 2022. This static read-only archive was pulled by NBR forum users between January 20 and January 31, 2022, in an effort to make sure that the valuable technical information that had been posted on the forums is preserved. For current discussions, many NBR forum users moved over to NotebookTalk.net after the shutdown.
Problems? See this thread at archive.org.

    6 core laptop CPUs?

    Discussion in 'Hardware Components and Aftermarket Upgrades' started by cdoublejj, Sep 28, 2014.

  1. cdoublejj

    cdoublejj Notebook Deity

    Reputations:
    195
    Messages:
    1,123
    Likes Received:
    92
    Trophy Points:
    66
    do true blue 6 core laptops cpus exist? like a 6 core i7 mobile?
     
  2. octiceps

    octiceps Nimrod

    Reputations:
    3,147
    Messages:
    9,944
    Likes Received:
    4,194
    Trophy Points:
    431
    No but there are mobile workstations and servers with 6+ core desktop i7's and Xeons.
     
  3. Jarhead

    Jarhead 恋の♡アカサタナ

    Reputations:
    5,036
    Messages:
    12,168
    Likes Received:
    3,133
    Trophy Points:
    681
    Nope, unless Clevo has desktop replacement laptops with desktop 6-core CPUs.
     
  4. Maru

    Maru Notebook Consultant

    Reputations:
    136
    Messages:
    235
    Likes Received:
    89
    Trophy Points:
    41
    Not sure what you mean by 'true blue" but

    Intel i7 quad cores (MQ, HQ) have hyperthreading so they support two threads per core, or 8 logical threads.

    Samsung's Chromebook 2 uses its Exynos 5 octo-core ARM (in big.LITTLE, 4 big + 4 LITTLE core configuration)

    Both of those are 8 not 6, so not "true blue" in that sense.

    AMD counts "compute cores"
    3 core CPU + 2 core GPU = 5 compute cores
    4 core CPU + 4 core GPU = 8 compute cores
    4 core CPU + 6 core GPU = 10 compute cores
    etc.
    But if you don't care about the GPU threads then those may not be "true blue" either.
     
  5. King of Interns

    King of Interns Simply a laptop enthusiast

    Reputations:
    1,329
    Messages:
    5,418
    Likes Received:
    1,096
    Trophy Points:
    331
    True desktop replacement laptops like the D900F have featured full desktop CPU support for 6 core CPU's since before 2010. The D900F is that old and can support up to 990X and 980X for example.

    So yes 6 cores in a laptop exists but I don't think any true laptop CPU's have been released yet featuring more than 4 physical cores.
     
  6. Karamazovmm

    Karamazovmm Overthinking? Always!

    Reputations:
    2,365
    Messages:
    9,422
    Likes Received:
    200
    Trophy Points:
    231
    There are no laptops cpus that come with 6+ cores.

    there are however other notebooks that come with desktop cpus and those can have 6+ cores. Basically only clevo makes those
     
  7. n=1

    n=1 YEAH SCIENCE!

    Reputations:
    2,544
    Messages:
    4,346
    Likes Received:
    2,600
    Trophy Points:
    231
    Those extra cores don't add much for routine light activities (surfing, watching movies, office productivity). Where it'll really shine is for video encoding, games that scale beyond 4 cores (BF4 and Crysis 3 comes to mind) and other heavily multi-threaded tasks.

    My 4930K desktop does feel snappier but only because it's overclocked to 4.5GHz on all 6 cores as opposed to my 4900MQ that's running stock (3.6-3.8GHz). Drop it back to stock and it feels about the same.
     
  8. Krane

    Krane Notebook Prophet

    Reputations:
    706
    Messages:
    4,653
    Likes Received:
    108
    Trophy Points:
    131
    Or even non routine where only a single core is utilized. Something to keep in mind.
    In that case, the difference can be substantial. Where these application are designed to scale and take advantage of the additional cores.


    Yes, in its Eurocom application. And as you might expect, it is mobile, but just barely.
     
  9. n=1

    n=1 YEAH SCIENCE!

    Reputations:
    2,544
    Messages:
    4,346
    Likes Received:
    2,600
    Trophy Points:
    231
    Yeah I should've added core speed is more important than having 2 extra cores for most situations, unless you know you can utilize those extra cores.

    That said having both will guarantee maximum performance.
     
  10. Charles P. Jefferies

    Charles P. Jefferies Lead Moderator Super Moderator

    Reputations:
    22,339
    Messages:
    36,639
    Likes Received:
    5,075
    Trophy Points:
    931
    We reviewed some of those Clevo's with the six-core CPUs. Here's the most recent:
    AVADirect Clevo P570WM Review

    As noted they are mobile but not really - 2.5" thick, 13.8 lbs and a 53 minute battery life. And don't forget the humongous dual 330W power supplies (I believe they can operate off of one depending on the configuration; our review unit was likely supplied with two due to the SLI'd Nvidia graphics cards). The whole package ends up being about 20 pounds. I always wanted one despite how ridiculous they are. lol.
     
    alexhawker likes this.
  11. Meaker@Sager

    Meaker@Sager Company Representative

    Reputations:
    9,431
    Messages:
    58,189
    Likes Received:
    17,897
    Trophy Points:
    931
    They are fun ;) No power limits and you can do what you want with it.

    Plus by clamping the GPUs to P8 you can extend the battery life a fair amount.

    Definitely heavy though!
     
  12. cdoublejj

    cdoublejj Notebook Deity

    Reputations:
    195
    Messages:
    1,123
    Likes Received:
    92
    Trophy Points:
    66
    AMD Bulldozer based CPUs are halfcores half modules. Before the computer core crap. i won't even go there, their marketing sucks. Stream processors and CPU cores, clean and simple.
     
  13. Dufus

    Dufus .

    Reputations:
    1,194
    Messages:
    1,336
    Likes Received:
    548
    Trophy Points:
    131
    No true blue ones, only white ones. :p

    [​IMG]
     
  14. nipsen

    nipsen Notebook Ditty

    Reputations:
    694
    Messages:
    1,686
    Likes Received:
    131
    Trophy Points:
    81
    A just as relevant question is what benefit you really get from a "true" 8-core or higher setup on the normal industry standard bus interface and an x86 instruction set - on a desktop. Or wherever.

    The thing is that outside microcode reduction (that doesn't really exist between "cores" as of yet anyway, unless a shared store currently contains the result of a new request), more cores can only be used to effectively run independent contexts. Virtual servers, that sort of thing. Outside of that - parallelism on an x86 instruction set, or any instruction set that isn't based on an explicitly parallel assembly language, isn't going to benefit from increased number of cores.

    Instead, what you're actually running into if you try to run a few tests, is that the overhead from context shifts is going to be higher from maintaining many active threads (in spite of many "cores", never mind that the response times towards the bus become unpredictable), than from running a scheduler programmatically. It's not useless to trust in high-level optimisations and creating threads, but you can't expect more cores to increase calculation performance unless you're running independent contexts (and they don't draw too much common resources and bandwidth - such as memory and IO, bus width, etc.).

    And.. that's not really a surprise. Because that's how the instruction set is made to function. And that's what the platform and the current industry standard was meant to achieve.

    ...just making the point that there's not going to be any point whatsoever to get more cores, or to have microcores with their own instruction set acceleration perhaps. Or having OpenCL "scale" over gpu units and peripheral devices (basically the same thing as those microcores) -- because the response time through the current standard for buses isn't fast enough or broad enough.

    *shrug*
     
    Jarhead likes this.
  15. n=1

    n=1 YEAH SCIENCE!

    Reputations:
    2,544
    Messages:
    4,346
    Likes Received:
    2,600
    Trophy Points:
    231
    Just thought of something: how would a true 8 core with HT off compare to a quad core with HT on?
     
  16. tijo

    tijo Sacred Blame

    Reputations:
    7,588
    Messages:
    10,023
    Likes Received:
    1,077
    Trophy Points:
    581
    If you're doing something compute heavy, night and day, it could even go to twice as fast. HT is useless for some things while physical cores aren't and even if the best of cases, you can expect 20-25% more performance from HT.
     
  17. octiceps

    octiceps Nimrod

    Reputations:
    3,147
    Messages:
    9,944
    Likes Received:
    4,194
    Trophy Points:
    431
    You mean like a 5960X w/o HT vs. a 4770K/4790K w/HT? Because, as you know, the AMD chips are not "true" 8-core CPU's due to the module-based Bulldozer architecture.

    The 5960X would win for sure, but at a much higher TDP ofc. HT can't make up the massive gulf in core count and cache sizes.

    In the most optimized applications, Hyper-Threading can improve performance by almost 30%, which is definitely a net win considering the marginal increases in die area, heat, and power consumption it brings, hence its popularity on Intel chips for what has been a very long time now.
     
  18. tijo

    tijo Sacred Blame

    Reputations:
    7,588
    Messages:
    10,023
    Likes Received:
    1,077
    Trophy Points:
    581
    It's not just a matter of optimization as far as I understand it. When I was running finite elements calculations, hyper-threading did basically nothing, I'm expecting that it's due to the nature of the task, so thread don't stall (this is just speculation on my part, but what I've observed seems to go that way). Running the calculations on one to 4 cores did yield considerable improvements, however, forcing the program to run on more than 4 cores on an Intel quad core resulted in no perceivable gain. The only difference I observed were in the fluidity of the OS, it was still fluid as long as the 8 logical cores weren't loaded with the simulation.

    Video encoding on the other hand did perform noticeable better with hyperthreading.
     
  19. n=1

    n=1 YEAH SCIENCE!

    Reputations:
    2,544
    Messages:
    4,346
    Likes Received:
    2,600
    Trophy Points:
    231
    Yeah I meant 5960X vs something like 4770K.

    Actually I should probably clarify something. What got me thinking was if an application was not optimized for multithreading and to scale beyond 4 physical cores, how said application would perform with a true 8 physical core processor compared to one with 4 physical cores but 8 logical cores.
     
  20. octiceps

    octiceps Nimrod

    Reputations:
    3,147
    Messages:
    9,944
    Likes Received:
    4,194
    Trophy Points:
    431
    Sounds like the simulation doesn't schedule beyond 4 threads. Hyper-Threading won't help if the workload isn't optimized, so software is key like I said. Video editing, 3D rendering, extreme multitasking, and other heavily multi-threaded workloads will benefit the most from Hyper-Theading. HT is better than no HT but not as good as more physical cores. More physical cores in addition to HT would obviously be best.

    Interesting to note, 4 cores w/HT is still not as fast as 6 cores w/o HT, let alone 8.
     
  21. octiceps

    octiceps Nimrod

    Reputations:
    3,147
    Messages:
    9,944
    Likes Received:
    4,194
    Trophy Points:
    431
    It would perform the same.
     
  22. n=1

    n=1 YEAH SCIENCE!

    Reputations:
    2,544
    Messages:
    4,346
    Likes Received:
    2,600
    Trophy Points:
    231
    <del>It makes sense on paper, but I'm curious why that would still be the case for legacy applications that can't even multithread properly? (or would it still hold true more cores = better in that scenario?)</del>

    ninja'd

    and good to know I'm not going crazy lol
     
    octiceps likes this.
  23. octiceps

    octiceps Nimrod

    Reputations:
    3,147
    Messages:
    9,944
    Likes Received:
    4,194
    Trophy Points:
    431
    Single-threaded app = no performance benefit to more cores or HT, only more clock speed. :p
     
  24. tijo

    tijo Sacred Blame

    Reputations:
    7,588
    Messages:
    10,023
    Likes Received:
    1,077
    Trophy Points:
    581
    Actually, it does, I can even manually schedule it to whatever amount of threads I want (launch the program with -np=# of threads). I've run the same simulation on a 6-core CPU as well and the difference was there. What I meant was that on a quad core with HT, doesn't matter whether 4 or 8 threads are scheduled, the computation will take roughly the same amount of time.

    Eventually, the simulation wouldn't scale too well, but that will happen at cluster levels where you can schedule dozens of cores.
     
  25. octiceps

    octiceps Nimrod

    Reputations:
    3,147
    Messages:
    9,944
    Likes Received:
    4,194
    Trophy Points:
    431
    So it's one of those programs that responds favorably to more physical cores rather than more efficient thread scheduling (HT).
     
  26. tijo

    tijo Sacred Blame

    Reputations:
    7,588
    Messages:
    10,023
    Likes Received:
    1,077
    Trophy Points:
    581
    Yup, but as far as I know, it's generalized to the numerical method used rather than the particular code/program.

    EDIT: looking a little more into it, it seems to be the same for other numerical methods like finite volumes. To be honest, I've used finite elements before, but it's not really my domain, I'll ask friends who are doing their PhDs on CFD when I get the chance.

    Seems like HT may help in some cases, but it's not easy to determine since a lot of factors will come into play. If you go nuts with the # of cores, you'll end up with the communication between the CPUs being a bottleneck, then there's memory access, depending on how many memory channels you have, you could end up I/O bound with memory and so on.
     
  27. n=1

    n=1 YEAH SCIENCE!

    Reputations:
    2,544
    Messages:
    4,346
    Likes Received:
    2,600
    Trophy Points:
    231
    I feel like I should install Spartan and fire it up to see if it's able to take advantage of 6 cores.
     
  28. Dufus

    Dufus .

    Reputations:
    1,194
    Messages:
    1,336
    Likes Received:
    548
    Trophy Points:
    131
    More cores can help with scheduling so the single threaded app can get it's own core and not be thrown off by some other process and have to wait for it's next quantum while having it's L1/L2 cache possibly wrecked.

    Also remember that in some situations HTT can actually cause a degrade in performance.
     
  29. nipsen

    nipsen Notebook Ditty

    Reputations:
    694
    Messages:
    1,686
    Likes Received:
    131
    Trophy Points:
    81
    I guess... assuming you had infinite amounts of subdivisions that had to complete eventually, then more threads and more cores would work in theory. But like you say, the way it's set up doesn't make things actually go faster without more cores (assuming there are more subdivisions that can run independently).

    The reason for that is perfectly demonstrated with a video decode over increased threads, or with something that really consists of the same instruction set resolution running over and over on different data. What happens here is that the context changes are cheap, and when each of the threads run in turn, the actual instruction to run on memory is already completed. And the new threads take advantage of that.

    If, on the other hand, you ran threads that require completely different instruction resolution every time. The the context changes are expensive and running things "concurrently" isn't really giving you many benefits on the actual computation element. Instead, we're really talking about being able to prepare for the calculation runs asynchronously. Which is useful if the math operation on the core that's reduced from a more complex instruction set is repeated over and over again.

    But the thing is that because the cores on an intel platform don't actually have access to main memory, you end up waiting for the main thread to complete something anyways. It's completely useless if you wanted to maintain a main thread and asynchronously run different instruction sets on working memory.

    Basically, you can't scale over more intel cores unless you have operations that can be divided into separate and independent tasks, that also are small enough to have a short preparation stage.
     
  30. cdoublejj

    cdoublejj Notebook Deity

    Reputations:
    195
    Messages:
    1,123
    Likes Received:
    92
    Trophy Points:
    66
    Ht uses un used parts of the core, sometimes time sensitive or priory calls will get handed over to hyper thread instead of actual core and either has to work at diminished capacity OR has to wait for the for current computation to finish. some games and apps work better with HT turned off.

    I was more interested in a six core due to VMing and multitasking.
     
  31. Dufus

    Dufus .

    Reputations:
    1,194
    Messages:
    1,336
    Likes Received:
    548
    Trophy Points:
    131
    Depends what your using.

     
  32. HTWingNut

    HTWingNut Potato

    Reputations:
    21,580
    Messages:
    35,370
    Likes Received:
    9,877
    Trophy Points:
    931
    Bottom line, there are no 6-core mobile CPU's. Your only option is for a six core desktop CPU in the Clevo beast.
     
  33. n=1

    n=1 YEAH SCIENCE!

    Reputations:
    2,544
    Messages:
    4,346
    Likes Received:
    2,600
    Trophy Points:
    231
    Which sadly has been discontinued. :(
     
  34. HTWingNut

    HTWingNut Potato

    Reputations:
    21,580
    Messages:
    35,370
    Likes Received:
    9,877
    Trophy Points:
    931
  35. nipsen

    nipsen Notebook Ditty

    Reputations:
    694
    Messages:
    1,686
    Likes Received:
    131
    Trophy Points:
    81
    ..the thing is that even very craftily optimized assembly code isn't going to take advantage of either hyperthreading or multiple cores directly. The way it's done is that at the point where the compiler has created assembly language, then this code is interpreted and executed in a different layer. When you specifically address a register (outside the reserved ones), that register is duplicated in a different memory area. And at that hardware layer is where the optimisations that can take advantage of hyperthreading happen.

    But since reduced code like that tends to end up being identical instructions, it's actually rare that hyperthreading (as in duplicating the low-level registers and sharing two instances over the same core) doesn't give you free execution cycles. Or, that there's room on the computation elements while one instruction set queue is waiting for another. Basically, the processor can make a couple of extra identical runs for the other duplicated memory area before continuing. And.. counterintuitively, in general you're better off writing code that has a wide range of different calls rather than identical ones. Because that mix on the assembly code level is more likely to yield opportunities for the machine level interpreter to steal cycles. It's a form of instruction level parallelism, I think the term is. And that is really where most improvements have been made since the Pentium was introduced. They employ a limited form of simd in a sense, transparently on the hardware layer - and kept on that level is what keeps the overhead acceptable. Programming your own instructions is not an option on this architecture.. Closest we're getting to that at the moment is OpenCL over a graphics card's execution units. That you write your own low-level code specifically designed to have simple maths executed in several semi-independent parts on a memory area. Cheating a curve with some interpolation, that sort of thing.

    Cores on the level above that only exchange information via cache ("level 3" cache. Level 2 and level 1 is for each core). But that's how the cores are able to support core-level parallelism, in the sense that one core can maintain several hardware-level threads (pointing back to the crunched assembly code - one function call could spawn several hardware threads). And the thing is that outside the hardware level optimisations, cores don't usually share execution time on the same thread (to speed up the execution of one single hardware thread). So that's the second great advance. That a single core can switch between hardware threads with very little overhead

    Instead, multiple cores on the same processor is what can support execution of different OS-level threads. Which again is usually handled most effectively by the compiler. If you program threads and assign them to a specific core, and expect something to execute in time for the main thread, etc. - that's always risky, because you won't necessarily have complete control over what else is run on the core, and the execution time could be variable anyway (or turn out to be more effectively handled by the computer and the compiler). So like with the instruction level parallelism, the goal with creating threads is really to create threads with tasks that easily can take advantage of the OS' scheduler. One common way games-programmers do it is to have one thread for graphics and one for core logic, for example. That's.. really it. Optimising one thread to fetch and place information on the graphics card array (then optimise that code independently, perhaps phasing in physics or particular routines that scale well over gpu elements), and having another to do maintenance and updates in memory otherwise.

    And as said, the problem here is that increasing the number of cores on the same independent processing element has a nasty tendency to end up with diminishing returns very quickly. In particular for real-time computation tasks. Whether it is games or graphics intense programs, and all the way down to just dealing with input and scrolling gracefully in a graphical ui. You know, stuff like drawing a perfect curve in real-time incurs the wrath of the computer gods. This is kind of a problem, and it's what hopefully will force the next.. 50 years or so to end up with companies dusting off the simd designs from the 70s again, and force ahead a common industry standard for an explicitly parallel assembly language. One that scales over different processing elements with their own registers, and where those processing elements share execution memory in a decent size capable of storing the program logic.

    This.. doesn't require all that much, though. Size isn't really that much of a concern as many think - you could have slower external ram on top of it. The tech exists, but it requires a different approach to the architecture design. And ddr3&4 and so on isn't really sufficient. You'd need ability to lock and write concurrently to different areas at the same time with the different processing elements to be at all able to exploit multicore parallelism. In the sense of having several processor elements execute instructions concurrently on the same memory area to complete that software "OS"-level thread.

    That.. contrary to the impression people seem to be having very often.. isn't actually occurring on a hexacore intel or amd processor. Where shifting context between processors is, practically by definition, less efficient than using one core. Because you have only one core executing useful computations towards the product at one time. I.e., adding another core to the same thread doesn't increase computation potential, but only overhead during context shifts.

    Funny thing, yeah?
     
    Jarhead likes this.
  36. Dufus

    Dufus .

    Reputations:
    1,194
    Messages:
    1,336
    Likes Received:
    548
    Trophy Points:
    131
  37. nipsen

    nipsen Notebook Ditty

    Reputations:
    694
    Messages:
    1,686
    Likes Received:
    131
    Trophy Points:
    81
    :) ..arm chipsets have cores with custom instruction sets, though. Set from the manufacturer as per order. The "graphics card" on several arm-units is actually an arm core with a custom instruction set, for example.

    But I guess that's the way we're going. Cheap "special purpose" cores for each typical common task the computer performs. That each operate independently, which in turn cheapens the manufacturing cost and makes factory-line reuse easier. Nintendo from the 80's must be laughing pretty hard at that one..
     
    Mr.Koala likes this.
  38. cdoublejj

    cdoublejj Notebook Deity

    Reputations:
    195
    Messages:
    1,123
    Likes Received:
    92
    Trophy Points:
    66
    how bad do they over heat and throttle ...or jet make jet engine sounds?
     
  39. tijo

    tijo Sacred Blame

    Reputations:
    7,588
    Messages:
    10,023
    Likes Received:
    1,077
    Trophy Points:
    581
    Jet engine sound is what happens. There are a ton of fans too 3 or 4 instead of two and the heatsinks are monstrous.
     
  40. TomJGX

    TomJGX I HATE BGA!

    Reputations:
    1,456
    Messages:
    8,707
    Likes Received:
    3,315
    Trophy Points:
    431
    Yup, those laptops are huggeee... Weighs over 12 pounds... Not to mention, 2 330W PSU's are a must to run at full perofmance due to the SLI/CF + 130W desktop CPU...