How many transistors in i7 2600k
The most notable difference is the core count, as the Core i is a native quad-core chip using a new, smaller die than the flagship Sandy Bridge-E. This reduction in cores greatly helps reduce the price. Those Hexa-core chips, remember, are actually eight-core dies with two cores disabled for both technical and marketing reasons.
With eight cores aboard plus boatloads of cache, those Hexa-core Sandy Bridge-E chips are gigantic in size, which lowers yields and contributes to the cost. Despite being a quad-core, it actually packs far more transistors than standard Sandy Bridge chips such as the Core iK. It has the same 40 PCIe 3. The core can detect before an instruction is decoded if that particular instruction has been decoded recently, and use the result from the previous decode rather than doing a full decode which wastes power.
Now the uOps are now in an allocation queue, which for modern cores usually means that the core can detect if the instructions are part of a simple loop, or if it can fuse uOps together to make the whole thing go quicker, it can. In the back end, starting with the re-order buffer, uOps can be rearranged depending on where the data each micro-op needs is. This buffer can rename and allocate uOps depending on where they need to go integer vs FP , and depending on the core, it can also act as a retire station for complete instructions.
After the re-order buffer, uOps are fed into the scheduler in a desired order to ensure data is ready and the uOp throughput is as high as possible. In the scheduler, it passes the uOps into the execution ports what does the compute as required.
Some cores have a unified scheduler between all the ports, however some split the scheduler depending on integer operations or vector style operations. Most out-of-order cores can have anywhere from 4 to 10 ports or more , and these execution ports will do the math required on the data given the instruction passed through the core.
Execution ports can take the form of a load unit load from cache , a store unit store into cache , an integer math unit, a floating point math unit, vector math units, special division units, and a few others for special operations. After the execution port is complete, the data can then be held for reuse in a cache, be pushed to main memory, while the instruction feeds into the retire queue, and finally retired. The biggest change for Sandy Bridge and all microarchitectures since is the micro-op cache uOp cache.
The decode hardware is a very complex part of the x86 pipeline, turning it off saves a significant amount of power. The cache is direct mapped and can store approximately 1.
You get slightly higher and more consistent bandwidth from the micro-op cache vs. I've personally asked TechArp to correct the specifications on few of the products. Some of them they just have to take a guess as official information is hard to come by. Though specifically regarding this I'm pretty sure they have gotten the same number as the review sites did.
Yes David, this is all kinda weird. Weird in that why would Intel not include the transistor count spec in the official CPU information page on their web site? If it is a little below or above one billion, either is a huge accomplishment IMO. If it was actually less than say an i series four-core CPU, which are million, so what?
Better overall performance with less transistors would be a great accomplishment as well. Or due to the Western culture's paradigm of more and bigger always meaning better, would that be perceived as not better? Is there a marketing concern over that? I could understand that, as the semi-informed would always consider more as better. Another factor could be the inclusion of the on-CPU graphics processor, whose transistors are not part of the regular CPU functioning. Are the graphics core transistors part of the count or not?
I'm sure they are, but what does all this nit-picking really do for us? Nothing IMO. Can you imagine if the reason for not including the transistor count as million is so the Intel executives that stated it as 1.
Frankly, I could care less, blame an engineer for providing an incorrect figure, done and over. Don't get me wrong here David, I like all the specs too, but does it really matter all that much? They don't make separate die for what's a trivial difference. For your information, I'd like to direct your attention to the i and the i These are two processors in the Sandy Bridge line and the same series that have a difference of million transistors. On the surface, the difference is HD vs.
They have like specs. This is repeated throughout Intel's processor history so in my opinion, it's even more incorrect to suggest that Intel is locked into a specific die, TC or size for any series. Even for a reason as "trivial" as graphics presentation as you seem to think. Intel has admitted publicly if you can read between the lines that they add un-necessary circuits to dies and that makes transistor count trivial in itself.
The problem was caused by un-necessary circuits. CarWiz, Apparently David was not aware of the difference in the graphics core between the i and ik. There are actually three different versions of the Sandy Bridge die shipping at launch. Finally, the slimmest variation sports two cores and a graphics engine composed of six EUs.
In its current state, Sandy Bridge-based processors are available with four cores with and without Hyper-Threading and two cores dual-core models all have Hyper-Threading enabled. Still present are the 32 KB L1 instruction and data caches along with KB L2 cache per core , though Sandy Bridge now incorporates what Intel calls a L0 instruction cache that holds up to decoded micro-ops. This feature has the dual effect of saving power and improving instruction throughput.
I ran these two single-threaded tests as a synthetic comparison of performance, clock for clock. As you can see, just the architectural shift makes a significant impact on Sandy Bridge's performance versus the Nehalem-based Lynnfield design. The impetus behind AVX comes from the high-performance computing world, where floating-point-intensive applications demand more horsepower than ever.
Unfortunately, there aren't any real-world apps optimized for AVX that we can test as a gauge of the capability's potential.
0コメント