embryo2 wrote:About performance - smaller switches lead to lesser current and lesser power requirements and in turn it leads to the lesser heat, while with the same heat it gives the ability to increase number of current switches per second. But from some point processor's frequency had stopped to grow. What is the reason for it? There's still too much heat despite the lesser switch current. But why?
As process node sizes decrease, sure the smaller switches become more power-efficient, but this improvement is counterweighted by the fact that the circuit is now physically more dense and now heat removal becomes trickier.
There are a lot of conceptual issues with running up the clock rates that don't relate directly to the transistors themselves. For example, as the clock speed increases, digital circuits start to behave much more like analog systems, where it becomes much more difficult to control things like impedance, reflections, etc. EMI/RFI begins to become an issue, because at higher speeds now the trace length becomes a bigger fraction of the effective wavelength of the signal that trace is transmitting. Things like that require specialized dielectric controls to maintain the proper RF qualities in the silicon.
At any rate, many systems external to the CPU fail to scale in exact proportion to CPU speeds, such as I/O components and RAM. There's a reason we're in the DDR4 era now and core RAM clock frequencies have not exceeded 1.6 GHz or so. Granted, we can muss with that a little with the dual-timing of double data rate. But generally, running up the core clock has the potential to generate or exacerbate instabilities between the CPU and the rest of the system.
Probably the biggest factor in it is oddly enough, Intel. Processor cores (and Intel cores especially) were on the rise across the industry generally up to the Pentium 4. The NetBurst microarchitecture was designed to scale to above 8 GHz, but it didn't anticipate the power concerns of running generally above 3 GHz, and above that speed (>= 4 GHz or so) the thermal dissipation shot through the roof. Those cores are the ones favored by overclockers because they hit 8.1 GHz, but it takes liquid nitrogen to get them there. After NetBurst, Intel figured that it was impractical to produce desktop processors that went that fast because they'd command exotic cooling systems simply to function, and therefore high-3s was the highest they could push the speed, and so most devices' PLLs are designed to cap their multipliers around that range.
The subsequent effect was that Intel pursued increasing performance by other means; namely, multicore processors, multithreading, larger caches, deeper and more superscalar pipelines, higher bus speeds. Intel pushes fab technology along quite well - they were one of the big spurs behind the development of EUV lithography - and so generally if they do something the industry at large generally likes to read the writing on the wall. Hence, no CPUs coming out higher than 3.5 GHz or so. At any rate, once the systems were developed for handling multicore processing well, everyone probably just figured they'd keep doing that (jamming on more cores, optimizing the core communication schemes) rather than trying to force single cores into the millimeter wave range.
Recently Intel has started to show signs of questioning that philosophy, as recent Xeons have scaled (via Turbo Boost) to 4 GHz. So, maybe in the grand scheme of things, the "speed versus parallel" argument might begin to lean in favor of speed again.
That being said, there's nothing that says that transistors have not continued to get faster. IBM several (~10) years back announced graphene-based transistors with a practical switch cutoff of 500 GHz. Nowadays, the development of "exotic" transistors such as MODFETs are routinely being shown to have power gain capacity into the 1 THz range. Why we aren't building microprocessors with them is a separate discussion, however.
embryo2 wrote:Is it only theoretical issue (lack of proper mathematical model to calculate some technology parameters)?
I wouldn't say so. If we're able to hit 13nm we're able to go lower. IBM announced right about a year ago that they managed to demonstrate functional transistors on a 7nm SiGe process. The first commercial production at 7nm is likely to start within the next five years or so.
I'm just meaning that as we approach the really small scales, suddenly quantum effects are going to become more and more of an issue. The notion that transistors have properties that can be explained by formulas, such as transconductance, intrinsic charge, etc., is increasingly having to yield to statistical mechanics and quantum mechanics approaches because we're approaching the ranges where those things suddenly become relevant. It's sort of like saying if you're the size of an flea, now an ant is a much more formidable and closely-sized enemy.
embryo2 wrote:What is the advantage of such layout in terms of performance or cost or power consumption? Isn't it similar to a number of separate devices or it could deliver some essential advantage in speed, power, cost?
Are you talking about 3D ICs or IC design in general? The laying out of a circuit into distinct areas based on function and technology type is a core stage in semiconductor design known as
floorplanning - it's been standard practice ever since integration became a thing. One might also consider the influence business models such as ARM have on the industry. ARM sells processor cores and peripherals to companies who then lay them out and fabricate them. By running a highly-compartmentalized footprint, different varieties of the same system-on-chip family can be created simply by changing out the amount of RAM/Flash and the set of peripherals, but the processor core itself remains constant. That's why when you go to buy embedded processors, you often find a multitude different options varying in number of communication buses, amount of program/data memory, 2/4 core options, specialty units like cryptographic coprocessors, etc. That's all because most of the stock of those variations is the same - they're punched from the same die.
As to 3D ICs, the big bonus is decrease in routing congestion, both on the die and on the circuit board, as well as a general increase in board density because of the vertical stacking. There might also be power considerations - most devices nowadays run at a voltage level of 1.0V or 1.2V as their internal logic voltage. For interfacing to the external world, they have to use driver stages to step that up to 1.8V, 2.5V, 3.3V, standard CMOS voltage levels. 3D IC constructions allow for the different stacked layers to continue to talk to each other at their internal voltage levels without requiring the power driver stages. That's a power bonus, although how significant - I can't say.