top | item 25904609

New Transistor Structures At 3nm/2nm

112 points| mrnode | 5 years ago |semiengineering.com | reply

80 comments

order
[+] tyingq|5 years ago|reply
This drawing is a bit clearer and more complete than the one in the article: https://images.anandtech.com/doci/16041/SamGAA_575px.png

Note that "MBCFET" is Samsung's name for their "nanosheet" FET.

And the Anandtech article it comes from: https://www.anandtech.com/show/16041/where-are-my-gaafets-ts...

[+] kens|5 years ago|reply
Are there drawings with more detail? I'm unclear where the gate dielectric is and the channel. The silicon is doped differently "inside" the gate to form the channel? That seems hard to fabricate. (I searched around a bit but couldn't find a better diagram.)
[+] ashz8888|5 years ago|reply
Just wanted to point out that 3nm or 2nm are nothing but the marketing terms. The physical channel length will continue to remain at around 10 nm in this decade, irrespective of the device architecture. See Fig. 1.2 of my PhD thesis: https://etheses.whiterose.ac.uk/22492/1/Novel%20Approaches%2...
[+] colejohnson66|5 years ago|reply
Why is that? What’s the point in calling it “3nm” when it’s not?

<rant>

Also, how is that not false advertising then? If I’m told a device is 5nm, and that traditionally meant the gate length, why isn’t it a lie when they’re actually 10nm or whatever? We need to stop calling it “marketing terms” and call it what it is: lies. Just because everyone does it doesn’t make it right.

</rant>

[+] ericlewis|5 years ago|reply
Why is that? (Sorry, the paper is beyond my understanding)
[+] ur-whale|5 years ago|reply
One thing I've always wondered about when it comes to new process design at smaller scale: how much actual quantum mechanics is actually needed to get the job done?

And ... if the answer is, as I suspect, a lot, what kind of numerical methods and processes are used to design and simulate these tiny quantum mechanical machines?

[EDIT] I mean, when taking a basic QM course, there is a lot of contorsions to try and find analytical solutions to the Schrödinger equation, but as soon as you have three particle interacting with each other, analytical methods run into a wall.

Am I right to think that sub-10nm process design is all done numerically?

Anyone who happens to work on this type of problems care to give pointers?

[+] typon|5 years ago|reply
Depends on what you mean by get the job done. (Rather which job)

If you are doing research into designing advanced transistors with new geometry or new materials (which is what I did my graduate research in), you would be using something like DFT (Density functional theory) for equilibrium analysis and NEGF, Huckel theory etc. for simulating current. These methods only realistically work on ~500-1000 atom systems, beyond which the simulation takes too long to run even on supercomputers (which is what i was using). I think GPUs here would be very useful, but there weren't any tools at the time that were seriously optimized for GPU. The codes I was using were SIESTA/TransSIESTA, Atomistix, QuantumEspresso and others.

For simulating multiple transistors, or transistors with a large geometry (for example 14nm gate length), you would use TCAD simulators that use FEM + measured parameters to simulate the transistors. The equations behind these are traditional semiconductor equations with a bunch of heuristics and curve fitting. The main tool I used was Sentaurus TCAD.

For simulating larger circuits, say a low-noise amplifier or maybe a small DAC, you would use tools like Cadence Virtuoso + the provided PDK from your foundry. The equations here are simpler than the ones used in Sentaurus and they are also calibrated to measurement.

[+] siver_john|5 years ago|reply
Full disclosure, I don't work on processors but am in a tangentially related field.

However, I don't feel that you need to have an analytical solution to the Schrödinger equation. In fact even in chemistry we don't do analytical solutions instead using fancy basis sets which allow us to do approximations.

Regardless, I don't think even that is particularly necessary, as quantum at that level means you basically have some level of leakage where the electrons can just tunnel through the barrier created by the transistor when off. So if I had to guess most of it is just ways to rectify this leakage so it doesn't effect calculations, probably similar to a form of error correcting.

(This ignores that you may have to do some initial quantum calculations using Density Functional Theorem to get a guess at how much leakage based off the materials you are using, though if I had to guess most of that work was done a while ago.)

[+] patcassidy2000|5 years ago|reply
Solid State Physics has a bunch of different types of models to try to explain the behavior of electrons in semiconductors and conductors. Most of them are only valid under specific situations and they sometimes give erroneous results unless you apply them carefully. That being said, your absolutely right that the industry uses approximations. They usually use some kind of simulation physics package similar to the ones used by EE engineers when designing circuits.
[+] mlindner|5 years ago|reply
Hacker news comment sections on articles about semiconductors always seems to be full of people talking authoritatively about things they clearly don't understand. And this is all visible to someone who only majored in computer engineering and got basic training in transistors and logic design who's day job is now in software. I can only imagine the people who actually know transistors cringing in here (and thus avoid commenting).
[+] pvarangot|5 years ago|reply
I worked in an aerospace startup for seven years, and while not an expert I really know my way around some spacey stuff. People speaking authoritatively about that on HN also make me cringe, I rarely intervene because they almost always engage in a discussion about how they are right even when I send references to papers and book chapters. In Reddit it's even worse.

With law and politics I feel there's a similar attitude going around but it's of course more an up for debate topic.

[+] mrfusion|5 years ago|reply
It’s not just articles about semi conductors ... you just happen to have some knowledge in that area so you notice it.

You’ve just discovered another instance of the gellman amnesia effect.

[+] guardiangod|5 years ago|reply
I agree with you (I have been following silicon fabrication news for 20 years already). I cringed when I saw people commenting on, say Intel replacing its CEO, or that RISC-V will replace ARM for design houses in the next few year, when it's clear the commentators haven't taken the time to do even a cursory research.

I don't even want to think of the comments I've seen for my main subject (cybersecurity, exploits, and vulnerabilities). I will never be able to correct all of them.

I think some humility would be nice for all of us. We know what we know and we should be aware of what we don't know. I don't know much about web programming, and I freely admit to it. I certainly don't pretend to have anything insightful to say about them.

[+] lizknope|5 years ago|reply
I do semiconductor digital physical design for the last 23 years. But almost everything I do is at the standard cell (AND/OR/flip flop) level. I generally don't do anything at the transitor level. I have seen some good comments here in the past.
[+] deeeeplearning|5 years ago|reply
Gell-Man Amnesia might interest you.

"Briefly stated, the Gell-Mann Amnesia effect is as follows. You open the newspaper to an article on some subject you know well. In Murray’s case, physics. In mine, show business. You read the article and see the journalist has absolutely no understanding of either the facts or the issues. Often, the article is so wrong it actually presents the story backward—reversing cause and effect. I call these the “wet streets cause rain” stories. Paper’s full of them.

In any case, you read with exasperation or amusement the multiple errors in a story, and then turn the page to national or international affairs, and read as if the rest of the newspaper was somehow more accurate about Palestine than the baloney you just read. You turn the page, and forget what you know.”

[+] newen|5 years ago|reply
It's not just computer engineering. I'm in natural language processing and was in academia and I see a lot of mistakes and misunderstandings in the comments in those topics. Probably the only thing reliable in HN comments is website design.
[+] MayeulC|5 years ago|reply
I tend to disagree; while some comments seem to be clearly wrong, mistakes tend to be pointed out by other commenters, and you can always check bios. It's a bit riskier on stories that don't stay on the front page, I'll grant you that.

And there can be a bit of Dunning-Kruger if you work in a related field too.

[+] mrRandomGuy|5 years ago|reply
You can expand this astute observation to _many_ things other than semiconductors around these parts
[+] LatteLazy|5 years ago|reply
Silicon atoms are about 0.13nm apart. 2nm is about 15 atoms wide. Mass producing anything on that scale is an exceptional feat.
[+] Denvercoder9|5 years ago|reply
It's been a while since node names were actually physical measurements though. For example, at the 3nm node the fin width is 5nm (as noted in the article).
[+] ur-whale|5 years ago|reply
>Mass producing anything on that scale is an exceptional feat.

Mass producing anything on that scale that will then go on working for years at ~ 3Billion movements per second reliably is simply astounding.

[+] jabl|5 years ago|reply
It's the closest thing to black magic mankind has achieved.
[+] everdrive|5 years ago|reply
What is the benefit of smaller transistors? Serious question. Why does it matter if I have 7nm vs. 5nm vs. anything else?
[+] mng2|5 years ago|reply
Besides being able to pack more transistors into a given area, a smaller transistor has less gate capacitance (to first order). This means it can switch faster (smaller RC time constant) and less energy is expended in switching. Thus, going from generation to generation, the overall energy expenditure of a chip can be kept within a reasonable range despite adding many more transistors. You also may have heard of a "die shrink", where an existing design gets shrunk to the next technology node, using less power and clocking faster.

Shrinking isn't always a walk in the park though. Some nodes ago subthreshold leakage became a big problem until they figured out how to solve it.

[+] tryptophan|5 years ago|reply
Smaller distance -> lower resistance -> less heat -> higher clocks/more stuff per clock allowed with same amount of heat produced -> higher performance.
[+] jabl|5 years ago|reply
The marginal cost of manufacturing a chip is roughly proportional to area. So with smaller transistors you can fit more of them into a given area.
[+] t-writescode|5 years ago|reply
It's crazy to think about if you've never thought about this, but the speed of light is a bottleneck for processors. When we get smaller devices, there's literally less distance that needs to be traversed, so more can be done!
[+] femto113|5 years ago|reply
The numbers roughly correspond to the width of wires in the circuits, but the number of circuits you can fit per unit area depends on the square of that number so going from 7nm to 5nm roughly doubles density. The first microprocessor[1] was around 10,000nm so we're approaching 5,000x thinner wires or almost 25 million times more circuits (the latest Apple M1[2] is at 5nm have about 8,000,000x as many transistors as the 4004).

[1] https://www.intel.com/content/www/us/en/history/museum-story... [2] https://www.apple.com/newsroom/2020/11/apple-unleashes-m1/

[+] lizknope|5 years ago|reply
Smaller distance to travel so signals get from one gate to another quicker which will enable a higher clock speed.

Smaller devices use less power so less heat and longer battery life.

Smaller devices mean a smaller chip which is cheaper (although mask costs will be more expensive) or use the extra area for more features like more cache or another processor core.

[+] ksec|5 years ago|reply
> Serious question...

To Oversimplify.

With a Fixed Yield, and an exact 100% increase in Transistor Density that translate to 50% smaller Die Size.

On a Wafer, that would equate to Double the amount Die you have. All of a sudden your profits increase dramatically.

5nm also have a better power curve so within the same clock speed you have lower energy usage. Hence you can push for higher performance if needed.

The first point of Uni Economics is important for the industry. If you have high enough volume, say hundreds Million of chips per year then it make sense to move to the next node for cost saving. If you have small volume or low margin chip then the Design Cost, which is the most expensive part of chip making, would not work to your benefits.

And it also depends on Wafer price, If 5nm is Double the Price of 7nm then in the above example your unit cost would be exactly the same.

The second point is important for CPUs, and other things that are increasingly computational expensive like WiFi 6 and 5G Modem. You want your Smartphone to last longer on battery so they work better on an energy efficient node.

So basically it is a Cost / Performance trade offs.

[+] LatteLazy|5 years ago|reply
Smaller means closer together. Closer together means less time for a signal to move from one to another. Less time means higher clock speeds.

If you CPU is 100mm across, the speed of light limits it to 3GHz because that's how many times you can cross the cpu travelling at c. At 10mm you get 30GHz.

[+] bigmattystyles|5 years ago|reply
You can pack more within the same area for one. A multitude of other constraints come in but that’s a big reason.
[+] mjevans|5 years ago|reply
Rather than the change in transistor design, I think the bigger news is the switch from silicon with dopants to silicon with germanium and dopants. The drop in threshold voltage from ~0.7v to ~0.3v might be one of the last levers left in extracting even more performance; at the cost of making semi production and equipment even more hazardous.
[+] j_walter|5 years ago|reply
What do you mean "silicon with germanium and dopants"? Implanting germanium as a dopant is already done at much larger geometries than 2/3nm. It's also not any more hazardous than implanting any other ion.
[+] x86_64Ubuntu|5 years ago|reply
Would you mind explaining how the production becomes more hazardous?
[+] amelius|5 years ago|reply
According to the article, the new transistor design offers the promise of lower leakage.
[+] phkahler|5 years ago|reply
Why can't they use a vertical channel? Layers are a thing, and vias are a thing? Why aren't vertical channel transistors a thing?
[+] wallacoloo|5 years ago|reply
From what I can gather looking at the images and such, the device is one gate with 3 or more isolated channels, each with separate source/drains. Are these processes constrained such that all the sources/drains have to be linked together later? Or can they be used independently, allowing the designer to construct 3 or more transistors with a shared gate?
[+] riskable|5 years ago|reply
I think what the world needs more right now is more foundries and not smaller manufacturing processes. The global chip shortage totally sucks and it's because there's so few players and all the innovation is focused on stuff like this instead of figuring out ways to produce ICs faster and cheaper.