Тёмный

End of the silicon era. Processors of the future 

My Computer
Подписаться 22 тыс.
Просмотров 313 тыс.
50% 1

The era of silicon chips is coming to an end. New processors come out hot, and everyone forgot about Moore's law. Will the development of electronics stop? What will be the processors of the future? Is there a replacement for silicon?
00:00 - The purest polysilicon
01:26 - Silicon limit
03:32 - What if not silicon?
08:50 - Rejection of CMOS
13:29 - Changing electrons to photons
14:53 - Quantum computer
#mycomputer #processor #silicon

Наука

Опубликовано:

 

17 май 2024

Поделиться:

Ссылка:

Скачать:

Готовим ссылку...

Добавить в:

Мой плейлист
Посмотреть позже
Комментарии : 800   
@blinded6502
@blinded6502 Год назад
Quantum computers are NOT a replacement for the classical computers, holy hell
@DocWolph
@DocWolph Год назад
A Classic Computer (C-Comp) becomes the I/O device to talk to the Q-Comp. Input-> C-comp-> Q-comp-> (back to)C-Comp-> Output
@cebo494
@cebo494 Год назад
​@@DocWolph It doesn't matter what the connection is, quantum computers simply are not meant to do the same things that classical computers do. That's what the guy meant by "are not a replacement". They are not better than classical computers for most of the tasks that the vast majority of people use computers for, like web browsing, text/image processing, gaming, etc. They have little to no values for consumers. They will likely forever remain in labs, datacenters, and industry for highly specialized tasks and research.
@fridolinkoch
@fridolinkoch Год назад
Until someone discovers something you can do with them as a general person….just like when the internet arrived
@Galacticfungus
@Galacticfungus Год назад
@@fridolinkoch There are a few Quantum ray tracing algorithms with papers now.
@DocWolph
@DocWolph Год назад
@@cebo494 Well at this point it is implied, at least, that you can not directly interface with a Q-Comp. That is, you need some kind of bridging software running from the C-Comp to the Q-Comp and back. This is regardless of application. Aside from scientific, engineering, security work, among other things, I can see Q-Comps being used for animation, rendering, or simulation (for example, I can see water sims taking only a few minutes rather than many hours or days for a FEW high quality frames with a Q-Comp) and that is not even the color of the tip of the iceberg of what is possible with Q-Comps. For most things people use Computers for, C-Comps will be at least enough.THIS is agreed. But there are things that are way bigger that even best current C-comp technology just is not adequate for. and Q-Comps at a price a small studio, or a very dedicated hobbyist, can buy to radically accelerate their sim work, animation, and rendering work (albeit the software may be coming a few years after the fact) is just what the Studio head asked for. And that is JUST animation. Anybody can become a "Doc Brown". pursue science, mathematics, engineering and MORE at home. Again, I generally agree with what you saying BUT you are not thinking broadly enough.
@cursedgamer2778
@cursedgamer2778 Год назад
I love these dual language channels, I don't know Russian and would have never have been able to understand your phenomenal video otherwise. Thank you, truly.
@tankerock
@tankerock Год назад
No wonder, I was wondering why my video/audio was out of sync lol
@ysa1023
@ysa1023 Год назад
Ah so that's why his disgusting face seemed so familiar and audio was out of sync! He is one of the idiots that supports war in my country. He is one of the people that thinks I should die... Yeah, hope at least somebody is going to see this comment, and stop watching this genocide supporter.
@mka2
@mka2 Год назад
Спасибо друг! Это очень приятно читать! Привет из холодной России!
@headSoup
@headSoup Год назад
@@mka2 привет друг! я из Американский
@headSoup
@headSoup Год назад
@@mka2 я говорю просто 😢
@DunnickFayuro
@DunnickFayuro Год назад
Very good content. I just find the desync between your video and sound a bit annoying.
@PareshPatel-xc2vu
@PareshPatel-xc2vu Год назад
I think it's dubbed, probably from Russian
@DunnickFayuro
@DunnickFayuro Год назад
@@PareshPatel-xc2vu Nope. Just desync. The lips almost match the sound, with a little delay.
@namthainam
@namthainam Год назад
Nope video is all ai generate from stock video so it's all jibberish. Might as well listen with your eyes closed
@samlovebutter
@samlovebutter Год назад
@@DunnickFayuro it is dubbed from russian. The original channel called "Мой Компьютер"
@vadulme8640
@vadulme8640 Год назад
@@DunnickFayuro Yeah, it is dubbed. I am the guy who recorded the dub. I synced it the best I could to the original voice, but I could only do so much. And yes, I get it that it's annoying.
@ajbowers44
@ajbowers44 Год назад
So glad RU-vid recommended this video. Nicely done, great balance of information and presentation without coming off overly optimistic or pessimistic.
@Andrew-rc3vh
@Andrew-rc3vh Год назад
Yes indeed, so far it sounds intelligent. I'm hoping I'm not going to be lectured about Moore's law. I'm convinced that is a marketing meme for Intel.
@hyll6700
@hyll6700 Год назад
Quantum Computing isn't a replacement.
@AliensKillDevils.
@AliensKillDevils. Год назад
Mt.Fuji is about to explode. Excess silicone should be mined. ru-vid.com/group/PLen6gjewxJdCB0VDrHJiylc3B3qh4b91m
@aqua-bery
@aqua-bery Год назад
Quantum computers can't replace regular computers. Yes they could calculate some very complex equations at a fraction of the speed. But they suck at doing simple calculations.
@hyll6700
@hyll6700 Год назад
@@Andrew-rc3vh Bruh 😂
@supriyochakrabortybme
@supriyochakrabortybme Год назад
A very interesting compilation, the RU-vid algorithm recommended this hidden gem.
@kevinmitchell3168
@kevinmitchell3168 Год назад
You mentioned photonics use in communications but didn't mention photonics switching. They've been trying to make purely photonic chips for decades, and it's always just around the corner much like fusion. It's that research that led to photonics being integrated into silicon chips. It seems like that research was very close to yielding results but focus and funding got taken over by quantum computing. From all I've read purely photonic chips would run 100-1000 times faster than silicon at far lower power and heat. Hopefully as silicon reaches it's limitations there will be renewed funding and research for it.
@Luizfernando-dm2rf
@Luizfernando-dm2rf Год назад
It's a bit sad really... Quantum computing while exciting, is very far from producing anything useful in terms of actual computing. There has been no more than 10 algorithims made to run on quantum computers for example.
@id104335409
@id104335409 Год назад
The comment I was searching for. I was hoping the video is on that topic. Not quantum computing.
@bdjfw2681
@bdjfw2681 Год назад
isn't fusion have focus and funding but still no result ? is it really that close for purely photonic chips ?
@dinozaurpickupline4221
@dinozaurpickupline4221 Год назад
light switches have their application in networking & data transfer
@jadedandbitter
@jadedandbitter Год назад
Intel's working on it. Have a buddy that works in the photonics dept. Can never talk about where they're at though.
@yamatsukami987
@yamatsukami987 Год назад
also worth considering in the mean time, their are computers that can run on trits (-1, 0, 1) which requires some fundamental changes, but could theoretically be more effective. This could even be extended further though it gets less practical the more you add.
@ctsd623
@ctsd623 Год назад
Re: "silicon being near the end of its life cycle," this is just sensationalistic journalism (or an uneducated remark). It's not going anywhere in our lifetimes due to the deeply entrenced and highly refined manufacturing process. More advanced materials will have a much higher and impractical cost of scale. The most certain outcome is slowly more advanced materials will be used in conjunction with silicon, but it isn't anywhere near ending its life cycle, not for a hundred years or more. We can even do photonic and quantum compute on silicon.
@josgraha
@josgraha Год назад
oh boy, quantum computers are even further out than tfets, optical compute, or memristors and you still need a von-neuman silicon computer to work with it. thanks so much for the great video, fantastic questions and background material. much appreciated
@ararune3734
@ararune3734 Год назад
Not only are they far away from any practical implementation, but they will simply never replace your standard computers in the way people think. Quantum computers will be better at performing certain tasks, but those tasks are not in our every day usage, so for your personal needs, you won't use a quantum computer, you'll stick to the standard computer.
@tony_T_
@tony_T_ Год назад
Silicon photonics is what I have the most hope in for in the next couple of decades. Imo, transitioning from electricity to light is just the most logical step forward. It will set moores law back by quite a bit, but the insane clock rate of the processors will make up for it. Most modern keyboards already use light to transmit signals.
@HeLrAiSiNg1
@HeLrAiSiNg1 Год назад
well light transmision exists for decades now unless you mean signals like hid interface that uses speed bandwidth like in hz its known that for ex most audio devices ps5 xbox etc have a light slot on the back to connect with optical fiber to send sound now a kb uses a microcontroller if its mechanical that works like microcomputers with speeds of some mhz so either at the usb there is a decoder that reads the signal or the computer itself can read it which i think its the first to be honestly but the photonic is promising concept but dont expect pcs out of the world sure there will be decent speeds we already have 6ghz on new gens cpu with 10+ cores and in reality the gains will be diminishing if for ex you play games you might just squeeze a bit more performance but thats it unless they make games with the most realistic graphics to look like real life which i think we have some decades to achieve such feat and at the end its also depend on gpus will they be made with the silicon photon ? and if yes when
@tony_T_
@tony_T_ Год назад
@@HeLrAiSiNg1 Had a really hard time understanding what you typed out. The point is that the speed of light, being about 100x faster than electrons, will counteract the size increase of transistors and gates with its sheer speed. This would give Moore's law a good amount more headroom to keep progressing.
@dinozaurpickupline4221
@dinozaurpickupline4221 Год назад
@@HeLrAiSiNg1 I ALWAYS THOUGHT THERE WAS MORE TO GHZ STORY,CANT THE CPU MADE TO THINK ITS OPERATING ON HIGHER FREQUENCIES?
@HeLrAiSiNg1
@HeLrAiSiNg1 Год назад
@@tony_T_ dream on what i wrote were about light transmision exists for decades unless as the original post states keyboards with light transmision the logical thing is the usb is a decoder and the keyboard an encoder that transmits light with the information of what is pressed anyway it sure will bring some improvements but nothing super super wow that will drop jaws
@HeLrAiSiNg1
@HeLrAiSiNg1 Год назад
@@dinozaurpickupline4221 nope it cant ? You see a cpu has trillions of transistors gates etc the more small they make em to cramp more transistors everytime the less electricity they can stand if you see older cpus were able to hit 8ghz with nitro but newer is yet to hit anywhere that cause they cant moores law thing is thats apply everywhere not only pc cause overclock on cpu is by increasing voltages on cpu to gain speed that heats it if you for example put a 12v 2A the fan will be fine as long its 12v stable while its amperage draw is only 0.2A versus 2A but if voltage goes 12.1v it will heat a bit and slowly burn the same goes to cpus no mater what you do the only option is to give it more voltage for speed which means it need better cooling there is no magic programm the best you can improve is buy a rx 6900 and flash an rx 6900xt bios if it exist it will make it work a bit better cpu doesnt have anything stored that makes it work the way it does
@mlytle0
@mlytle0 Год назад
Silicon carbide is used in power transitors, handles high temperatures and very high frequencies, well above silicon. It already has a supply chain, and looks like a candidate.
@SirBlot
@SirBlot Год назад
It expands and contracts a lot.
@CrashCarson14
@CrashCarson14 Год назад
What about gan
@thenoobgamer9225
@thenoobgamer9225 Год назад
This video is more informativ than I thought. I love the web for channels like this.
@luketurnbull5648
@luketurnbull5648 Год назад
I just saw this is a dub channel of another. I didn’t know I was looking for content like this and I love it. I’ve only seen dub channels go to other languages from English so it’s really cool to see that it actually does work!
@chrisglaze658
@chrisglaze658 Год назад
Really good run down of the limits of silicon. Great research and presentation!
@Polar_Onyx
@Polar_Onyx Год назад
great research? most of the information given was a vast oversimplification or just outright wrong. I wish there wasn't so much dishonesty in tech channels. Being able to accurately depict advanced technology is difficult and it's much easier to make it sound fantastical and get more viewer retention
@thebogsofmordor7356
@thebogsofmordor7356 Год назад
Ayyy I am your 1000th subscriber! Great video
@shipsey2
@shipsey2 Год назад
Great video I found today I don't know if it's just me but the video and audio was out and not synced
@vadulme8640
@vadulme8640 Год назад
It's actually a dubbed video and we didn't have the orignal voice track without background music baked in, so yeah, doesn't look particularly stunning. We'll fix that with new vids. I mean the really new vids. Those that are already produced, will have to be dubbed the same way unfortunately.
@patelhitesh-ux2uf
@patelhitesh-ux2uf Год назад
Very informative and explained in very interesting way. Thank you.
@devbites77
@devbites77 Год назад
Thanks. I enjoyed your fascinating journey into the chip realm, exploring the potential future.
@Fernando-wz6no
@Fernando-wz6no Год назад
Quantum computer will *NEVER* happen. Forget that *FANTASY* !
@xxamulyaxx
@xxamulyaxx Год назад
RU-vid recommendations at its finest. Keep up.
@khlorghaal
@khlorghaal Год назад
"that is, if it continues to exist for us, in this line of events" that hit hard
@florenciabruck
@florenciabruck Год назад
Great information and you are very carismatic you can give tones of information without feeling overwelming.
@lamebubblesflysohigh
@lamebubblesflysohigh Год назад
The biggest problem with Quantum computer is the lack of software. Writing code for them is so complicated it practically eliminates 99% if not more people who currently make their living by writing code. That means less developers and thus less development. Unless coding will become heavily AI assisted.
@timothywilliams8530
@timothywilliams8530 Год назад
I'd say the biggest problem with them is that their cooling systems are the size of a room but, ya know.
@amentco8445
@amentco8445 Год назад
@@timothywilliams8530 Seeing how much of a mess it's been just to get quantum computers to the point they are now, I have doubts about all their pros in general.
@lamebubblesflysohigh
@lamebubblesflysohigh Год назад
@@timothywilliams8530 My mom worked with a computer that took an entire room before I was born. Her company actually had 2 of those. Size, price, power consumption do not matter if it has a purpose... a profitable use. Current quantum computer have no use whatsoever... not because they are bad but because there is no software that would make them better than standard computers for them. And there is no software because there is only a handful of people who can write some and they they probably call each other by their given names ;-)
@BadgerGamePlay
@BadgerGamePlay Год назад
only ~800 subs oO such high production, nice video man cant wait to see more!
@human_shaped
@human_shaped Год назад
Nice job researching and pulling all these alternatives together.
@robertburton432
@robertburton432 Год назад
Keep it simple >~
@RWBHere
@RWBHere Год назад
4:56 Not quite true. I have some Germanium transistors here which were used at 10.7 MHz, and some other low power ones which were used at over 100 MHz. They were expensive, but very capable at low currents. Bipolar Silicon transistors have been made which would amplify at frequencies close to 20GHz (I made some of them, back in the 1980's.) For higher frequencies, JFET, then MOSFET, IGFET and other types are needed. But yes; Germanium transistors are definitely very limited at higher frequencies, in comparison to Silicon devices. Leakage currents are very problematic, and they lead to high noise levels. It would still be interesting to know how a Germanium MOSFET would perform at somewhat higher frequencies, however.
@kykykyykykyk2851
@kykykyykykyk2851 Год назад
finally yt algorithm doing his job well
@Notfound4747
@Notfound4747 Год назад
hey i realy liked the video nicely balanced and no stale moments, i got the random recomendation video in auto play and didn't regret a BIT XD
@ryox82
@ryox82 Год назад
What I like about quantum computing is it could eventually be used to work the problems we face much faster, like disease cures, and silicon limitations.
@halrichard1969
@halrichard1969 Год назад
No one wants to cure disease. That should be obvious. If even half a dozen major diseases were cured, it would result in Trillions of dollars in losses for the Medical profession, big Pharma and medical hardware industries. They only want to "treat" diseases because that is where the profits are.
@richardnorbron4969
@richardnorbron4969 Год назад
Great video. Thanks for posting.
@John-vw4qz
@John-vw4qz Год назад
Thanks sir. I never knew this information... Amazing developments..
@stormevans6897
@stormevans6897 Год назад
First time I've ever seen a channel translated like this, cool
@invertexyz
@invertexyz Год назад
The next frontier will likely be 3D chips. We already do build them with a few layers, especially memory chips which are dozens of cell layers thick now. We will need to do the same with processors, and also switch to more thermally efficient materials to help avoid the increased heat per sqr unit as a processor gets thicker (building heat exchange tubes into the design itself can also drastically help combat this). A 3D design also opens up pathways for much more optimized computation methods that a chiplet design does not really lend itself well to as it scales. Chiplets primarily communicate to eachother through side channels, whereas a 3D chip would simply be one single complex chip of interconnected logic in all directions. The potential performance gains as we add layers is insane to think about, when you consider how thin a given layer of a processor is. You could have hundreds of thousands of layers eventually. A single processor rivaling billion dollar super computer warehouses of today.
@karotgamin790
@karotgamin790 Год назад
da cube chip
@dylanhecker6686
@dylanhecker6686 Год назад
Would the AMD 5800X3D be considered a 3D chip?
@invertexyz
@invertexyz Год назад
@@dylanhecker6686 Not really, it's only the cache, which is essentially on-die RAM that is stacked, which was already being done for your SYSRAM too. The core logic is still single layer.
@dylanhecker6686
@dylanhecker6686 Год назад
@@invertexyz thanks!
@Pythoner
@Pythoner 11 месяцев назад
Well we already do have '3D chips'. A processor consists of many layers of circuits stacked on top of each other.
@MonkeyManMechanicals
@MonkeyManMechanicals Год назад
Thank you. That was very informative with a thought processing progression that was spot on without someone knocking on your door. I've been imagining crystal tech ever since I learned about piezoelectric when I was 5. Magnetic cooling, heating and generators are my insanity. It's all about switches and ball bearings. Space and Storage. I still don't know why they just didn't cut the 0 into an 8??? Cheers
@JamesHill-vs4kn
@JamesHill-vs4kn Год назад
Thank you! We support you.
@onlinetech8077
@onlinetech8077 Год назад
Superb man👍💐👌♥️ #Mycomputer
@pneumantic6297
@pneumantic6297 Год назад
I don't see quantum processors replacing processors but if they become cheap enough I can definitely see them becoming a new component of the computer. If they do replace anything it will probably be graphics cards. All in all, we probably won't see quantum computers hit mass market until probably decades. What is more likely to happen than most theories is the integration of supercomputers, powered by more rare metals, that use the internet to give you your computer as a service rather than owning a computer. The system we have will be a lot more stupid and only decode the sent information.
@nyalan8385
@nyalan8385 Год назад
Yeah I feel like cloud computing will be big, but hopefully the next big "breakthrough" will be just moving away from the x86 architecture. This should buy us a couple more decades at least before we have to start replacing silicon, or whatever other new innovation occurs during that time
@Jack-he8jv
@Jack-he8jv Год назад
god i hope not, i hate this rent mentality. fuking WEF kikes will better write their wills if they try this.
@meppeorga
@meppeorga Год назад
We still got some ways to go with classical architecture, you can also gain some performance by improving how quickly we can access memory and how much of that fast memory we got, there are also architectural adjustments with concepts like RISC, bigLITTLE, vcash, new schedulers in OSs etc. so both on hardware and software level. With many new technologies being developed like Quantum, DNA-computing and Optical-computing, I don't think these will replace classic computer but with interconnects between them then they can work in tandem in the tasks they are best at.
@KingSteven77
@KingSteven77 Год назад
Good video but the video and sound is not synced and it's annoying but other than that it's 10/10
@jmr1920
@jmr1920 Год назад
good content. i really enjoyed it up to the end
@tombouie
@tombouie Год назад
Well-Done, summary: SILICON: Above vs below a 5nm gate width silicon is discrete on eor off vs statisticial on&off (aka tunnel field effect) GERMANIUM: Germanium Has much better lab performance but silicon is extraordinarily more pragmatic/practical (availability/cost, heat dissipation/tolerance, oxidation, freq band, etc). However germanium might be modified to improve its characteristics (ex: Molybdenite in development, lab germanane, etc). CARBON nano-tubes (in development): Graphene is a one-atom-thick layer of carbon atoms arranged in a hexagonal lattice. A carbon nano-tube is a tube of graphene GALLIUM NITRIDE: has some better performance than silicon & can be manufactured with silicon based equipment/industry Better CMOS: design based on statisticial on&off (aka tunnel field effect) instead of discrete on eor off to lower power/heat. Only work for graphene & at super low temps. MEMRISTOR: A memristor is an electrical component that limits or regulates the flow of electrical current in a circuit and remembers the amount of charge that has previously flowed through it. Memristors are important because they are non-volatile, meaning that they retain memory without power. OPTICAL COMPUTING: QUANTUM COMPUTING: In theory quantum computing can find least-worst solutions to problems no-matter the number of potential candidates (ex: NP problems like the traveling salement, decryption/password breaking, etc)
@Luizfernando-dm2rf
@Luizfernando-dm2rf Год назад
I like how you said nothing about optical computing xD
@tombouie
@tombouie Год назад
@@Luizfernando-dm2rf *I wasn't trying to show-off. These was just quick notes to remind me of technology tends I was fuzzy-on (thks to this good video). I'm a retired physicist & am already familar with optics & optical computing. *There is Trillion$ ... Trillion$ of semiconductor electronic infrastructure around the-world & comparatively optic/quantum infrastructures are very small. Quantum computer infrastructure is small but growing much faster than optic computing infrastructure. So we just might jump from semiconductor dominance to quantum dominance computing.
@Mew178
@Mew178 Год назад
Good new channel subbed. Well done.
@theosib
@theosib Год назад
Years ago, I did a calculation to determine when the speed and power benefits of process scaling would no longer offset the process variation. I got 5nm for that number as well.
@gabriel2fan
@gabriel2fan Год назад
Goood video man, best wishes from Switzerland
@avetruetocaesar3463
@avetruetocaesar3463 Год назад
I wholeheartedly disagree with the concluding section of this video. It insinuates that quantum-cloud-compute backend reliant hardware, which are about as functional as a storage device with some added connectivity, but mostly hollow toys which will serve as GUIs is the future of computers. I reject such a plane of existence, even if there is simplicity, or even necessity, in it; such a paradigm shift would basically mark the end of personal computers and ownership of anything related to computers. Quantum computers may as well exist in their own space and continue to get more sophisticated; they will still require close-to-0-Kelvin temperatures to even function, so an average person cannot carry one such device, let alone drive one in a home setting in today's society. I cannot visualise a future in which these devices become so compact, efficient, easy-to-manage and affordable as to be the de facto standard for computers because somehow, for some godforsaken reason, they are the only way going forward. How dreadful.
@Luizfernando-dm2rf
@Luizfernando-dm2rf Год назад
Ignoring the severe limitations Q-computing has as of today, the prospect of not being able to own a full computer is actual garbage. I'd rather stay with my classical slow piece of junk than having to trust big techs and cloud providers.
@antonnym214
@antonnym214 Год назад
Very nice reporting. I subscribed immediately. Thank you!
@qm3ster
@qm3ster Год назад
Dear Ivan. At 9:05, there's one frame where the cmos construction steps are shrinking, and the diagram there is not translated from Russian. Please enjoy it!
@WahteverYouWant
@WahteverYouWant Год назад
Dam man you need to give credit to Eugene Khutoryanski for using his graphics
@Daojyn
@Daojyn Год назад
Very good video, but I have two questions: -I wonder if it took around a half a century to reach this performance on silicon/CMOS doesn't it mean that it would take around the same time for others technologies to catch up or do improvements already discovered accelerate the rest? -After the huge rise of chip price in the last years can we expect before others technologies are catching up, let's say in the next decades that chip prices will decrease a lot because of the silicon limit that meanwhile we will see a comeback of dual CPU and SLI/crossfire for PC to keep improving performances?
@Jack-he8jv
@Jack-he8jv Год назад
amd chiplet design and intel vastly inferior big,little design is already a form of dual cpus. sli is very unlikely to ever come back due to how frustrating it was for everyone. for gpu its most likely gonna be software improvements.
@joefish6091
@joefish6091 Год назад
RISC-V is the future.
@samirkumargupta3502
@samirkumargupta3502 Год назад
Awesome video, thanks.from India.
@francisdebriey3609
@francisdebriey3609 Год назад
Very good ... subscribed !
@larrysal8866
@larrysal8866 Год назад
Can't wait to get my own silicone chip manufacturing system for cheap in a few decades B)
@diegoochoa572
@diegoochoa572 Год назад
Bro amazing content!
@GrandNecro
@GrandNecro Год назад
i think just coming up with better architechture and firmware is the easier route for now than finding a replacement for silicon
@lucifermorningstar8051
@lucifermorningstar8051 Год назад
amazing video. made me subscribe. ill watch more
@ebindanjan
@ebindanjan Год назад
Very good and interesting content. Thank you
@SarcastSempervirens
@SarcastSempervirens Год назад
great video, keep it up!
@AzErLoc
@AzErLoc Год назад
Silicon is also relatively easy to work with and process.
@leonel9110
@leonel9110 Год назад
Exactly and they’re working on 3nm already
@GuigEspritDuSage
@GuigEspritDuSage Год назад
According to some scientists, quantic computer may never become real fully capable computers and instead became at best some kind of accelerator or limited to very specific tasks.
@Wbfuhn
@Wbfuhn Год назад
Look at it this way. Human DNA is 2nm. To achieve human DNA size would be not only astounding but also open the potential for micro robots. At some point all manufacturers will have no choice but to develop more cores to handle more tasks. In addition to this, it might be possible to develop cybernetics and simulate human nerves.
@ouch1223
@ouch1223 Год назад
as an old viewer of yours, i'd like to wish u luck with ur new channel :D
@aaronsmith593
@aaronsmith593 8 месяцев назад
Silicone has so many uses. Pretty soon humans will completely turn into their favorite element.
@19vangogh94
@19vangogh94 Год назад
Silicon era will continue for at least another century m8, anything else will most likely be 10x-1000x more expensive and used in niche applications.
@paulssnfuture2752
@paulssnfuture2752 Год назад
yeah, until any tech becomes consumer grade and priced like silicon/lower it won't be viable replacement anytime within the century
@cj09beira
@cj09beira Год назад
@@paulssnfuture2752 which is very hard to do considering right now the base material is simply sand, and we got extremely good at making perfect crystals of Si, good luck beating either of those things. i think we will just have more and more silicon around the cpu ( more cache, integrated ram, accelerators etc) then we might get a cpu where one of the more important bits is a different faster material and the rest stays silicon
@redstefan6515
@redstefan6515 Год назад
very good video, this truly is a hidden gem
@JakubHohn
@JakubHohn Год назад
Thank you, I was lacking any news about memristors for about 8 years, so I stopped looking for updates. It is great to hear there is finally some progress
@mnomadvfx
@mnomadvfx Год назад
The main problem with memristors is the success of and investment in NAND flash memory. The advent of 3D multi layer NAND devices allowed an otherwise dying technology a new lease on life in the middle of the last decade, and it is still pushing forward. Even though ReRAM / memristors would be far superior in power consumption, latency and speed they are still leagues behind NAND flash in density even at the 2D level, and at the 3D level it's not even worth mentioning. Weebit is making a headway, but it will be years longer before we see anything truly commercial come out of it.
@xp.949
@xp.949 Год назад
What version of knock knock on heavens door it is bro? Thnks
@claudiohase296
@claudiohase296 Год назад
Very GOOD !!!!
@tombouie
@tombouie Год назад
Thks again & request you update your most excellent video once a ~quarter.
@alexvincheu
@alexvincheu Год назад
This thinking is what we humans do 🤔 and how we achieve it! Just by solving problems one by one. Then we wait for another or more problems and we solve them again, and again. That's the beauty in science and what our world represents.
@ma2i485
@ma2i485 Год назад
Great video bro
@AlephCasara
@AlephCasara Год назад
Excelent channel, you've got a brazilian subscriber here
@mka2
@mka2 Год назад
Olá! Obrigada!
@bakedbeings
@bakedbeings 4 месяца назад
There's good news coming out of georgia tech this week about silicon epigraphene transistors, compatible with current manufacturing methods 👍
@ZAcharyIndy
@ZAcharyIndy Год назад
Quite informative
@infinnite4938
@infinnite4938 Год назад
Nice video! I thought you would have 90k subs at least, but only 900!
@gmvpmb
@gmvpmb Год назад
Loved the information 🐱
@MommeSherif
@MommeSherif Год назад
Amazing content 🥰
@Kie-7077
@Kie-7077 Год назад
Cool, but the audio is badly skewed relative to the picture in this video.
@feey1443
@feey1443 Год назад
Thanks for the info
@ProDigit80
@ProDigit80 Год назад
Overheating becomes less of an issue with the reduction of transistor size. The only reason modern transistors are emanating more heat than older models, is because manufacturers make use of that reduction in heat output, by increasing cores, core frequency, and core complexity (more transistors per core).
@cj09beira
@cj09beira Год назад
that's only happening because the transistors no longer use less power when Shrank at the same rate of the volume reduction which used to be the case now a 2x volume reduction only leads to 20-30% power improvement, which is not enough,
@radomiami
@radomiami Год назад
If I were to guess which one would become the first step in a post-silicon CMOS world, it would probably be a room temperature tunnel field effect transistor.
@SupaKoopaTroopa64
@SupaKoopaTroopa64 Год назад
I'd put my money on silicon/photonic hybrid processors. There are already ASICs out there using photonic processing which are hundreds of times more energy efficient than silicon equivalents. We already have the technology, it's just a matter of integrating them into a full CPU/GPU.
@absolute___zero
@absolute___zero Год назад
​@@SupaKoopaTroopa64 photons are mass-less packets of energy and thus it is very difficult to make them interact between themselves and with any piece of matter we are made or our chips are made of. that's the reason why we still use CMOS transistors with gate delay of 30 picoseconds instead of using light and get attosecond "gate delays" with "light transistors", if they would exist. Everybody has it very clear, creating a chip purely using light will make it millions of times faster than silicon chips. The reason why light is so fast, it is because it has no mass. But you can't make a flip-flop out of photons. That is why accelerating CMOS chips with analog-style logic (implemented using interference, like some startups are doing right now) is a bad idea. It is like putting a F1 engine on a bicycle and expecting it is going to do speeds of 400 km/hour. The bearings of the bicycle wheels will simply melt because they aren't designed for such high RPM (not talking even about aerodynamics here). It would be a good idea for a college student working in his garage, but not for a company that rises 100 million or so for this kind of project. I suggest you to learn quantum physics and chip-design before you invest. The rate of bankrupcy of new companies is like 9 to 10, if not more. Only understanding the physics of the computation you can make a fair judgement of which company is going right direction or not. There are always tons of options to invest but only few of them are real opportunities.
@SupaKoopaTroopa64
@SupaKoopaTroopa64 Год назад
@@absolute___zero I'm aware of these limitations. I'm just saying that production-ready photonics already exist, so they have a head start over many other technologies. Also, I don't plan on investing in any photonics startups, or even anything in the microprocessor industry right now, I was just saying that If I had to make a bet on which of these technologies would first appear in a consumer product, I'd go with photonics.
@absolute___zero
@absolute___zero Год назад
@@SupaKoopaTroopa64 photonics of course has future, but not as co-processors, rather a standalone computing logic with only a few CMOS. But there are many other computing methods that have been developed by scientists but not exploited yet. For example, there are mechanical integrated circuits made of mechanical relays (they achieve 10 nanosecond times to switch on/off), or there are gold transistors working in vacuum like vacuum tubes, there are DNA computers, and so on... the next big idea will be a new computing paradigm. There is also lots of work to do in software to gain performance, for example, development of massively parallel operating system.
@kwgm8578
@kwgm8578 4 месяца назад
Even if Moore's law may not apply to Si applications, they are still very useful in solving many of our computational applications at 16 and even 32 nm technologies.
@squigglesmcjr199
@squigglesmcjr199 Год назад
Glad i found this channel premium content
@MusingsFromTheJohn00
@MusingsFromTheJohn00 Год назад
Quantum computing can't replace existing systems, but it can be an add on for some special types of problems. What was really missed here, in my opinion, is (1) true 3D chips, like a solid 1" cube, and (2) wetware which uses living cells, DNA, RNA, and proteins.
@NineSeptims
@NineSeptims Год назад
And what kind of processing speeds will dna have? Surely a downgrade.
@MusingsFromTheJohn00
@MusingsFromTheJohn00 Год назад
@@NineSeptims the human brain operates on two general layers, an upper layer of neurological swarm intelligence that is complexly interwoven with the lower layer of DNA/RNA swarm intelligence. We do not know enough to give an accurate measure of how intelligent a human brain but we do know that it is still vastly more powerful than our most powerful supercomputer, the Frontier, which can perform 1,102,000,000,000,000,000 to 1,685,000,000,000,000,000 floating point operation per second, and uses about 21,000,000 watts of power. The human brain uses about 20 watts of power. Then there is the physical size difference, here is what Frontier looks like: en.wikipedia.org/wiki/Frontier_(supercomputer) So, while Frontier can do things you can't, because it is specialized for that, Frontier still is not capable of matching the computational power of your brain. Now, consider the power of a cybernetic brain which has all the power of at least one human brain and the power of a silicon based system more powerful than Frontier and some specialized quantum computing addons for those special problems where quantum computing works. That will be a true Artificial General Super Intelligence with Personality (AGSIP) system. Probably will not have that for a few decades or so.
@twinkytwinklier4047
@twinkytwinklier4047 Год назад
@@NineSeptims I don’t think so. A DNA contais about a million GB, so imagine having a pot full of them, mix some protiens and some fancy ribosomes, with the right temperature, you can essentially process tons of information IN PARALLEL. I don’t know if this is how wetware is gunna look like, but it’s just a guess. I think the issue would be temperature control. Imagine when you forget a vile of DNA in your hot car and it turns to soup, or when the synthesis of some protiens are a little off and you just acidentally destroy your expensive batch. This gives a new meaning to “computer virus” …and also imagine running a neural network on an actual neural network… damn I wanna see
@twinkytwinklier4047
@twinkytwinklier4047 Год назад
@@NineSeptims Add to that, when these protiens break down RNAs, it looks just like a turing machine. Imagine programming a ribosome as a state machine, it’ll be turing complete!
@Sciptopia
@Sciptopia Год назад
@@twinkytwinklier4047 a neural network on an actual neural network, we're in a simulation!
@ZMacZ
@ZMacZ 4 месяца назад
When it comes to silicon being a pillar of computing, that will never really go away. Why ? Mass computing. Basically, when wanting to do computing en masse, given that power usage is not a problem with enough silicon-based photovoltaics in mind, computing can still use mostly silicon-based computing for when computing something that requires so much computing power that using the more scarce resource based computing options become prohibitively expensive. It's like creating a very large computer that can be like a workhorse, while using the more scarce resource based computing option for small applications. The workhorse can be used as an off-site computing 'monster' to offload work on. This may seem rather inefficient, but given that electricity can easily be farmed with solar voltaics, the power requirements are a lesser drawback, while the mass computing becomes cheaper over time, and can yet benefit from increased performance. Once people go into space and build space stations, remote computing this way can be much cheaper than supplying the more scarce resource based computing options for all/everyone. The problem is not one about creating a few million small computing options that require scarce resources (and thus becoming expensive), but once the newer options need to be created by the billions, yearly. The amount of mining required to get enough of the scarcer resource would be bad for the environment in such a fashion that humanity would 'compute' itself to death. Since, computing to death does not compute, en masse computing with the cheaper silicon option makes more sense. Also notice that en masse computing is much more efficient. When a resource isn't used by someone, someone else may be allocated more computing resource aiding his or her task to complete more rapidly. When it comes to remote computing for things like gaming, this may also work well. Silicon will always stay the best and easiest resource for computing, since it's most abundant. Also, the new computing options will have their own size limitations, albeit somewhat smaller. This would be a fixed factor. 4:28 Yes, the cost of fabrication of non silicon processors is much higher due to the scarcity of the required materials. This will never change. Also, due to the limitation of availability the newer processors would be fabricated in much smaller quantity, while being much more expensive. Due to rising cost of scarce materials the cost over time would rise, not become lower. For regular silicon-based processors, these would become cheaper over time. Combined with photovoltaics the increased power usage would not be the limiting factor. 13:56 Photonics are best used for transfer of data, not computing. Extreme short relays, let's say between chip layers and external to chip pathways would be an option, since these can transfer data really fast, and yet use very little power, which results in less heat release. Photons travel light so to speak while electron travel is heavy. This means less heat release. There's one problem though, the creation of very small light emitters/receivers and very small fiber optics or such is much harder than a regular circuit. It's at this point in time at least not very useful. A similar thing is with the wattage per computing. If the wattage becomes low enough, further reduction of such becomes less useful, and with that only extreme high end computing purposes would be served by further reduction. Let's examine an example, where you can have a fully capable processor using a 10 cm by 10 cm solar panel for continuous usage. What would the difference make if it's then operable by an 8 x 8 cm solar panel instead, by increasing the cost tenfold ? It would mean less power used, yes, by as much as 33%, but then the tenfold cost ? So, at some point in time the resource cost will define the computing solution, rather than its capability, once again referring to the scarcity of materials. 14:46 Photonics will not remove the limitations of distance. Any signal will still be limited to the speed of light, whether electronic or photonic based. Latency stays the same no matter which of these mediums are used. Basically, the best way to make processors smaller and faster, currently, is to provide on-chip memory, with comparatively slow memory being used for the externals. On chip memory would increase processing by removing much of the latency of memory instructions, like cache does, but then used on a larger scale, with only one cache and larger memory area. Then you'd have 16 GB base memory on the chip, with like 4-8 megs in cache. When a large-scale external reference would occur (like writing to SSD) the memory controller would copy the on-chip memory to the external one, subsequently transferring it to SSD, while optimizing the transfer speed, between on-chip and external memory. On chip the memory could have the same latency as regular cache, but transfer between the on-chip memory and external (DDR) memory would still be many times greater than the transfer speed between regular memory and SSD. Also, when using stuff like virtual memory on an SSD, this could then be replaced by the external (DDR) memory, increasing computational output per unit by that much as well. Knowing that most of the CPU's and GPU's time is wasted on waiting for memory return, you know that this transfer of operations from external memory to on-chip memory can make a big difference, even when the computational output of the processor is yet smaller. Solving each bottleneck in turn is also a way to speed up computing, in some cases making a bigger difference than making it smaller.
@projectw.a.a.p.f.t.a.d7762
@projectw.a.a.p.f.t.a.d7762 Год назад
I was watching how atom's are being controlled and manipulated on 2d materials. That would suggest we have a ways to go, regarding Moore's Laws future.
@roboticroom7283
@roboticroom7283 Год назад
GE is used in military industry that's why. + Gallium arsenide is used in military because chips made from that materials has lowest fail rate and extended temperature limits.
@olagarto1917
@olagarto1917 Год назад
Photonics can also do prossessing via constructive/distructive wave interference
@agsystems8220
@agsystems8220 Год назад
Not really, as it messes with phase information. You can build simple gates, but as soon as you try to do something like an and gate either your 1+1 value is different amplitude from 1+0, or its phase is different. This is fine if you are just throwing it at a detector, but if you try to throw it into another gate built around interference it is not going to work right. There might be some quantum mechanical way to align the phases, but the naive approach does not work.
@SC-zq6cu
@SC-zq6cu Год назад
Photonics cannot store memory yet as that would imply being able to trap light for very long periods of time which is still impossible.
@JorgetePanete
@JorgetePanete Год назад
processing*
@aaronsmith593
@aaronsmith593 8 месяцев назад
How does one remember which position all those switches are in ?!? Either On or Off. Well that is what computer glitches are, the chip is forgetting where, what, who and when the switches are doing at any given time. Software errors can multiply these error glitches.😮😊
@ERRORFACE1
@ERRORFACE1 Год назад
thanks for the vid. what's the outro song thanks
@kaiperdaens7670
@kaiperdaens7670 5 месяцев назад
During the video I was thinking what if we make a superconductor chip?
@gab882
@gab882 Год назад
Is it possible to combine all the tech using graphene carbon nanotubes transmitting light instead of electrons in memristors/neuronmorphic architecture?
@ashishchauhan2759
@ashishchauhan2759 Месяц назад
Legendary content 😮
@nathanwest2304
@nathanwest2304 Год назад
I really think that graphene is the right way to move forward, we don't know how or when, but ultimately they will be the solution
@mr.electronx9036
@mr.electronx9036 Год назад
I think the "next thing" will be photonic cpu or/and ai neuromorphic chiplet, which both are extryemly fast. The issue is still a silicon as a substrate. I heard that specific "glass" can be used even more better than silicon
@DespaceMan
@DespaceMan Год назад
Interesting, keep an eye out for Biphenylene I believe this will have a wider application field than graphene.
@DeTrOiTXX12
@DeTrOiTXX12 Год назад
I really enjoyed this video!
@shows1234
@shows1234 Год назад
Hi there, it seems that the audio and video aren't synched.
@antonnym214
@antonnym214 Год назад
Check into computing with Ternary. It is mathematically more efficient than binary and requires just one extra voltage level. This could lead to more efficient and faster CPUs than we have today and some are already being built. All good wishes.
@nadergadelrab2669
@nadergadelrab2669 Год назад
amazing
@DuckyThePilot
@DuckyThePilot Год назад
your audio isnt lined up with the video.
Далее
This New Photonic Chip Computes in Femtoseconds
18:14
Просмотров 185 тыс.
Analog Computing is GENIUS - Here's Why!
15:28
Просмотров 464 тыс.
Interstellar Expansion WITHOUT Faster Than Light Travel
21:14
How Apple Just Changed the Entire Industry (M1 Chip)
26:28
How are Microchips Made?
27:48
Просмотров 121 тыс.
Why making chips is so hard
16:03
Просмотров 899 тыс.
A New Way to Achieve Nuclear Fusion: Helion
30:48
Просмотров 7 млн
The Most Powerful Computers You've Never Heard Of
20:13
Best Gun Stock for VR gaming. #vr #vrgaming  #glistco
0:15