End of the silicon era. Processors of the future

  Переглядів 311,917

My Computer

My Computer

День тому

The era of silicon chips is coming to an end. New processors come out hot, and everyone forgot about Moore's law. Will the development of electronics stop? What will be the processors of the future? Is there a replacement for silicon?
00:00 - The purest polysilicon
01:26 - Silicon limit
03:32 - What if not silicon?
08:50 - Rejection of CMOS
13:29 - Changing electrons to photons
14:53 - Quantum computer
#mycomputer #processor #silicon

КОМЕНТАРІ: 796
@blinded6502
@blinded6502 Рік тому
Quantum computers are NOT a replacement for the classical computers, holy hell
@DocWolph
@DocWolph Рік тому
A Classic Computer (C-Comp) becomes the I/O device to talk to the Q-Comp. Input-> C-comp-> Q-comp-> (back to)C-Comp-> Output
@cebo494
@cebo494 Рік тому
​@@DocWolph It doesn't matter what the connection is, quantum computers simply are not meant to do the same things that classical computers do. That's what the guy meant by "are not a replacement". They are not better than classical computers for most of the tasks that the vast majority of people use computers for, like web browsing, text/image processing, gaming, etc. They have little to no values for consumers. They will likely forever remain in labs, datacenters, and industry for highly specialized tasks and research.
@fridolinkoch
@fridolinkoch Рік тому
Until someone discovers something you can do with them as a general person….just like when the internet arrived
@Galacticfungus
@Galacticfungus Рік тому
@@fridolinkoch There are a few Quantum ray tracing algorithms with papers now.
@DocWolph
@DocWolph Рік тому
@@cebo494 Well at this point it is implied, at least, that you can not directly interface with a Q-Comp. That is, you need some kind of bridging software running from the C-Comp to the Q-Comp and back. This is regardless of application. Aside from scientific, engineering, security work, among other things, I can see Q-Comps being used for animation, rendering, or simulation (for example, I can see water sims taking only a few minutes rather than many hours or days for a FEW high quality frames with a Q-Comp) and that is not even the color of the tip of the iceberg of what is possible with Q-Comps. For most things people use Computers for, C-Comps will be at least enough.THIS is agreed. But there are things that are way bigger that even best current C-comp technology just is not adequate for. and Q-Comps at a price a small studio, or a very dedicated hobbyist, can buy to radically accelerate their sim work, animation, and rendering work (albeit the software may be coming a few years after the fact) is just what the Studio head asked for. And that is JUST animation. Anybody can become a "Doc Brown". pursue science, mathematics, engineering and MORE at home. Again, I generally agree with what you saying BUT you are not thinking broadly enough.
@DunnickFayuro
@DunnickFayuro Рік тому
Very good content. I just find the desync between your video and sound a bit annoying.
@PareshPatel-xc2vu
@PareshPatel-xc2vu Рік тому
I think it's dubbed, probably from Russian
@DunnickFayuro
@DunnickFayuro Рік тому
@@PareshPatel-xc2vu Nope. Just desync. The lips almost match the sound, with a little delay.
@namthainam
@namthainam Рік тому
Nope video is all ai generate from stock video so it's all jibberish. Might as well listen with your eyes closed
@samlovebutter
@samlovebutter Рік тому
@@DunnickFayuro it is dubbed from russian. The original channel called "Мой Компьютер"
@vadulme8640
@vadulme8640 Рік тому
@@DunnickFayuro Yeah, it is dubbed. I am the guy who recorded the dub. I synced it the best I could to the original voice, but I could only do so much. And yes, I get it that it's annoying.
@cursedgamer2778
@cursedgamer2778 Рік тому
I love these dual language channels, I don't know Russian and would have never have been able to understand your phenomenal video otherwise. Thank you, truly.
@tankerock
@tankerock Рік тому
No wonder, I was wondering why my video/audio was out of sync lol
@ysa1023
@ysa1023 Рік тому
Ah so that's why his disgusting face seemed so familiar and audio was out of sync! He is one of the idiots that supports war in my country. He is one of the people that thinks I should die... Yeah, hope at least somebody is going to see this comment, and stop watching this genocide supporter.
@mka2
@mka2 Рік тому
Спасибо друг! Это очень приятно читать! Привет из холодной России!
@headSoup
@headSoup Рік тому
@@mka2 привет друг! я из Американский
@headSoup
@headSoup Рік тому
@@mka2 я говорю просто 😢
@ajbowers44
@ajbowers44 Рік тому
So glad UKposts recommended this video. Nicely done, great balance of information and presentation without coming off overly optimistic or pessimistic.
@Andrew-rc3vh
@Andrew-rc3vh Рік тому
Yes indeed, so far it sounds intelligent. I'm hoping I'm not going to be lectured about Moore's law. I'm convinced that is a marketing meme for Intel.
@hyll6700
@hyll6700 Рік тому
Quantum Computing isn't a replacement.
@AliensKillDevils.
@AliensKillDevils. Рік тому
Mt.Fuji is about to explode. Excess silicone should be mined. ukposts.info/slow/PLen6gjewxJdCB0VDrHJiylc3B3qh4b91m
@aqua-bery
@aqua-bery Рік тому
Quantum computers can't replace regular computers. Yes they could calculate some very complex equations at a fraction of the speed. But they suck at doing simple calculations.
@hyll6700
@hyll6700 Рік тому
@@Andrew-rc3vh Bruh 😂
@kevinmitchell3168
@kevinmitchell3168 Рік тому
You mentioned photonics use in communications but didn't mention photonics switching. They've been trying to make purely photonic chips for decades, and it's always just around the corner much like fusion. It's that research that led to photonics being integrated into silicon chips. It seems like that research was very close to yielding results but focus and funding got taken over by quantum computing. From all I've read purely photonic chips would run 100-1000 times faster than silicon at far lower power and heat. Hopefully as silicon reaches it's limitations there will be renewed funding and research for it.
@Luizfernando-dm2rf
@Luizfernando-dm2rf Рік тому
It's a bit sad really... Quantum computing while exciting, is very far from producing anything useful in terms of actual computing. There has been no more than 10 algorithims made to run on quantum computers for example.
@id104335409
@id104335409 Рік тому
The comment I was searching for. I was hoping the video is on that topic. Not quantum computing.
@bdjfw2681
@bdjfw2681 Рік тому
isn't fusion have focus and funding but still no result ? is it really that close for purely photonic chips ?
@dinozaurpickupline4221
@dinozaurpickupline4221 Рік тому
light switches have their application in networking & data transfer
@jadedandbitter
@jadedandbitter Рік тому
Intel's working on it. Have a buddy that works in the photonics dept. Can never talk about where they're at though.
@supriyochakrabortybme
@supriyochakrabortybme Рік тому
A very interesting compilation, the UKposts algorithm recommended this hidden gem.
@yamatsukami987
@yamatsukami987 Рік тому
also worth considering in the mean time, their are computers that can run on trits (-1, 0, 1) which requires some fundamental changes, but could theoretically be more effective. This could even be extended further though it gets less practical the more you add.
@tony_T_
@tony_T_ Рік тому
Silicon photonics is what I have the most hope in for in the next couple of decades. Imo, transitioning from electricity to light is just the most logical step forward. It will set moores law back by quite a bit, but the insane clock rate of the processors will make up for it. Most modern keyboards already use light to transmit signals.
@HeLrAiSiNg1
@HeLrAiSiNg1 Рік тому
well light transmision exists for decades now unless you mean signals like hid interface that uses speed bandwidth like in hz its known that for ex most audio devices ps5 xbox etc have a light slot on the back to connect with optical fiber to send sound now a kb uses a microcontroller if its mechanical that works like microcomputers with speeds of some mhz so either at the usb there is a decoder that reads the signal or the computer itself can read it which i think its the first to be honestly but the photonic is promising concept but dont expect pcs out of the world sure there will be decent speeds we already have 6ghz on new gens cpu with 10+ cores and in reality the gains will be diminishing if for ex you play games you might just squeeze a bit more performance but thats it unless they make games with the most realistic graphics to look like real life which i think we have some decades to achieve such feat and at the end its also depend on gpus will they be made with the silicon photon ? and if yes when
@tony_T_
@tony_T_ Рік тому
@@HeLrAiSiNg1 Had a really hard time understanding what you typed out. The point is that the speed of light, being about 100x faster than electrons, will counteract the size increase of transistors and gates with its sheer speed. This would give Moore's law a good amount more headroom to keep progressing.
@dinozaurpickupline4221
@dinozaurpickupline4221 Рік тому
@@HeLrAiSiNg1 I ALWAYS THOUGHT THERE WAS MORE TO GHZ STORY,CANT THE CPU MADE TO THINK ITS OPERATING ON HIGHER FREQUENCIES?
@HeLrAiSiNg1
@HeLrAiSiNg1 Рік тому
@@tony_T_ dream on what i wrote were about light transmision exists for decades unless as the original post states keyboards with light transmision the logical thing is the usb is a decoder and the keyboard an encoder that transmits light with the information of what is pressed anyway it sure will bring some improvements but nothing super super wow that will drop jaws
@HeLrAiSiNg1
@HeLrAiSiNg1 Рік тому
@@dinozaurpickupline4221 nope it cant ? You see a cpu has trillions of transistors gates etc the more small they make em to cramp more transistors everytime the less electricity they can stand if you see older cpus were able to hit 8ghz with nitro but newer is yet to hit anywhere that cause they cant moores law thing is thats apply everywhere not only pc cause overclock on cpu is by increasing voltages on cpu to gain speed that heats it if you for example put a 12v 2A the fan will be fine as long its 12v stable while its amperage draw is only 0.2A versus 2A but if voltage goes 12.1v it will heat a bit and slowly burn the same goes to cpus no mater what you do the only option is to give it more voltage for speed which means it need better cooling there is no magic programm the best you can improve is buy a rx 6900 and flash an rx 6900xt bios if it exist it will make it work a bit better cpu doesnt have anything stored that makes it work the way it does
@mlytle0
@mlytle0 Рік тому
Silicon carbide is used in power transitors, handles high temperatures and very high frequencies, well above silicon. It already has a supply chain, and looks like a candidate.
@SirBlot
@SirBlot Рік тому
It expands and contracts a lot.
@CrashCarson14
@CrashCarson14 Рік тому
What about gan
@josgraha
@josgraha Рік тому
oh boy, quantum computers are even further out than tfets, optical compute, or memristors and you still need a von-neuman silicon computer to work with it. thanks so much for the great video, fantastic questions and background material. much appreciated
@ararune3734
@ararune3734 Рік тому
Not only are they far away from any practical implementation, but they will simply never replace your standard computers in the way people think. Quantum computers will be better at performing certain tasks, but those tasks are not in our every day usage, so for your personal needs, you won't use a quantum computer, you'll stick to the standard computer.
@tombouie
@tombouie Рік тому
Well-Done, summary: SILICON: Above vs below a 5nm gate width silicon is discrete on eor off vs statisticial on&off (aka tunnel field effect) GERMANIUM: Germanium Has much better lab performance but silicon is extraordinarily more pragmatic/practical (availability/cost, heat dissipation/tolerance, oxidation, freq band, etc). However germanium might be modified to improve its characteristics (ex: Molybdenite in development, lab germanane, etc). CARBON nano-tubes (in development): Graphene is a one-atom-thick layer of carbon atoms arranged in a hexagonal lattice. A carbon nano-tube is a tube of graphene GALLIUM NITRIDE: has some better performance than silicon & can be manufactured with silicon based equipment/industry Better CMOS: design based on statisticial on&off (aka tunnel field effect) instead of discrete on eor off to lower power/heat. Only work for graphene & at super low temps. MEMRISTOR: A memristor is an electrical component that limits or regulates the flow of electrical current in a circuit and remembers the amount of charge that has previously flowed through it. Memristors are important because they are non-volatile, meaning that they retain memory without power. OPTICAL COMPUTING: QUANTUM COMPUTING: In theory quantum computing can find least-worst solutions to problems no-matter the number of potential candidates (ex: NP problems like the traveling salement, decryption/password breaking, etc)
@Luizfernando-dm2rf
@Luizfernando-dm2rf Рік тому
I like how you said nothing about optical computing xD
@tombouie
@tombouie Рік тому
@@Luizfernando-dm2rf *I wasn't trying to show-off. These was just quick notes to remind me of technology tends I was fuzzy-on (thks to this good video). I'm a retired physicist & am already familar with optics & optical computing. *There is Trillion$ ... Trillion$ of semiconductor electronic infrastructure around the-world & comparatively optic/quantum infrastructures are very small. Quantum computer infrastructure is small but growing much faster than optic computing infrastructure. So we just might jump from semiconductor dominance to quantum dominance computing.
@MusingsFromTheJohn00
@MusingsFromTheJohn00 Рік тому
Quantum computing can't replace existing systems, but it can be an add on for some special types of problems. What was really missed here, in my opinion, is (1) true 3D chips, like a solid 1" cube, and (2) wetware which uses living cells, DNA, RNA, and proteins.
@NineSeptims
@NineSeptims Рік тому
And what kind of processing speeds will dna have? Surely a downgrade.
@MusingsFromTheJohn00
@MusingsFromTheJohn00 Рік тому
@@NineSeptims the human brain operates on two general layers, an upper layer of neurological swarm intelligence that is complexly interwoven with the lower layer of DNA/RNA swarm intelligence. We do not know enough to give an accurate measure of how intelligent a human brain but we do know that it is still vastly more powerful than our most powerful supercomputer, the Frontier, which can perform 1,102,000,000,000,000,000 to 1,685,000,000,000,000,000 floating point operation per second, and uses about 21,000,000 watts of power. The human brain uses about 20 watts of power. Then there is the physical size difference, here is what Frontier looks like: en.wikipedia.org/wiki/Frontier_(supercomputer) So, while Frontier can do things you can't, because it is specialized for that, Frontier still is not capable of matching the computational power of your brain. Now, consider the power of a cybernetic brain which has all the power of at least one human brain and the power of a silicon based system more powerful than Frontier and some specialized quantum computing addons for those special problems where quantum computing works. That will be a true Artificial General Super Intelligence with Personality (AGSIP) system. Probably will not have that for a few decades or so.
@twinkytwinklier4047
@twinkytwinklier4047 Рік тому
@@NineSeptims I don’t think so. A DNA contais about a million GB, so imagine having a pot full of them, mix some protiens and some fancy ribosomes, with the right temperature, you can essentially process tons of information IN PARALLEL. I don’t know if this is how wetware is gunna look like, but it’s just a guess. I think the issue would be temperature control. Imagine when you forget a vile of DNA in your hot car and it turns to soup, or when the synthesis of some protiens are a little off and you just acidentally destroy your expensive batch. This gives a new meaning to “computer virus” …and also imagine running a neural network on an actual neural network… damn I wanna see
@twinkytwinklier4047
@twinkytwinklier4047 Рік тому
@@NineSeptims Add to that, when these protiens break down RNAs, it looks just like a turing machine. Imagine programming a ribosome as a state machine, it’ll be turing complete!
@Sciptopia
@Sciptopia Рік тому
@@twinkytwinklier4047 a neural network on an actual neural network, we're in a simulation!
@luketurnbull5648
@luketurnbull5648 Рік тому
I just saw this is a dub channel of another. I didn’t know I was looking for content like this and I love it. I’ve only seen dub channels go to other languages from English so it’s really cool to see that it actually does work!
@invertexyz
@invertexyz Рік тому
The next frontier will likely be 3D chips. We already do build them with a few layers, especially memory chips which are dozens of cell layers thick now. We will need to do the same with processors, and also switch to more thermally efficient materials to help avoid the increased heat per sqr unit as a processor gets thicker (building heat exchange tubes into the design itself can also drastically help combat this). A 3D design also opens up pathways for much more optimized computation methods that a chiplet design does not really lend itself well to as it scales. Chiplets primarily communicate to eachother through side channels, whereas a 3D chip would simply be one single complex chip of interconnected logic in all directions. The potential performance gains as we add layers is insane to think about, when you consider how thin a given layer of a processor is. You could have hundreds of thousands of layers eventually. A single processor rivaling billion dollar super computer warehouses of today.
@karotgamin790
@karotgamin790 Рік тому
da cube chip
@dylanhecker6686
@dylanhecker6686 Рік тому
Would the AMD 5800X3D be considered a 3D chip?
@invertexyz
@invertexyz Рік тому
@@dylanhecker6686 Not really, it's only the cache, which is essentially on-die RAM that is stacked, which was already being done for your SYSRAM too. The core logic is still single layer.
@dylanhecker6686
@dylanhecker6686 Рік тому
@@invertexyz thanks!
@Pythoner
@Pythoner 11 місяців тому
Well we already do have '3D chips'. A processor consists of many layers of circuits stacked on top of each other.
@chrisglaze658
@chrisglaze658 Рік тому
Really good run down of the limits of silicon. Great research and presentation!
@Polar_Onyx
@Polar_Onyx Рік тому
great research? most of the information given was a vast oversimplification or just outright wrong. I wish there wasn't so much dishonesty in tech channels. Being able to accurately depict advanced technology is difficult and it's much easier to make it sound fantastical and get more viewer retention
@thenoobgamer9225
@thenoobgamer9225 Рік тому
This video is more informativ than I thought. I love the web for channels like this.
@RWBHere
@RWBHere Рік тому
4:56 Not quite true. I have some Germanium transistors here which were used at 10.7 MHz, and some other low power ones which were used at over 100 MHz. They were expensive, but very capable at low currents. Bipolar Silicon transistors have been made which would amplify at frequencies close to 20GHz (I made some of them, back in the 1980's.) For higher frequencies, JFET, then MOSFET, IGFET and other types are needed. But yes; Germanium transistors are definitely very limited at higher frequencies, in comparison to Silicon devices. Leakage currents are very problematic, and they lead to high noise levels. It would still be interesting to know how a Germanium MOSFET would perform at somewhat higher frequencies, however.
@devbites77
@devbites77 Рік тому
Thanks. I enjoyed your fascinating journey into the chip realm, exploring the potential future.
@shipsey2
@shipsey2 Рік тому
Great video I found today I don't know if it's just me but the video and audio was out and not synced
@vadulme8640
@vadulme8640 Рік тому
It's actually a dubbed video and we didn't have the orignal voice track without background music baked in, so yeah, doesn't look particularly stunning. We'll fix that with new vids. I mean the really new vids. Those that are already produced, will have to be dubbed the same way unfortunately.
@meppeorga
@meppeorga Рік тому
We still got some ways to go with classical architecture, you can also gain some performance by improving how quickly we can access memory and how much of that fast memory we got, there are also architectural adjustments with concepts like RISC, bigLITTLE, vcash, new schedulers in OSs etc. so both on hardware and software level. With many new technologies being developed like Quantum, DNA-computing and Optical-computing, I don't think these will replace classic computer but with interconnects between them then they can work in tandem in the tasks they are best at.
@ryox82
@ryox82 Рік тому
What I like about quantum computing is it could eventually be used to work the problems we face much faster, like disease cures, and silicon limitations.
@halrichard1969
@halrichard1969 Рік тому
No one wants to cure disease. That should be obvious. If even half a dozen major diseases were cured, it would result in Trillions of dollars in losses for the Medical profession, big Pharma and medical hardware industries. They only want to "treat" diseases because that is where the profits are.
@MonkeyManMechanicals
@MonkeyManMechanicals Рік тому
Thank you. That was very informative with a thought processing progression that was spot on without someone knocking on your door. I've been imagining crystal tech ever since I learned about piezoelectric when I was 5. Magnetic cooling, heating and generators are my insanity. It's all about switches and ball bearings. Space and Storage. I still don't know why they just didn't cut the 0 into an 8??? Cheers
@patelhitesh-ux2uf
@patelhitesh-ux2uf Рік тому
Very informative and explained in very interesting way. Thank you.
@florenciabruck
@florenciabruck Рік тому
Great information and you are very carismatic you can give tones of information without feeling overwelming.
@human_shaped
@human_shaped Рік тому
Nice job researching and pulling all these alternatives together.
@robertburton432
@robertburton432 Рік тому
Keep it simple >~
@Notfound4747
@Notfound4747 Рік тому
hey i realy liked the video nicely balanced and no stale moments, i got the random recomendation video in auto play and didn't regret a BIT XD
@khlorghaal
@khlorghaal Рік тому
"that is, if it continues to exist for us, in this line of events" that hit hard
@John-vw4qz
@John-vw4qz Рік тому
Thanks sir. I never knew this information... Amazing developments..
@lamebubblesflysohigh
@lamebubblesflysohigh Рік тому
The biggest problem with Quantum computer is the lack of software. Writing code for them is so complicated it practically eliminates 99% if not more people who currently make their living by writing code. That means less developers and thus less development. Unless coding will become heavily AI assisted.
@timothywilliams8530
@timothywilliams8530 Рік тому
I'd say the biggest problem with them is that their cooling systems are the size of a room but, ya know.
@amentco8445
@amentco8445 Рік тому
@@timothywilliams8530 Seeing how much of a mess it's been just to get quantum computers to the point they are now, I have doubts about all their pros in general.
@lamebubblesflysohigh
@lamebubblesflysohigh Рік тому
@@timothywilliams8530 My mom worked with a computer that took an entire room before I was born. Her company actually had 2 of those. Size, price, power consumption do not matter if it has a purpose... a profitable use. Current quantum computer have no use whatsoever... not because they are bad but because there is no software that would make them better than standard computers for them. And there is no software because there is only a handful of people who can write some and they they probably call each other by their given names ;-)
@theosib
@theosib Рік тому
Years ago, I did a calculation to determine when the speed and power benefits of process scaling would no longer offset the process variation. I got 5nm for that number as well.
@pneumantic6297
@pneumantic6297 Рік тому
I don't see quantum processors replacing processors but if they become cheap enough I can definitely see them becoming a new component of the computer. If they do replace anything it will probably be graphics cards. All in all, we probably won't see quantum computers hit mass market until probably decades. What is more likely to happen than most theories is the integration of supercomputers, powered by more rare metals, that use the internet to give you your computer as a service rather than owning a computer. The system we have will be a lot more stupid and only decode the sent information.
@nyalan8385
@nyalan8385 Рік тому
Yeah I feel like cloud computing will be big, but hopefully the next big "breakthrough" will be just moving away from the x86 architecture. This should buy us a couple more decades at least before we have to start replacing silicon, or whatever other new innovation occurs during that time
@Jack-he8jv
@Jack-he8jv Рік тому
god i hope not, i hate this rent mentality. fuking WEF kikes will better write their wills if they try this.
@qm3ster
@qm3ster Рік тому
Dear Ivan. At 9:05, there's one frame where the cmos construction steps are shrinking, and the diagram there is not translated from Russian. Please enjoy it!
@Daojyn
@Daojyn Рік тому
Very good video, but I have two questions: -I wonder if it took around a half a century to reach this performance on silicon/CMOS doesn't it mean that it would take around the same time for others technologies to catch up or do improvements already discovered accelerate the rest? -After the huge rise of chip price in the last years can we expect before others technologies are catching up, let's say in the next decades that chip prices will decrease a lot because of the silicon limit that meanwhile we will see a comeback of dual CPU and SLI/crossfire for PC to keep improving performances?
@Jack-he8jv
@Jack-he8jv Рік тому
amd chiplet design and intel vastly inferior big,little design is already a form of dual cpus. sli is very unlikely to ever come back due to how frustrating it was for everyone. for gpu its most likely gonna be software improvements.
@joefish6091
@joefish6091 Рік тому
RISC-V is the future.
@BadgerGamePlay
@BadgerGamePlay Рік тому
only ~800 subs oO such high production, nice video man cant wait to see more!
@ZMacZ
@ZMacZ 3 місяці тому
When it comes to silicon being a pillar of computing, that will never really go away. Why ? Mass computing. Basically, when wanting to do computing en masse, given that power usage is not a problem with enough silicon-based photovoltaics in mind, computing can still use mostly silicon-based computing for when computing something that requires so much computing power that using the more scarce resource based computing options become prohibitively expensive. It's like creating a very large computer that can be like a workhorse, while using the more scarce resource based computing option for small applications. The workhorse can be used as an off-site computing 'monster' to offload work on. This may seem rather inefficient, but given that electricity can easily be farmed with solar voltaics, the power requirements are a lesser drawback, while the mass computing becomes cheaper over time, and can yet benefit from increased performance. Once people go into space and build space stations, remote computing this way can be much cheaper than supplying the more scarce resource based computing options for all/everyone. The problem is not one about creating a few million small computing options that require scarce resources (and thus becoming expensive), but once the newer options need to be created by the billions, yearly. The amount of mining required to get enough of the scarcer resource would be bad for the environment in such a fashion that humanity would 'compute' itself to death. Since, computing to death does not compute, en masse computing with the cheaper silicon option makes more sense. Also notice that en masse computing is much more efficient. When a resource isn't used by someone, someone else may be allocated more computing resource aiding his or her task to complete more rapidly. When it comes to remote computing for things like gaming, this may also work well. Silicon will always stay the best and easiest resource for computing, since it's most abundant. Also, the new computing options will have their own size limitations, albeit somewhat smaller. This would be a fixed factor. 4:28 Yes, the cost of fabrication of non silicon processors is much higher due to the scarcity of the required materials. This will never change. Also, due to the limitation of availability the newer processors would be fabricated in much smaller quantity, while being much more expensive. Due to rising cost of scarce materials the cost over time would rise, not become lower. For regular silicon-based processors, these would become cheaper over time. Combined with photovoltaics the increased power usage would not be the limiting factor. 13:56 Photonics are best used for transfer of data, not computing. Extreme short relays, let's say between chip layers and external to chip pathways would be an option, since these can transfer data really fast, and yet use very little power, which results in less heat release. Photons travel light so to speak while electron travel is heavy. This means less heat release. There's one problem though, the creation of very small light emitters/receivers and very small fiber optics or such is much harder than a regular circuit. It's at this point in time at least not very useful. A similar thing is with the wattage per computing. If the wattage becomes low enough, further reduction of such becomes less useful, and with that only extreme high end computing purposes would be served by further reduction. Let's examine an example, where you can have a fully capable processor using a 10 cm by 10 cm solar panel for continuous usage. What would the difference make if it's then operable by an 8 x 8 cm solar panel instead, by increasing the cost tenfold ? It would mean less power used, yes, by as much as 33%, but then the tenfold cost ? So, at some point in time the resource cost will define the computing solution, rather than its capability, once again referring to the scarcity of materials. 14:46 Photonics will not remove the limitations of distance. Any signal will still be limited to the speed of light, whether electronic or photonic based. Latency stays the same no matter which of these mediums are used. Basically, the best way to make processors smaller and faster, currently, is to provide on-chip memory, with comparatively slow memory being used for the externals. On chip memory would increase processing by removing much of the latency of memory instructions, like cache does, but then used on a larger scale, with only one cache and larger memory area. Then you'd have 16 GB base memory on the chip, with like 4-8 megs in cache. When a large-scale external reference would occur (like writing to SSD) the memory controller would copy the on-chip memory to the external one, subsequently transferring it to SSD, while optimizing the transfer speed, between on-chip and external memory. On chip the memory could have the same latency as regular cache, but transfer between the on-chip memory and external (DDR) memory would still be many times greater than the transfer speed between regular memory and SSD. Also, when using stuff like virtual memory on an SSD, this could then be replaced by the external (DDR) memory, increasing computational output per unit by that much as well. Knowing that most of the CPU's and GPU's time is wasted on waiting for memory return, you know that this transfer of operations from external memory to on-chip memory can make a big difference, even when the computational output of the processor is yet smaller. Solving each bottleneck in turn is also a way to speed up computing, in some cases making a bigger difference than making it smaller.
@richardnorbron4969
@richardnorbron4969 Рік тому
Great video. Thanks for posting.
@Wbfuhn
@Wbfuhn Рік тому
Look at it this way. Human DNA is 2nm. To achieve human DNA size would be not only astounding but also open the potential for micro robots. At some point all manufacturers will have no choice but to develop more cores to handle more tasks. In addition to this, it might be possible to develop cybernetics and simulate human nerves.
@stormevans6897
@stormevans6897 Рік тому
First time I've ever seen a channel translated like this, cool
@GuigEspritDuSage
@GuigEspritDuSage Рік тому
According to some scientists, quantic computer may never become real fully capable computers and instead became at best some kind of accelerator or limited to very specific tasks.
@ProDigit80
@ProDigit80 Рік тому
Overheating becomes less of an issue with the reduction of transistor size. The only reason modern transistors are emanating more heat than older models, is because manufacturers make use of that reduction in heat output, by increasing cores, core frequency, and core complexity (more transistors per core).
@cj09beira
@cj09beira Рік тому
that's only happening because the transistors no longer use less power when Shrank at the same rate of the volume reduction which used to be the case now a 2x volume reduction only leads to 20-30% power improvement, which is not enough,
@xxamulyaxx
@xxamulyaxx Рік тому
UKposts recommendations at its finest. Keep up.
@jmr1920
@jmr1920 Рік тому
good content. i really enjoyed it up to the end
@radomiami
@radomiami Рік тому
If I were to guess which one would become the first step in a post-silicon CMOS world, it would probably be a room temperature tunnel field effect transistor.
@SupaKoopaTroopa64
@SupaKoopaTroopa64 Рік тому
I'd put my money on silicon/photonic hybrid processors. There are already ASICs out there using photonic processing which are hundreds of times more energy efficient than silicon equivalents. We already have the technology, it's just a matter of integrating them into a full CPU/GPU.
@absolute___zero
@absolute___zero Рік тому
​@@SupaKoopaTroopa64 photons are mass-less packets of energy and thus it is very difficult to make them interact between themselves and with any piece of matter we are made or our chips are made of. that's the reason why we still use CMOS transistors with gate delay of 30 picoseconds instead of using light and get attosecond "gate delays" with "light transistors", if they would exist. Everybody has it very clear, creating a chip purely using light will make it millions of times faster than silicon chips. The reason why light is so fast, it is because it has no mass. But you can't make a flip-flop out of photons. That is why accelerating CMOS chips with analog-style logic (implemented using interference, like some startups are doing right now) is a bad idea. It is like putting a F1 engine on a bicycle and expecting it is going to do speeds of 400 km/hour. The bearings of the bicycle wheels will simply melt because they aren't designed for such high RPM (not talking even about aerodynamics here). It would be a good idea for a college student working in his garage, but not for a company that rises 100 million or so for this kind of project. I suggest you to learn quantum physics and chip-design before you invest. The rate of bankrupcy of new companies is like 9 to 10, if not more. Only understanding the physics of the computation you can make a fair judgement of which company is going right direction or not. There are always tons of options to invest but only few of them are real opportunities.
@SupaKoopaTroopa64
@SupaKoopaTroopa64 Рік тому
@@absolute___zero I'm aware of these limitations. I'm just saying that production-ready photonics already exist, so they have a head start over many other technologies. Also, I don't plan on investing in any photonics startups, or even anything in the microprocessor industry right now, I was just saying that If I had to make a bet on which of these technologies would first appear in a consumer product, I'd go with photonics.
@absolute___zero
@absolute___zero Рік тому
@@SupaKoopaTroopa64 photonics of course has future, but not as co-processors, rather a standalone computing logic with only a few CMOS. But there are many other computing methods that have been developed by scientists but not exploited yet. For example, there are mechanical integrated circuits made of mechanical relays (they achieve 10 nanosecond times to switch on/off), or there are gold transistors working in vacuum like vacuum tubes, there are DNA computers, and so on... the next big idea will be a new computing paradigm. There is also lots of work to do in software to gain performance, for example, development of massively parallel operating system.
@alexvincheu
@alexvincheu Рік тому
This thinking is what we humans do 🤔 and how we achieve it! Just by solving problems one by one. Then we wait for another or more problems and we solve them again, and again. That's the beauty in science and what our world represents.
@JakubHohn
@JakubHohn Рік тому
Thank you, I was lacking any news about memristors for about 8 years, so I stopped looking for updates. It is great to hear there is finally some progress
@mnomadvfx
@mnomadvfx Рік тому
The main problem with memristors is the success of and investment in NAND flash memory. The advent of 3D multi layer NAND devices allowed an otherwise dying technology a new lease on life in the middle of the last decade, and it is still pushing forward. Even though ReRAM / memristors would be far superior in power consumption, latency and speed they are still leagues behind NAND flash in density even at the 2D level, and at the 3D level it's not even worth mentioning. Weebit is making a headway, but it will be years longer before we see anything truly commercial come out of it.
@kwgm8578
@kwgm8578 4 місяці тому
Even if Moore's law may not apply to Si applications, they are still very useful in solving many of our computational applications at 16 and even 32 nm technologies.
@ZMacZ
@ZMacZ 3 місяці тому
16:58 Classic mistake, the qubits are not in every state at once, each can only have one state at any given time. They fluctuate really fast, though. The boundaries between reading 1 and 0's digitally limit these state's readings, while the qubit itself is factually an analog computing bit, with the state varying between 0 and x, where x means the highest possible potential contained in said qubit. 18:07 Well, I'd like to see quantum graphics cards, with a nice new Unreal Engine 6 or so, personally. Nanites + quantum => huge open worlds with extreme distant view capability.
@onlinetech8077
@onlinetech8077 Рік тому
Superb man👍💐👌♥️ #Mycomputer
@GrandNecro
@GrandNecro Рік тому
i think just coming up with better architechture and firmware is the easier route for now than finding a replacement for silicon
@raduradu1289
@raduradu1289 Місяць тому
0:20 Credit for obtaining the first pure silicon goes to Swedish chemist Jöns Jacob Berzelius. He succeeded in this in the year 1823. 4:20 Germanium semiconductors have a maximum operating temperature of up to 70°C, which is much lower than silicon semiconductors.
@thebogsofmordor7356
@thebogsofmordor7356 Рік тому
Ayyy I am your 1000th subscriber! Great video
@bakedbeings
@bakedbeings 3 місяці тому
There's good news coming out of georgia tech this week about silicon epigraphene transistors, compatible with current manufacturing methods 👍
@gabriel2fan
@gabriel2fan Рік тому
Goood video man, best wishes from Switzerland
@AzErLoc
@AzErLoc Рік тому
Silicon is also relatively easy to work with and process.
@leonel9110
@leonel9110 Рік тому
Exactly and they’re working on 3nm already
@kykykyykykyk2851
@kykykyykykyk2851 Рік тому
finally yt algorithm doing his job well
@ctsd623
@ctsd623 Рік тому
Re: "silicon being near the end of its life cycle," this is just sensationalistic journalism (or an uneducated remark). It's not going anywhere in our lifetimes due to the deeply entrenced and highly refined manufacturing process. More advanced materials will have a much higher and impractical cost of scale. The most certain outcome is slowly more advanced materials will be used in conjunction with silicon, but it isn't anywhere near ending its life cycle, not for a hundred years or more. We can even do photonic and quantum compute on silicon.
@Mew178
@Mew178 Рік тому
Good new channel subbed. Well done.
@mr.electronx9036
@mr.electronx9036 Рік тому
I think the "next thing" will be photonic cpu or/and ai neuromorphic chiplet, which both are extryemly fast. The issue is still a silicon as a substrate. I heard that specific "glass" can be used even more better than silicon
@roboticroom7283
@roboticroom7283 Рік тому
GE is used in military industry that's why. + Gallium arsenide is used in military because chips made from that materials has lowest fail rate and extended temperature limits.
@redstefan6515
@redstefan6515 Рік тому
very good video, this truly is a hidden gem
@asdfghyter
@asdfghyter Рік тому
Another direction we could take to fundamentally change how computers work in order to increase efficiency is reversible computing. The laws of thermodynamics gives us a minimal possible energy cost per bit of computation, which we are slowly approaching. However, there is a workaround, since the actual cost is in erasing information. So long as the computation does not erase information, there is no theoretical lower limit in energy use. With technology based on this, it might be possible to build 3d-chips having to worry about the heat problem at all. That said, it's at least as fundamentally different from how computers function today as quantum computers are, so it will be very difficult to transition and we will still need components that actually erase information in many cases.
@JamesHill-vs4kn
@JamesHill-vs4kn Рік тому
Thank you! We support you.
@humanistwriting5477
@humanistwriting5477 Рік тому
Oh I think you've missed quite a bit; Photonic computers are not just transmitting data with light, they are CPU's, and very powerful ones at that. Memristors do not just simulate neurons closely they are physically closer to logic gates then transistors
@larrysal8866
@larrysal8866 Рік тому
Can't wait to get my own silicone chip manufacturing system for cheap in a few decades B)
@antonnym214
@antonnym214 Рік тому
Check into computing with Ternary. It is mathematically more efficient than binary and requires just one extra voltage level. This could lead to more efficient and faster CPUs than we have today and some are already being built. All good wishes.
@diegoochoa572
@diegoochoa572 Рік тому
Bro amazing content!
@projectw.a.a.p.f.t.a.d7762
@projectw.a.a.p.f.t.a.d7762 Рік тому
I was watching how atom's are being controlled and manipulated on 2d materials. That would suggest we have a ways to go, regarding Moore's Laws future.
@ouch1223
@ouch1223 Рік тому
as an old viewer of yours, i'd like to wish u luck with ur new channel :D
@SarcastHandleNotAvailable
@SarcastHandleNotAvailable Рік тому
great video, keep it up!
@olagarto1917
@olagarto1917 Рік тому
Photonics can also do prossessing via constructive/distructive wave interference
@agsystems8220
@agsystems8220 Рік тому
Not really, as it messes with phase information. You can build simple gates, but as soon as you try to do something like an and gate either your 1+1 value is different amplitude from 1+0, or its phase is different. This is fine if you are just throwing it at a detector, but if you try to throw it into another gate built around interference it is not going to work right. There might be some quantum mechanical way to align the phases, but the naive approach does not work.
@SC-zq6cu
@SC-zq6cu Рік тому
Photonics cannot store memory yet as that would imply being able to trap light for very long periods of time which is still impossible.
@JorgetePanete
@JorgetePanete Рік тому
processing*
@samirkumargupta3502
@samirkumargupta3502 Рік тому
Awesome video, thanks.from India.
@aaronsmith593
@aaronsmith593 8 місяців тому
How does one remember which position all those switches are in ?!? Either On or Off. Well that is what computer glitches are, the chip is forgetting where, what, who and when the switches are doing at any given time. Software errors can multiply these error glitches.😮😊
@avetruetocaesar3463
@avetruetocaesar3463 Рік тому
I wholeheartedly disagree with the concluding section of this video. It insinuates that quantum-cloud-compute backend reliant hardware, which are about as functional as a storage device with some added connectivity, but mostly hollow toys which will serve as GUIs is the future of computers. I reject such a plane of existence, even if there is simplicity, or even necessity, in it; such a paradigm shift would basically mark the end of personal computers and ownership of anything related to computers. Quantum computers may as well exist in their own space and continue to get more sophisticated; they will still require close-to-0-Kelvin temperatures to even function, so an average person cannot carry one such device, let alone drive one in a home setting in today's society. I cannot visualise a future in which these devices become so compact, efficient, easy-to-manage and affordable as to be the de facto standard for computers because somehow, for some godforsaken reason, they are the only way going forward. How dreadful.
@Luizfernando-dm2rf
@Luizfernando-dm2rf Рік тому
Ignoring the severe limitations Q-computing has as of today, the prospect of not being able to own a full computer is actual garbage. I'd rather stay with my classical slow piece of junk than having to trust big techs and cloud providers.
@lucifermorningstar8051
@lucifermorningstar8051 Рік тому
amazing video. made me subscribe. ill watch more
@tombouie
@tombouie Рік тому
Thks again & request you update your most excellent video once a ~quarter.
@xp.949
@xp.949 Рік тому
What version of knock knock on heavens door it is bro? Thnks
@MommeSherif
@MommeSherif Рік тому
Amazing content 🥰
@antonnym214
@antonnym214 Рік тому
Very nice reporting. I subscribed immediately. Thank you!
@claudiohase296
@claudiohase296 Рік тому
Very GOOD !!!!
@SuperSnotgun
@SuperSnotgun Рік тому
Silicon still has a way to go before its phased out. The key is better packaging (things like 3D stacking).
@DespaceMan
@DespaceMan Рік тому
Interesting, keep an eye out for Biphenylene I believe this will have a wider application field than graphene.
@feey1443
@feey1443 Рік тому
Thanks for the info
@morallycorruptmaleficentti1194
@morallycorruptmaleficentti1194 Рік тому
I worked on photonic semiconductors at Berkeley thats the next wave for chips for sure if yall are interested look it up
@WahteverYouWant
@WahteverYouWant Рік тому
Dam man you need to give credit to Eugene Khutoryanski for using his graphics
@ERRORFACE1
@ERRORFACE1 Рік тому
thanks for the vid. what's the outro song thanks
@ma2i485
@ma2i485 Рік тому
Great video bro
@BongoFerno
@BongoFerno Рік тому
You forgot multi valued transistors. Non binary transistors which can switch between more than 2 values. A 4 state transistor can emulate 2 binary transistor, without miniaturization. So it can preserve Moore's law, without requiring smaller atoms, or getting into tunneling troubles.
@nathanwest2304
@nathanwest2304 Рік тому
I really think that graphene is the right way to move forward, we don't know how or when, but ultimately they will be the solution
@dinozaurpickupline4221
@dinozaurpickupline4221 Рік тому
actually a silicon graphene polymer has already shown upto 60% improvements in to the polarities & movement of electrons so maybe silicon will move to that
@fk319fk
@fk319fk Рік тому
This is what I like best about UKposts, you got me thinking... 1) The way we are doing things now is coming to an end. 1a) we may get smaller features, but we are close to the limit 1b) due to size, we may find different circuits that get better results, but again an incremental change 1c) chiplets, and other die layouts, can give up more performance, but nothing major. 2) We need to think about pipeline, Risc/Cisc/Vliw, process. 2a) my personal opinion is we need to rethink compilers and CPU so they make optimal code, not optimizing compilers for a CPU 3) The software process is broken 3a) with ~8m lines of code for Linux and ~50m for Windows 10, something is wrong 4) I am a product of 16-bit words and 64K of memory. 4a) My OS was 4k 4b) I could run a word processor or a spreadsheet 4c) my machine was 10 MHz 5) when we buy a new machine, we generally get all new things 5a) New OS do NOT need to be backward compatible I really think we need to look at how we move the software to a CPU and fix that issue.
@cj09beira
@cj09beira Рік тому
that was a different time, with complexity many orders of magnitude simpler, no need to support a hundred different processors, and thousands of different IO devices, yes it has to be mostly backwards compatible or else you waste billions of dollars of software development for little to no gain. chiplets still has a long road ahead, we can go really far with it specially when we start stacking chiplets.
@ebindanjan
@ebindanjan Рік тому
Very good and interesting content. Thank you
@shows1234
@shows1234 Рік тому
Hi there, it seems that the audio and video aren't synched.
@johndoh5182
@johndoh5182 Рік тому
Umm, the death of silicon as a base for ICs is GREATLY overstated. There is this thought about some END coming that just fails to think clearly about what will happen. So let's evaluate some timelines. Sure, silicon is going to hit a limit, but so would any other material. If you want an IC based on on/off transistors, there is only so small you can make the transistors, but once you hit that theoretical size, is that the death of silicon. The answer is no. YOu can measure CPU performance in multiple ways. One is raw performance. Raw performance is based on work done in a certain amount of time, but to calculate this you use multiple types of applications to calculate this performance. If I want a render machine, do I need a CPU that's great at gaming? The answer is no. I need something that's great at stream processing, and that could be a combination of CPUs and GPUs and accelerators, or a computer that integrates all three in a single circuit to run any type of stream processing programs. Maybe with the accelerators I can program it first before running the job, so the first part of the render application checks for an FPGA and if it exists it programs the accelerators before running the render tasks. Next, at the beginning, which is as far as I got because these kind of videos are silly, the talk is on chiplets and IPC (Instructions per clock cycle). What was skipped is making a process node much more efficient at running higher speeds. There's no law that says a CPU can't be clocked over 6GHz. In fact I think over the next 5 years TSMC and Intel will both have process nodes that run 6 - 6.5GHz and it won't be very high power consumption. Also what isn't mentioned is compute power in 3 dimensional space. We're getting to the point to where ICs can have multiple layers, and this is for different reasons, one being the shrinking of transistor size and every shrink usually comes with power reduction. I think it's safe to say that CPU design is going to involve 3D layouts and this will involve layering in such a way to where one layer isn't producing so much heat and the other layer can, so planning out the layout of circuits will become more complex. But here is the REAL meat and potatoes of this. Home users already have access to so much compute power it's crazy. A home user can now do with large servers had to do 10 years ago, and in another 5 years the amount of power that can exist in a laptop is competing with servers from a decade ago. The move to 2nm or Intel 20A which are the same thing is SO monumental that for home users there's no need to get past that. The issue won't be not having enough compute power, but having too much with rare exception. One area of growth is PC gaming. Graphics compute still has to improve to get high quality 4K gaming. But, we're there NOW. The GPUs that AMD and Nvidia are releasing right now are good enough for high quality 4K gaming, and this is using a 5nm node or modified 5nm node (TSMC N4). When semi-conductor companies get to 2nm the issue won't be not having enough transistors anymore. It will be hard to keep adding more compute units (graphics processors) and in fact for Nvidia GPUs they're using multiple types of compute in a single GPU, and that's basically the raw compute for making the image on the screen, a lighting and shading technique used for games called ray tracing for there are RT cores, and then the cores for AI, tensor cores. 2nm is SO small with SO MUCH transistor density that home computing just never needs to move to anything else, but there will be other nodes that come out with better transistor density, so like 18, 16A, etc.... and this can probably go all the way down to 10A. By this time we're about 2 decades from now. The REAL issue isn't going to be transistor density. It WILL be about making machines smarter, to where they can do more with the same number of transistors. IPC isn't a gimic. It's how you get more performance in a fixed time period from a CPU over an older CPU at the same clock speed. So, there are many tools for improving compute power, and they are NOT gimmicks to hide some flaw. It's PART of better engineering for computing. One is more processor cores. For home computing though you hit a limit because most programs home users use don't have so many processing threads and adding ever more cores is a total waste. For servers on the other hand, there are many types of server compute loads that can benefit for a LOT more cores. Another is clock speed. But this has to be solved while at the same time keeping power consumption down. THIS is a real issue and it's hard to solve, but it IS getting solved slowly, but every process node is different so you have to solve that problem for every process node. Another is instructions per clock (IPC). This also improves over time and one of the ways to solve this is simply to have more cache so the cores have more data and instructions next to the cores instead of having to access system memory to get data/instructions. But there are other ways to improve IPC too. Another is making better processors that can accelerate workloads. AMD, Intel and Nvidia are all working on this, and even Apple has done excellent work with acceleration and that's part of what makes the Apple M1 and M2 processors so fast and NOT the fact that it's based on ARM instead of X86-64. Another is being able to stack transistors which shrinks the area needed for the same circuit. Another IS chiplets. This helps with heat issues. The point is, yes silicon will hit a limit for transistor density for a 2 dimensional area, but that's a single problem, not THE problem. Solving the other problems I listed is more important. And lastly we can talk about how much silicon is needed. Well, as process nodes shrink transistor size over time, a single wafer can make a lot more ICs, so we are always improving the number of circuits that some from a fixed amount of silicon. It's not an issue at least not for another 50 years or so.
@markvisser388
@markvisser388 Рік тому
The Japanese have made it possible, and in mass. Smart move abandoning silicon 20 years ago to focus on developing GaN. Hands down!
@KingSteven77
@KingSteven77 Рік тому
Good video but the video and sound is not synced and it's annoying but other than that it's 10/10
@francisdebriey3609
@francisdebriey3609 Рік тому
Very good ... subscribed !
@Fernando-wz6no
@Fernando-wz6no Рік тому
Quantum computer will *NEVER* happen. Forget that *FANTASY* !
@mnomadvfx
@mnomadvfx Рік тому
Derp, you completely ignored spintronics as a possible future path. Which is funny as it easily has the most diverse amount of research going in with it.
@nadergadelrab2669
@nadergadelrab2669 Рік тому
amazing
@ashishchauhan2759
@ashishchauhan2759 Місяць тому
Legendary content 😮
NVIDIA'S HUGE AI Chip Breakthroughs Change Everything (Supercut)
26:08
Ticker Symbol: YOU
Переглядів 1,3 млн
The Race to Build a Perfect Computer Chip
26:02
Bloomberg Originals
Переглядів 774 тис.
0% Respect Moments 😥
00:27
LE FOOT EN VIDÉO
Переглядів 32 млн
ISSEI funny story😂😂😂Strange World | Magic Lips💋
00:36
ISSEI / いっせい
Переглядів 18 млн
Повістки у Києві: «Яке право вони мають забирати всіх мужиків?» #війна #мобілізація #військові
00:41
Слідство.Інфо | Розслідування, репортажі, викриття
Переглядів 1,6 млн
The Most Powerful Computers You've Never Heard Of
20:13
Veritasium
Переглядів 10 млн
How Quantum Computers Break The Internet... Starting Now
24:29
Veritasium
Переглядів 8 млн
The Gate-All-Around Transistor is Coming
15:44
Asianometry
Переглядів 428 тис.
CHIPLETS: Divide and Conquer | The Future of Processors
14:32
My Computer
Переглядів 13 тис.
Future Computers Will Be Radically Different (Analog Computing)
21:42
Something Strange Happens When You Follow Einstein's Math
37:03
Veritasium
Переглядів 6 млн
Will gallium nitride electronics change the world? | Upscaled
14:29
Why making chips is so hard
16:03
TechAltar
Переглядів 898 тис.
How Apple Just Changed the Entire Industry (M1 Chip)
26:28
ColdFusion
Переглядів 4,9 млн
Processor under microscope. Nanometer journey
12:41
My Computer
Переглядів 1 млн
Phone charger explosion
0:43
_vector_
Переглядів 35 млн
RTX 4070 Super слишком хороша. Меня это бесит
15:22
Рома, Просто Рома
Переглядів 89 тис.
Секретная функция ютуба 😱🐍 #shorts
0:14
Владислав Шудейко
Переглядів 2,1 млн