Ray Tracing: How NVIDIA Solved the Impossible!

  Переглядів 784,204

Two Minute Papers

Two Minute Papers

Рік тому

❤️ Check out Lambda here and sign up for their GPU Cloud: lambdalabs.com/papers
📝 The showcased papers are available here:
research.nvidia.com/publicati...
research.nvidia.com/publicati...
graphics.cs.utah.edu/research...
users.cg.tuwien.ac.at/zsolnai...
Link to the talk at GTC: www.nvidia.com/en-us/on-deman...
If you wish to learn more about light transport, I have a course that is free for everyone, no strings attached:
users.cg.tuwien.ac.at/zsolnai...
❤️ Watch these videos in early access on our Patreon page or join us here on UKposts:
- / twominutepapers
- / @twominutepapers
🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:
Aleksandr Mashrabov, Alex Balfanz, Alex Haro, Andrew Melnychuk, Benji Rabhan, Bryan Learn, B Shang, Christian Ahlin, Eric Martel, Geronimo Moralez, Gordon Child, Jace O'Brien, Jack Lukic, John Le, Jonas, Jonathan, Kenneth Davis, Klaus Busse, Kyle Davis, Lorin Atzberger, Lukas Biewald, Luke Dominique Warner, Matthew Allen Fisher, Michael Albrecht, Michael Tedder, Nevin Spoljaric, Nikhil Velpanur, Owen Campbell-Moore, Owen Skarpness, Rajarshi Nigam, Ramsey Elbasheer, Steef, Taras Bobrovytsky, Ted Johnson, Thomas Krcmar, Timothy Sum Hon Mun, Torsten Reil, Tybie Fitzhugh, Ueli Gallizzi.
If you wish to appear here or pick up other perks, click here: / twominutepapers
Thumbnail background design: Felícia Zsolnai-Fehér - felicia.hu
Károly Zsolnai-Fehér's links:
Instagram: / twominutepapers
Twitter: / twominutepapers
Web: cg.tuwien.ac.at/~zsolnai/

КОМЕНТАРІ: 915
@PsuedoNymrod
@PsuedoNymrod Рік тому
Thank you for demonstrating (and including the links!) to these papers. Glad we already have access to one of these techniques and excited to see more commercially-available tools based on the other pieces of research you've highlighted!
@TwoMinutePapers
@TwoMinutePapers Рік тому
Thank you so much! So delighted to see that you Fellow Scholars are also enjoying light transport - it's the best! 😊
@smanzoli
@smanzoli Рік тому
UE5 uses Lumen, and it’s not only awesome and blazing fast for Global Illymination, but also you can add Voxel 2.0 plugin
@kairu_b
@kairu_b Рік тому
Great research
@demp11
@demp11 Рік тому
@@smanzoli yes it's great but not nearly as accurate as real ray tracing. There are a couple vids showing how trees for example and some other objects appears mich brighter in lumen than they would ray traced. That's because lumen uses spacial light samples, not Pixel light sample like raytracing. So instead of shooting rays out of the camera, it samples 3d pixels (voxels) in a cube grid on the map. That's basically what they always used for rendering game lighting, with the difference that lumen works in real time and doesn't need to be baked.
@tbuk8350
@tbuk8350 Рік тому
Yeah, that realtime denoising is awesome! In terms of quality, it's still not quite as good as something like Intel's OpenImageDenoise (which runs on the CPU). The difference is their method is realtime, and OpenImageDenoise isn't, without a ridiculous CPU. Edit: nevermind, a developer working on OpenImageDenoise got it running in realtime on both AMD and Nvidia GPUs, meaning it isn't locked to just Nvidia. OpenImageDenoise is definitely the better solution, or I suppose will be once they finish up the build that functions on the GPU.
@kobby2g8
@kobby2g8 Рік тому
As the render times get shorter, the 2 minute papers get longer 😛
@iurivanastacio3081
@iurivanastacio3081 Рік тому
Lol
@calebmcurby8580
@calebmcurby8580 Рік тому
They're now two minutes to the power of four lol
@Twisted_Logic
@Twisted_Logic Рік тому
The real fundamental law of papers
@daria_c
@daria_c Рік тому
I enjoyed the longer format as I don’t watch every video.
@JorgetePanete
@JorgetePanete Рік тому
Heisenberg would be proud
@UON
@UON Рік тому
I feel there's no way I can hold my papers down with these ferocious winds of change
@alansmithee419
@alansmithee419 Рік тому
We're gonna need a bigger paperweight.
@UON
@UON Рік тому
​@@alansmithee419 only an RTX 4090 could possibly work
@sejemandhaha
@sejemandhaha Рік тому
@@UON yeah wont even have any papers left to hold down when that's bought
@jebus456
@jebus456 Рік тому
same
@Edbrad
@Edbrad Рік тому
I want a paper on trying take get a realistic render of a Paper being squeeze that’s 100% realistic
@TwoMinutePapers
@TwoMinutePapers Рік тому
It is such an honor to get an invitation to GTC to talk about light transport. Thank you so much! 🙏 So happy! 😊
@DaylightFactory
@DaylightFactory Рік тому
Thank you for presenting all of this astounding research. More than the technology itself, it is your life-affirming delivery that inspires optimism and enthusiasm for a future that otherwise contains so many bright unknowns as to be terrifying. You are the Bob Ross of deep learning.
@zeblanoue2642
@zeblanoue2642 Рік тому
You’ve more than earned it! :)
@shanekhiu9884
@shanekhiu9884 Рік тому
what a time to be alive!
@delpinsky
@delpinsky Рік тому
You really deserve this honor, sir! 👏
@immanueldavielgarcia
@immanueldavielgarcia Рік тому
Do a video on the new Jurassic-1 language model
@scaredyfish
@scaredyfish Рік тому
I’m always amazed how even as hardware gets faster, algorithms still keep getting smarter, so you get a double boost in performance.
@MadScientist512
@MadScientist512 Рік тому
Moore's Law has been dying this millenia so hardware improvements have become incremental and not exponential like in the 'good old days' when we expected a doubling of performance every generation, now the younger generation considers a 25% increase to be significant; so it's really only Algorithms/AI getting exponentially smarter that's giving these massive improvements, but like Moore's Law that's also bound to reach a limit, hopefully later than sooner. What's really needed is new computing technology that once again follows Moore's Law, but since capitalism is still making huge profits from old silicon tech I ain't holding my breath.
@thebcwonder4850
@thebcwonder4850 Рік тому
@@MadScientist512 we’re getting 50-70% performance uplift per generation in GPUs, I don’t know where you got the 25% from
@n9s3nse10
@n9s3nse10 Рік тому
@@MadScientist512 we are getting 2x improvement if you combine both hardware and software i.e. dlss
@atifarshad7624
@atifarshad7624 Рік тому
@@n9s3nse10 dlss is software. From strictly hardware point, we are nowadays getting 25% to 33% at most per generation
@iamwisdomsky
@iamwisdomsky Рік тому
our best bet right now is improvements in algorithms particularly AI or machine learning. machine learning is more on guessing rather than actual computation. but the thing with machine learning is learns so it also scales up as time goes on just like the transistors in a chip. what's rough guess right now could be an accurate outcome in the future.
@Wobbothe3rd
@Wobbothe3rd Рік тому
Still squeezing in 2022! I remember me and my sister trying to render Bryce pictures on a pentuim 90. It took HOURS to make a sphere over water, lol. Miraculous stuff here.
@jean_mollycutpurse_winchester
@jean_mollycutpurse_winchester Рік тому
Back in about 1987 I was using the forerunner of cinema 4D on an Amega 500 and I remember setting up a scene and going on holiday for three weeks and when I come back it had only done half the scene!
@LongTran-em6hc
@LongTran-em6hc Рік тому
It did not crashed in 3 weeks? What kind of sorcery is that
@jean_mollycutpurse_winchester
@jean_mollycutpurse_winchester Рік тому
@@LongTran-em6hc 🤣🤣🤣
@alinaqirizvi1441
@alinaqirizvi1441 Рік тому
@@LongTran-em6hc I ran my Linux PC crypto mining for like a month and a half without rebooting it and it never crashed
@SmallSpoonBrigade
@SmallSpoonBrigade Рік тому
One of the things that's changed is that we have more cores now. Even at the time, if you were fortunate enough to have a dual-processor rig, that would have been much closer to 45 hours as ray tracing is trivial to run in parallel processes. And with modern heardware, just being able to split the load to 8 or more processors alone would cut the time down to just over 10 hours. With the more efficient processors and increased RAM it would be done much faster.
@leynenslucker2991
@leynenslucker2991 Рік тому
I did my master on this. Computer Graphics. And I must confess this is insane. Like no way this is possible. I need to read all four of the papers. Wow.
@Musikpunx
@Musikpunx Рік тому
This is really an amazing time to be alive. In my particular case, I was a child, when the video game "Pong" was released. Looking at the developement of computers and AI from then to now is unbelieveble. Btw., I was born in 1967, which means 2 years before astronauts landed on the moon with a computer on board that had 80 kilobytes of memory.
@PumpyGT
@PumpyGT Рік тому
You're old, wow
@QuackZack
@QuackZack Рік тому
You're 65? What a time to be alive.
@Musikpunx
@Musikpunx Рік тому
@@QuackZack I am 54 and the time is amazing regarding computing power.
@wussrestbrook1200
@wussrestbrook1200 Рік тому
If you live long enough you may be one of the first generation to reverse aging
@trapfethen
@trapfethen Рік тому
You can get a better sense of what paper 4 is doing when you look at the edges of the image where the train is just coming into frame. You can see the detail jump levels as that spot on the train spends more time in the frame. A possible workaround to this particular detail is to render to a virtual screen slightly larger than the actual screen and cropping out the edges when displaying on screen.
@applepie9806
@applepie9806 Рік тому
ah good idea. depends on if the wider area will take up enough resources to drop framerate, but I doubt it would drop very much
@The_Forge_Master
@The_Forge_Master Рік тому
If you're talking about 13:56, you might be mistaking the raw noisy section for one with less detail. The first few columns on the left side show the noisy input to the denoising algorithm, which takes up the rest of the screen.
@lagrangewei
@lagrangewei Рік тому
this doesn't work because camera can rotate, it doesn't just go left or right. all these effect are not new, there are working example of this shader that u can apply on existing game with api hook because gamers always find way to make their game look better. all these implementation suffer from a problem when you rotate. the best "solution" i seen from is to hide this from user by implementing a camera edge motion blur when the camera rotates, u cannot predict how a player control camera will turn.
@xardinlume7345
@xardinlume7345 Рік тому
I remember last year, as I was browsing a few (not yet reviewed/incomplete/old) papers on arxiv, and stumbled across a thesis that was talking about cone tracing... I downloaded it and was like "hey one day I'll read it". I never did. This is probably not the same paper but the second you pronounced "voxel cone tracing" I was like "of COURSE". Thanks for the vid, love your work. To think they actually thought about that ten years ago... EDIT: turns out I still have the PDF of the thesis, it's "Audio and Visual Rendering with Perceptual Foundations", which doesn't match any of NVIDIA's papers, yet is dated 2009, so...
@guaposneeze
@guaposneeze Рік тому
The modern tech is fantastic. But it's also quite complex. One of the real amazing things about classical raytracing was that you could famously learn to get started "In One Weekend." Even as somebody who has been at least somewhat engaged in CG for many years (you've all seen stuff I worked on when I worked in VFX) the new stuff is less exciting because I have pretty much zero illusions that I am ever going to write an implementation of a full modern renderer by myself. Even just trying to use nVidia's Restir implementation library in my own engine rather than implementing it myself is more complex than writing a whole renderer used to be!
@alefratat4018
@alefratat4018 Рік тому
Totally agree, we've reached a point where amateurs or even low-budget professionals simply can't compete on their own, no matter how skilled they are.
@mahdiyussuf9804
@mahdiyussuf9804 Рік тому
The good old days for a lot of things are over, right!
@RamsesTheFourth
@RamsesTheFourth Рік тому
@@alefratat4018 I think low budget studios and amateurs can always compete. You dont have to use RTX if you dont want to. And the main reason I love indie gams is because they are much simpler and focused on the fun part, with great idea for game mechanics or story. AAA games usually go full bonkers on graphics, but the story is usually lacking, and they are sterile most of the time from some new interesting game mechanics.
@kennylaysh2776
@kennylaysh2776 Рік тому
@@RamsesTheFourth Doesn't he mean writing it into a custom engine? I mean just putting ray tracing in a game is easy for anyone at a basic level, it's included in UE5. I can enable it easily there - but I'd be clueless on adding it to a custom engine.
@RamsesTheFourth
@RamsesTheFourth Рік тому
@@kennylaysh2776 Thats what I understand he meant. Sure, If anyone uses commercial engine like UE/Unity then it should be easy to do.
@psiga
@psiga Рік тому
Oof. Sympathy and empathy to the people who saw that research from 11 years ago and mistakenly thought that widespread adoption and advancement of it would be just around the corner. At least it's in the right hands now. Thank you for bringing it to our attention, Károly!
@damysticalone87
@damysticalone87 Рік тому
Those read, listen to / bugged / eavesdrop / overheared / tapped, watched, spied, copied and steal many of my ideas (=inventions) initiated by me: + open, transparent 360° sunlight Buildings, construction methods, architectures, designs, concepts; + step-floor / -storey / -level / -tier pyramid (e.g., as a residential building); + 360° stepped floors / storey / level / tier buildings (constructions, architectures, construction methods, designs, concepts, 2D, 3D, models); + 3D 360° environment (surround), volume, space (room) audio / sound; + 360° environment (surround), volume, room screens (displays), monitors, TVs; + screens (displays), monitors, TVs without backlighting; + many things with magnets like Micro-OLED; + 360° screens, monitors, TVs, panels, glass; + curved screens, monitors, TVs, panels, glass; + 360° reflections and light digital: "ENB", "Ray-Tracing" / "RTX", “Lumen illumination” and whatever renaming! + Short Throw Projectors; + dark backgrounds / themes / skins for windows, browser windows, internet sites, programs / apps, etc.; + fanless, passive cooling; + thermal pads; + heat pipes; heat fins (no sure yet); + 360° heat fins; + Water cooling (unwanted, because it conducts electricity) and + liquid cooling (wanted) for the computer and personal computer / PC application area! + 360° open cooling, e.g., with air-flow by air-holes, no cases, no walls, no cover, etc.; - I am not the inventor of refrigeration by nitrogen, instant- / blast-freeze / -freezing; + 360° open, transparent PC cases; + 360° Sphere, Ball as a Wheel (2D, 3D) “Omniwheel” / “Omni-Wheel”; + from a standing position / standing a 360° turn / rotation in any direction on all objects, including logically and of course vehicles, transporters, robots of all kinds; + flying objects, planes that take off and land vertically, for example by hot air, example fighter jet; + skyscraper up to the weightlessness of space ("space scraper") - "mega construction", “orbital”, etc.; + bendable, flexible, foldable screens / displays / monitors; + bracket / mounting arms for PC, screens (not sure yet), + 360° "ergonomic" bracket / mounting arms; + LED lights (at least its implementation and intended use outside of tv and pc-monitor); RGB LED “Nanoleaf”, etc. + and much more! + I'm not the inventor of VR, but of "AR", AR glass, AR glasses; It was only later that I realized that they derived a lot from my ideas (=inventions), a lot came about that has to do with color and light, through me as an initiator, booster / catalyst, e.g., through my idea, invention of the Screens without a backlight and without a built-in / integrated backlight! It is no coincidence that only afterwards, after I initiated this, they built, built and are building those inventions thanks to my impetus! They sell my ideas (inventions) as theirs! They are not the inventors, but the first technical implementers of my ideas (=inventions)! And those are not the inventors, but the thieves of my ideas (=inventions) initiated by me, because those act as if I wasn't the first hand and the first domino, and they take unjustly, undeservedly a lot of money, stolen money (blood money), recognitions, awards, certificates, fame and history, they boast of my laurels / merits! Before me they all were stucked at LCD, Plasma TV and less lights! I have made a deep impact in evolution! Those manipulate, sabotage, falsify, distort images, paintings, digitized and real, animations, videos, films, also composed of many images, even the publication dates of mine, others and their posts, images, videos, etc.! You have to understand, those can distort everything that can be heard and seen in real life and digitally! They block and delete my pictures, videos, posts, comments, comments-answers and answers! They are poisoning and murdering the world with fake diseases, treatments, "vaccinations"/ "vaccines" and injections by syringe! Now they also make it out as if they haven't been ripping off, cheating, enslaving, murdering other countries with the money currency, money exchange by even -99% for more than two centuries! And as if I'm not the first to disclose that and more! As if I didn't disclose and initiate > 1.00 Ruble (₽) = 1.00 asian Yen (¥) = 1.00 Euro (€) = 1.00 Dollar ($) = any (X) any country < years ago! Each and every non-civilian you hear and see on TV is involved! You can hear and see their > blue blue blue
@terrsus7676
@terrsus7676 Рік тому
@@damysticalone87 ???
@damysticalone87
@damysticalone87 Рік тому
If / When you have money, health or "natural" disaster problems or if / when you are somehow affected by traffic accidents, traffic jams at some point, somewhere, then think of me, share my posts and my pages! Be sure to read my comments under my first videos completely via laptop or PC! Be sure to read my contributions, posts, comments under my first videos thoroughly, maybe no longer visible via phone app because they intentionally set it that way! Use laptop or PC!
@JoeARedHawk275
@JoeARedHawk275 Рік тому
@@damysticalone87 ///beep hoop////beepboop
@damysticalone87
@damysticalone87 Рік тому
If / When you have money, health or "natural" disaster problems or if / when you are somehow affected by traffic accidents, traffic jams at some point, somewhere, then think of me, share my posts and my pages! Be sure to read my comments under my first videos completely via laptop or PC! Be sure to read my contributions, posts, comments under my first videos thoroughly, maybe no longer visible via phone app because they intentionally set it that way! Use laptop or PC!
@aidanm5578
@aidanm5578 Рік тому
I've superglued my papers to my hands. Please help.
@Zylenxx
@Zylenxx Рік тому
This lineup of papers is not just perfect for high end graphics cards , too. Considering that theyre optimised where possible and still going for more, smaller graphics cards and possibly devices that arent normally MEANT to do such things could possibly opt in for a similar method for demos! This is a big change. Better light transports for the same amount of time and effort is wonderful to keep track of. Thank you alot!
@damysticalone87
@damysticalone87 Рік тому
Those read, listen to / bugged / eavesdrop / overheared / tapped, watched, spied, copied and steal many of my ideas (=inventions) initiated by me: + open, transparent 360° sunlight Buildings, construction methods, architectures, designs, concepts; + step-floor / -storey / -level / -tier pyramid (e.g., as a residential building); + 360° stepped floors / storey / level / tier buildings (constructions, architectures, construction methods, designs, concepts, 2D, 3D, models); + 3D 360° environment (surround), volume, space (room) audio / sound; + 360° environment (surround), volume, room screens (displays), monitors, TVs; + screens (displays), monitors, TVs without backlighting; + many things with magnets like Micro-OLED; + 360° screens, monitors, TVs, panels, glass; + curved screens, monitors, TVs, panels, glass; + 360° reflections and light digital: "ENB", "Ray-Tracing" / "RTX", “Lumen illumination” and whatever renaming! + Short Throw Projectors; + dark backgrounds / themes / skins for windows, browser windows, internet sites, programs / apps, etc.; + fanless, passive cooling; + thermal pads; + heat pipes; heat fins (no sure yet); + 360° heat fins; + Water cooling (unwanted, because it conducts electricity) and + liquid cooling (wanted) for the computer and personal computer / PC application area! - I am not the inventor of refrigeration by nitrogen, instant- / blast-freeze / -freezing; + 360° open, transparent PC cases; + 360° Sphere, Ball as a Wheel (2D, 3D) “Omniwheel” / “Omni-Wheel”; + from a standing position / standing a 360° turn / rotation in any direction on all objects, including logically and of course vehicles, transporters, robots of all kinds; + flying objects, planes that take off and land vertically, for example by hot air, example fighter jet; + skyscraper up to the weightlessness of space ("space scraper") - "mega construction", “orbital”, etc.; + bendable, flexible, foldable screens / displays / monitors; + bracket / mounting arms for PC, screens (not sure yet), + 360° "ergonomic" bracket / mounting arms; + LED lights (at least its implementation and intended use outside of tv and pc-monitor); RGB LED “Nanoleaf”, etc. + and much more! + I'm not the inventor of VR, but of "AR", AR glass, AR glasses; It was only later that I realized that they derived a lot from my ideas (=inventions), a lot came about that has to do with color and light, through me as an initiator, booster / catalyst, e.g., through my idea, invention of the Screens without a backlight and without a built-in / integrated backlight! It is no coincidence that only afterwards, after I initiated this, they built, built and are building those inventions thanks to my impetus! They sell my ideas (inventions) as theirs! They are not the inventors, but the first technical implementers of my ideas (=inventions)! And those are not the inventors, but the thieves of my ideas (=inventions) initiated by me, because those act as if I wasn't the first hand and the first domino, and they take unjustly, undeservedly a lot of money, stolen money (blood money), recognitions, awards, certificates, fame and history, they boast of my laurels / merits! Before me they all were stucked at LCD, Plasma TV and less lights! I have made a deep impact in evolution! Those manipulate, sabotage, falsify, distort images, paintings, digitized and real, animations, videos, films, also composed of many images, even the publication dates of mine, others and their posts, images, videos, etc.! You have to understand, those can distort everything that can be heard and seen in real life and digitally! They block and delete my pictures, videos, posts, comments, comments-answers and answers! They are poisoning and murdering the world with fake diseases, treatments, "vaccinations"/ "vaccines" and injections by syringe! Now they also make it out as if they haven't been ripping off, cheating, enslaving, murdering other countries with the money currency, money exchange by even -99% for more than two centuries! And as if I'm not the first to disclose that and more! As if I didn't disclose and initiate > 1.00 Ruble (₽) = 1.00 asian Yen (¥) = 1.00 Euro (€) = 1.00 Dollar ($) = any (X) any country < years ago! Each and every non-civilian you hear and see on TV is involved! You can hear and see their > blue blue blue
@Fingerblasterstudios
@Fingerblasterstudios Рік тому
@@damysticalone87 even if one of these things is true, an idea is not an invention. An idea is just an idea. An invention is the implementation of an idea.
@damysticalone87
@damysticalone87 Рік тому
@@Fingerblasterstudios You criminal or and supporter of criminals! Childmurderer or and supporter of chiuldmurderer! Laws! "Protection of Intellectual Property"! "Patent Rights"! Theft!
@damysticalone87
@damysticalone87 Рік тому
@@Fingerblasterstudios Laws! "Protection of Intellectual Property"! "Patent Rights"! Theft!
@damysticalone87
@damysticalone87 Рік тому
If / When you have money, health or "natural" disaster problems or if / when you are somehow affected by traffic accidents, traffic jams at some point, somewhere, then think of me, share my posts and my pages! Be sure to read my comments under my first videos completely via laptop or PC! Be sure to read my contributions, posts, comments under my first videos thoroughly, maybe no longer visible via phone app because they intentionally set it that way! Use laptop or PC!
@HAWXLEADER
@HAWXLEADER Рік тому
14:00 seems like motion vectors artifacts. It is a genius move by NVIDIA scientists to use these! 14:35 Democratizing but usually only as long as you have specific hardware ^^
@Winium
@Winium Рік тому
Specific, ready-to-use, affordable (relatively...) hardware that I can also use to game :D I think that more importantly, the papers being released mean that other chipmakers can also implement the hardware acceleration.
@michaelleue7594
@michaelleue7594 Рік тому
The first law of access is that democratization is a process. Do not look at where we are, look at where we will be, 2 more graphics cards down the line.
@katech6020
@katech6020 Рік тому
ReStir does use motion vector to benefit from light calculation done in previous frame
@kangsan2014
@kangsan2014 Рік тому
I wish you would tell us what kind of graphics card these researchers are using so that we can understand whether they are using the newest GPUs or CPUs as well as the what version of the software. I feel like that would be more informative. Your information is vastly and greatly informative but having those little bits of information will really tell us if they are using current software, GPUs and CPUs. Thank you so much and keep the excellent information coming! :)
@Nico1a5
@Nico1a5 Рік тому
They can't have a GPU much more powerful than 2x - 4x of any current high end one. So if they say 80 fps you can expect for sure to get above 20 fps with a good GPU. Unless it's something that requires a lot of VRAM
@descai10
@descai10 Рік тому
@@Nico1a5 The scene where they mentioned 80 fps, they showed that they were using a 3090.
@Nico1a5
@Nico1a5 Рік тому
@@descai10 I mean the video mentions "real time" many times, and 20 fps should be an ok number to consider "real time". If your concern is knowing what hardware was used just to run it yourself, then considering the same case you can expect a nice 5 fps. Which is great compared to a single frame taking minutes to render
@descai10
@descai10 Рік тому
@@Nico1a5 Not sure what you are talking about. They used a 3090. That means if you or I use a 3090, we will also get 80 fps in that scene.
@Nico1a5
@Nico1a5 Рік тому
@@descai10 on that scene, with that gpu, yes. Only that the video shows multiple technologies at multiple points in time, and there's no mention of the hardware used like the OP is asking
@ACGreviews
@ACGreviews Рік тому
Your stuff is constantly top tier
@trenton9
@trenton9 Рік тому
I grow more and more in love with how incredible solutions to incredible problems are often collecting dust in a drawer somewhere waiting for proper implementation.
@therealsourc3
@therealsourc3 Рік тому
Another way to make light transport faster is distance based light resolution downsampling. Far away caustics and reflections are not gonna be visible due to the amount of pixels in the image anyways so there is no need for them to be super refined, so as long as the stuff closer to the camera is good enough resolution everything further away can be progressively downsampled to reduce memory usage and possibly increase the accuracy where it matters more..
@clonkex
@clonkex Рік тому
How would you do that? Maybe render a depth map and fire more rays towards closer regions? Interesting concept
@therealsourc3
@therealsourc3 Рік тому
@@clonkex yeah, that might be quite a good trick. Maybe in combination with some ai based stuff you can downsample parts of the view that doesn't change considerably between frames too.
@torsaxberg2923
@torsaxberg2923 Рік тому
This sounds like what Unread Engine 5 does to concentrate its polygons closer to the viewer.
@protocol6
@protocol6 Рік тому
It's interesting how you think of the noise as a limitation of ray tracing when cameras and even your eyes have the same problem in low light.
@nolan412
@nolan412 Рік тому
"We simulate ISO 3200 film."
@protocol6
@protocol6 Рік тому
@@nagualdesign I think some of these are about choosing which rays to emit in the first place to improve problem areas rather than just the usual random distribution. But yes, general denoising algorithms used in post-processing for ray traced images could be used on photos taken with the wrong iso settings or in conditions below the sensor's rated light levels. I suspect some of the non-AI ones used for ray tracing were adapted from ones originally developed for photo and video denoising.
@keyreeves4631
@keyreeves4631 Рік тому
Actually cameras in modern smartphones use these sort of algorithms, to compensate for their small sensors.
@protocol6
@protocol6 Рік тому
@@nagualdesign ISO involves two different types of noise at opposite ends with a sweet spot that minimizes for a given light level if you fix the aperture and exposure time. I'm not certain but I think the most analogous type here would be selecting too low of an ISO for the light level rather than the sensor noise when the ISO is too high. Either way, the "right" ISO should minimize the noise.
@OrangeC7
@OrangeC7 Рік тому
I was wondering about whether they might want to limit the noise reduction in extremely low light areas of an image, since that may help add another subtle touch of realism. Of course, it all depends on where and how it's being applied I suppose.
@raytracer2651
@raytracer2651 Рік тому
With ray tracing , fluid dynamics , ai etc ,games are set to take a huge leap forward but the best thing is it looks like it's all in favour of the indie developer, this will make games riveting because it'll be like reading a book by a great author, one mind is the way.
@SmallSpoonBrigade
@SmallSpoonBrigade Рік тому
It kind of reminds me of when George Lucas made the prequel trilogies. It relied so much of CGI and effects, that nobody involved in the production thought to worry about the acting, writing or plot. The result are some of the worst movies in the franchise. (Although I will say seeing Yoda in that light saber fight was almost enough to make watching it worthwhile, almost)
@Analoque444
@Analoque444 Рік тому
I've been fascinated by raytracing since "Imagine" on the Amiga. It's unbelievable how far raytracing has developed today. Back then, it took me four weeks to do one picture and it was a dream that one day it would happen in real time, although it's not perfect yet. Thank you for your videos.
@damysticalone87
@damysticalone87 Рік тому
Those read, listen to / bugged / eavesdrop / overheared / tapped, watched, spied, copied and steal many of my ideas (=inventions) initiated by me: + open, transparent 360° sunlight Buildings, construction methods, architectures, designs, concepts; + step-floor / -storey / -level / -tier pyramid (e.g., as a residential building); + 360° stepped floors / storey / level / tier buildings (constructions, architectures, construction methods, designs, concepts, 2D, 3D, models); + 3D 360° environment (surround), volume, space (room) audio / sound; + 360° environment (surround), volume, room screens (displays), monitors, TVs; + screens (displays), monitors, TVs without backlighting; + many things with magnets like Micro-OLED; + 360° screens, monitors, TVs, panels, glass; + curved screens, monitors, TVs, panels, glass; + 360° reflections and light digital: "ENB", "Ray-Tracing" / "RTX", “Lumen illumination” and whatever renaming! + Short Throw Projectors; + dark backgrounds / themes / skins for windows, browser windows, internet sites, programs / apps, etc.; + fanless, passive cooling; + thermal pads; + heat pipes; heat fins (no sure yet); + 360° heat fins; + Water cooling (unwanted, because it conducts electricity) and + liquid cooling (wanted) for the computer and personal computer / PC application area! - I am not the inventor of refrigeration by nitrogen, instant- / blast-freeze / -freezing; + 360° open, transparent PC cases; + 360° Sphere, Ball as a Wheel (2D, 3D) “Omniwheel” / “Omni-Wheel”; + from a standing position / standing a 360° turn / rotation in any direction on all objects, including logically and of course vehicles, transporters, robots of all kinds; + flying objects, planes that take off and land vertically, for example by hot air, example fighter jet; + skyscraper up to the weightlessness of space ("space scraper") - "mega construction", “orbital”, etc.; + bendable, flexible, foldable screens / displays / monitors; + bracket / mounting arms for PC, screens (not sure yet), + 360° "ergonomic" bracket / mounting arms; + LED lights (at least its implementation and intended use outside of tv and pc-monitor); RGB LED “Nanoleaf”, etc. + and much more! + I'm not the inventor of VR, but of "AR", AR glass, AR glasses; It was only later that I realized that they derived a lot from my ideas (=inventions), a lot came about that has to do with color and light, through me as an initiator, booster / catalyst, e.g., through my idea, invention of the Screens without a backlight and without a built-in / integrated backlight! It is no coincidence that only afterwards, after I initiated this, they built, built and are building those inventions thanks to my impetus! They sell my ideas (inventions) as theirs! They are not the inventors, but the first technical implementers of my ideas (=inventions)! And those are not the inventors, but the thieves of my ideas (=inventions) initiated by me, because those act as if I wasn't the first hand and the first domino, and they take unjustly, undeservedly a lot of money, stolen money (blood money), recognitions, awards, certificates, fame and history, they boast of my laurels / merits! Before me they all were stucked at LCD, Plasma TV and less lights! I have made a deep impact in evolution! Those manipulate, sabotage, falsify, distort images, paintings, digitized and real, animations, videos, films, also composed of many images, even the publication dates of mine, others and their posts, images, videos, etc.! You have to understand, those can distort everything that can be heard and seen in real life and digitally! They block and delete my pictures, videos, posts, comments, comments-answers and answers! They are poisoning and murdering the world with fake diseases, treatments, "vaccinations"/ "vaccines" and injections by syringe! Now they also make it out as if they haven't been ripping off, cheating, enslaving, murdering other countries with the money currency, money exchange by even -99% for more than two centuries! And as if I'm not the first to disclose that and more! As if I didn't disclose and initiate > 1.00 Ruble (₽) = 1.00 asian Yen (¥) = 1.00 Euro (€) = 1.00 Dollar ($) = any (X) any country < years ago! Each and every non-civilian you hear and see on TV is involved! You can hear and see their > blue blue blue
@oysteinsoreide4323
@oysteinsoreide4323 Рік тому
A big difference in the real time ray tracing vs the long renderings are the amount of rays that you can calculate. The real time ray tracing is an approximation, but good enough for many applications.
@jem5636
@jem5636 Рік тому
It's interesting being able to watch this journey of ray tracing both from the coding and technical side of it, here, and watching it actually being implemented, such as in gaming communities. Admittedly, part of me is also interested in the optimization of code more than anything else. I'd love to see some projects focusing on doing the same things we can already do, but for less processing power and less storage space.
@OurTribeGaming
@OurTribeGaming Рік тому
Is there any chance that these papers/techniques will be available/implemented in blender in future?
@DasherDash
@DasherDash Рік тому
I'm an amateur 3d artist and rendering have always been pulling me back. I always have to make a lot of compromises to get a good enough look in "short" time. Or find fake solutions that will mimic what I want to achieve. Especially if I want to make an animation - I can't wait weeks to render a short video, as it means my PC will be out of use for this whole time. So this technology is salvation, if it will be implemented, it will allow me and many other artists let loose of imagination to create what we want, without fear of weeks of rendering ahead of us. So I agree. What a time to be alive.
@chex6121
@chex6121 Рік тому
Incredible video. The amount of times i have rendered little home made projects and watched videos about rendering i never knew why noise would show up in rendering. But In the first minute of this video I had learned why.
@mastershooter64
@mastershooter64 Рік тому
When I go to graduate school for physics, I wanna start a channel like yours but for pure mathematics and physics papers :D
@ImSquiggs
@ImSquiggs Рік тому
It's impossible to not get invested in something you talk about... you're just so passionate. These videos are really incredible.
@sk8erbyern
@sk8erbyern Рік тому
It is a bit of a turn off imo. I feel like if he was more professional then this video could be 2 minutes long and still be 10 times more informative.
@ImSquiggs
@ImSquiggs Рік тому
@@sk8erbyern Boring contrarianism!
@damysticalone87
@damysticalone87 Рік тому
Those read, listen to / bugged / eavesdrop / overheared / tapped, watched, spied, copied and steal many of my ideas (=inventions) initiated by me: + open, transparent 360° sunlight Buildings, construction methods, architectures, designs, concepts; + step-floor / -storey / -level / -tier pyramid (e.g., as a residential building); + 360° stepped floors / storey / level / tier buildings (constructions, architectures, construction methods, designs, concepts, 2D, 3D, models); + 3D 360° environment (surround), volume, space (room) audio / sound; + 360° environment (surround), volume, room screens (displays), monitors, TVs; + screens (displays), monitors, TVs without backlighting; + many things with magnets like Micro-OLED; + 360° screens, monitors, TVs, panels, glass; + curved screens, monitors, TVs, panels, glass; + 360° reflections and light digital: "ENB", "Ray-Tracing" / "RTX", “Lumen illumination” and whatever renaming! + Short Throw Projectors; + dark backgrounds / themes / skins for windows, browser windows, internet sites, programs / apps, etc.; + fanless, passive cooling; + thermal pads; + heat pipes; heat fins (no sure yet); + 360° heat fins; + Water cooling (unwanted, because it conducts electricity) and + liquid cooling (wanted) for the computer and personal computer / PC application area! - I am not the inventor of refrigeration by nitrogen, instant- / blast-freeze / -freezing; + 360° open, transparent PC cases; + 360° Sphere, Ball as a Wheel (2D, 3D) “Omniwheel” / “Omni-Wheel”; + from a standing position / standing a 360° turn / rotation in any direction on all objects, including logically and of course vehicles, transporters, robots of all kinds; + flying objects, planes that take off and land vertically, for example by hot air, example fighter jet; + skyscraper up to the weightlessness of space ("space scraper") - "mega construction", “orbital”, etc.; + bendable, flexible, foldable screens / displays / monitors; + bracket / mounting arms for PC, screens (not sure yet), + 360° "ergonomic" bracket / mounting arms; + LED lights (at least its implementation and intended use outside of tv and pc-monitor); RGB LED “Nanoleaf”, etc. + and much more! + I'm not the inventor of VR, but of "AR", AR glass, AR glasses; It was only later that I realized that they derived a lot from my ideas (=inventions), a lot came about that has to do with color and light, through me as an initiator, booster / catalyst, e.g., through my idea, invention of the Screens without a backlight and without a built-in / integrated backlight! It is no coincidence that only afterwards, after I initiated this, they built, built and are building those inventions thanks to my impetus! They sell my ideas (inventions) as theirs! They are not the inventors, but the first technical implementers of my ideas (=inventions)! And those are not the inventors, but the thieves of my ideas (=inventions) initiated by me, because those act as if I wasn't the first hand and the first domino, and they take unjustly, undeservedly a lot of money, stolen money (blood money), recognitions, awards, certificates, fame and history, they boast of my laurels / merits! Before me they all were stucked at LCD, Plasma TV and less lights! I have made a deep impact in evolution! Those manipulate, sabotage, falsify, distort images, paintings, digitized and real, animations, videos, films, also composed of many images, even the publication dates of mine, others and their posts, images, videos, etc.! You have to understand, those can distort everything that can be heard and seen in real life and digitally! They block and delete my pictures, videos, posts, comments, comments-answers and answers! They are poisoning and murdering the world with fake diseases, treatments, "vaccinations"/ "vaccines" and injections by syringe! Now they also make it out as if they haven't been ripping off, cheating, enslaving, murdering other countries with the money currency, money exchange by even -99% for more than two centuries! And as if I'm not the first to disclose that and more! As if I didn't disclose and initiate > 1.00 Ruble (₽) = 1.00 asian Yen (¥) = 1.00 Euro (€) = 1.00 Dollar ($) = any (X) any country < years ago! Each and every non-civilian you hear and see on TV is involved! You can hear and see their > blue blue blue
@dantecavallin8229
@dantecavallin8229 Рік тому
There seems to be a loss of texture resolution that you didnt bring up. Other than that its a great video. Look at the beam in 11:46 to see what i mean
@SwervingLemon
@SwervingLemon Рік тому
I saw that and thought "Couldn't I have just used a gaussian blur? Result looks similar..."
@YvesBodson
@YvesBodson Рік тому
I follow your posts with great attention and thank you for communicating on these topics so important in our industry.. Thank you.
@eyal.herlin
@eyal.herlin Рік тому
X Minutes Papers :) Not that I mind. Love your videos and grats on the GTC talk!
@aleksanderniemirka5258
@aleksanderniemirka5258 Рік тому
4:40 I would really like to get a full view of this room as the design is bloody awesome, does anybody know where to download the program/watch a full video of it?
@41A2E
@41A2E Рік тому
Here is the video. There's a watermark that told me what to look up ;) ukposts.info/have/v-deo/oGGZmHeIiGlepJ8.html&ab_channel=BR34K
@michalchik
@michalchik Рік тому
I like your videos but I am sometimes disappointed with how the commentary is heavily weighted towards your emotional responses to the technology rather than the specifics of the technical challenges of the problem, and more particularly how the solution works. If you could give us more education and less how you feel about it, I at least would appreciate that. Thank you for your work
@getsideways7257
@getsideways7257 Рік тому
Indeed, I'm getting hand cramps at this point already because of all the paper squeezing... Jokes aside, too much reused material in this video.
@jensenraylight8011
@jensenraylight8011 Рік тому
he only introduce the paper, if you want an explanation, you should dig your own information on the web. there are a lot of things that won't see the light of the day if it's not introduced. if he explain how the technology work, then nobody gonna watch his channel
@michalchik
@michalchik Рік тому
@@jensenraylight8011 I would totally watch his channel for that reason and there are a bunch of channels like that that are quite popular in other areas that go into deep Dives into technology. Unfortunately there's very little popularization of the Science and Technology of artificial intelligence. There's a large amount of stuff in physics and engineering and a decent amount of stuff in chemistry and biology. They take the latest news and give a not only analysis of what it can do and what it might be useful for, but usually spend the bulk of the time explaining how the new technology works, sometimes incorrectly but you know it's UKposts. I like this channel because it's clear that this wonderful gentleman is smart and does a lot of important reading. I'm just saying that for me personally as much as I enjoy hearing him happy, I wait a lot to hear more explanation of the problem and the solution.
@getsideways7257
@getsideways7257 Рік тому
@@michalchik It takes quite a bit of natural intelligence to comprehend the artificial one, and judging by the majority of commenters here the channel would lose most of its audience if it offered deep insights into the subject instead of the regular "ooh, shiny!" eye candy.
@michalchik
@michalchik Рік тому
@@getsideways7257 lol, perhaps you're right but I think it's worth a try and I think the author of the channel might actually appreciate the response. There are big markets for short and long high quality explanations. On the short side take a look at all the wonderful things that minute physics does for example.
@johntnguyen1976
@johntnguyen1976 Рік тому
What a time to be a ... new sub to Two Minute Papers! LOL. Sooooo much to catch up on (it's been my ENTIRE WEEKEND)! 😍 Thanks for all you do!
@Cyndadogo
@Cyndadogo Рік тому
Thanks Two Minute Papers, your videos are really amazing
@harunizi7506
@harunizi7506 Рік тому
We are right to think that voxel is the future. Good days when light calculations will be better with both physics and pixel graphics in a world filled with atoms instead of mesh. Pixels are powerful because images are also made up of pixels. I think a format such as png, jpeg, etc., where voxels and pixel matrices are compressed, can also be used for voxels.
@Razumen
@Razumen Рік тому
It's not rendering in voxels, they're just used to divide up the scene.
@harunizi7506
@harunizi7506 Рік тому
@@Razumen I didn't say it was. I'm talking about how voxel will be a strong investment for the future. The reason for this is the need for a voxel-based structure in many places. fluid physics, light-per-pixel calculation, etc. that is, dressing dimensional entities by forcing a dimensionless entity such as mesh. instead of weird stuff like uv in mesh technology for example. color, physical information and much more information can be given pixel by pixel. it will be perfect. and yes you are right my comment is a bit off topic in the video.
@xelusprime
@xelusprime Рік тому
I can't help but want Path-tracing in real-time by the day more. Just imagine how things would look. Be mind blowing.
@taloor3137
@taloor3137 Рік тому
Portal RTX has full path-tracing!
@DavyDavePapi
@DavyDavePapi Рік тому
Seems we'll be getting there sooner than we expected I feel. Look at Cyberpunk yberpunk for example
@DavidDavis311
@DavidDavis311 Рік тому
I know nothing of what it takes to render light but I love how excited you are.
@JSena-ff8we
@JSena-ff8we Рік тому
Hope people developing Cycles and EEVEE see these papers.
@TheFinagle
@TheFinagle Рік тому
I feel like in terms of getting it into real time one trick will be to actually put the speed of light into your simulation - that is a real life ray only moves so far in the few ns of a processor tic, so only calculate that much of the path before starting the next cycle. yes you would need very fast processor cycles, and more ram to track it all, but that is something were pushing to build more and more anyway
@Walker-hh7xf
@Walker-hh7xf Рік тому
I think you are correct. Current technique is by tracing the ray from the viewpoint. This is incorrect and is really just cheating, but they say the opposite is "impossible". I disagree, it is possible but just takes longer to compute. What really needs to happen is we need a space-time simulation that emanates the 'light' or photons throughout the cells of the scene, creating a wave format of light. The waves will naturally provide the specular and reflective, refractive properties that light in our real life has. But that's only the first step, we would then need to simulate a lens that can accurately capture these waves that are emanating and bouncing throughout the cells, a virtual model that implements the properties of a camera lens could be a good example. The limitations of this approach is immense, but it is the correct approach, and if we wish to truly have life like real time rendering, the hardware needs to address these limitations. Memory, computational power. Maybe someday when we have consumer level quantum mechanical machines, but for now we can start on the software side of things by creating slow prototypes.
@TheFinagle
@TheFinagle Рік тому
You dont even need to do the whole scene. For animations you know where the camera will be in the next several frames, you can do a quick "What will it be able to see" outgoing ray trace with say 1000 rays (instead of the normal 8.3 million per frame) and back track using the speed limited version in the high detail levels 4k rendering needs. In a video Game you could cut it down to what positions could the camera reach within the back track time of the longest ray. The player can only move so fast, and turn so much per frame. Theres a finite amount of scene they will be able to look at. This is still a ways away from being game engine ready, but the unpredictability of a physics driven and user interactive scene combined with the very tight deadline of having to wait for physics engine and still finish within 1 / 120th of a second 120 times a second is a much more challenging.
@TheRidiculousR
@TheRidiculousR Рік тому
i just wish this feature will make its way to blender, i am more then OK with the shortcommings if it means my renders will be any faster, and it could improve viewport clarity as currently you get some really funky artifacts in poorly lit scenes.
@Ortagonation
@Ortagonation Рік тому
this actually a very great research. The problem of water light refraction is not the noise, but the fireflies (like noise but it is the artifacts of bounching light). Until now there is no solution than tricking the scene. But now this, amazing...
@upolpi3171
@upolpi3171 5 місяців тому
I think it's one of your most underrated videos! People can truly bring such a level of vividness and clarity to the subject when they're actual qualified experts on the topic. I love how this is basically an ode to a cool algorithm, focusing on as much detail as UKposts's format would allow instead of glossing over it completely like in OpenAI's recent marketing brochures of papers. Wish you the best of luck!
@protocol6
@protocol6 Рік тому
By the way, Martin of Wintergatan referenced your catchphrase in his latest Marble Machine X video where he's testing repeatability of a marble gate and making improvements.
@badpuppy3
@badpuppy3 Рік тому
I always enjoy the enthusiasm in your presentations.
@sinonimo8719
@sinonimo8719 Рік тому
Saw this on the stream of DotCsv with Asier, glad to see you there ;)
@IuliusPsicofactum
@IuliusPsicofactum Рік тому
This video could be reduced 80% if we exclude all the awe and amazement expressions from Károly, BUT these papers improvement clearly deserve each one of them!
@12q8
@12q8 Рік тому
For the missing paths, the fireflies, can these spots be estimated by having the geometric position of that spot averaged with the values near it? Similar to anti-alising algorithms but with light rays?
@anlumo1
@anlumo1 Рік тому
That's basically the beginning of a denoising algorithm.
@thomaswalther7276
@thomaswalther7276 Рік тому
To hear so much enthusiasm for the topic in the voice of the narator cheered me up 😀
@RC_Engineering
@RC_Engineering Рік тому
My favorite part is that Ren from Ren and Stimpy is summarizing these papers for us.
@Sssseytha
@Sssseytha Рік тому
oh man this is epic news for CG artists, cannot wait to speed up my workflow!!! what a time to be alive
@nettsm
@nettsm Рік тому
Didn't we see this video about a year ago? Why are we seeing this again? 🤔
@michaelbiljon8249
@michaelbiljon8249 Рік тому
A 16 minute Two Minute paper. You know it's going to be good.
@elgentidwell9407
@elgentidwell9407 Рік тому
in 2 papers we wont even need denoising! This is nuts and I absolutely love it!!
@007LvB
@007LvB Рік тому
I love you, man! Such a great work sharing this. We truly live in wondrous times :))
@Rwdphotos
@Rwdphotos Рік тому
Why does it feel like the last five videos from this guy are basically the same video?
@andrewpoison
@andrewpoison Рік тому
Dude :D As much as I love your enthusiasm about it, your seemingly random emphasis and speaking with an invisible comma after every second word hurt my ears terribly ^^ Still thank you for the video ;)
@Tommo_
@Tommo_ Рік тому
I'm more impressed at how impressed the narrator sounds
@yabdelm
@yabdelm Рік тому
I wish there was no recap of all the previous videos he has done... it's called 2 min papers!
@garyhamilton2104
@garyhamilton2104 Рік тому
I'm so glad that you're becoming so popular in the field(s). Is there any way to invest in your success?
@TwoMinutePapers
@TwoMinutePapers Рік тому
You are very kind, thank you! Just being with us and watching the series is an honor for us already. No need to do more - thank you! 🙏
@olli3686
@olli3686 Рік тому
This is mindblowing!!! Honestly, render costs/times is the reason why I quit working in 3D art.
@egretfx
@egretfx Рік тому
duh
@facts9144
@facts9144 Рік тому
@@egretfx duh
@xxxod
@xxxod Рік тому
@@facts9144 ok
@Bleepurchin
@Bleepurchin Рік тому
It’s been pretty easy to render complex scenes for a while now… unless you quit over 10 years ago…
@user-mf3oc6mj5l
@user-mf3oc6mj5l Рік тому
One computer graphics advancement that I hope will be enabled by ray tracing is vector geometry. Since we aren't doing computations per polygon anymore but per light ray, we can potentially have "infinite" number of polygons, i.e. smooth vector 3d models.
@tantumDicoQuodCogito
@tantumDicoQuodCogito Рік тому
Great work, congratulations!!
@sebbes333
@sebbes333 Рік тому
*Why not "solve backwards" instead?* Set the default "un-traced" space to either some impossible color (eg: 100% transparent), or (bright pink, or the opposite of the nearest "traced" pixels) something. Then just scan for one of these "impossible areas", and shoot rays from the camera towards that area & see where they end up. (I bet this has already been tried, so-) Why wouldn't this work?
@davidmcgill1000
@davidmcgill1000 Рік тому
That'd get you to the first bounce of light, but can't handle the interaction of all sources of lights interfering with every pixel on screen. Basically what pixel shaders have done for the past 20 years. The benefit of this is getting actual diffuse lighting.
@teslaromans1023
@teslaromans1023 4 місяці тому
Love the enthusiasm in this video 😃, however it’d be nice if you went into more details about how do these new gamechangers work and manage to improve performances
@NoPantsBaby
@NoPantsBaby Рік тому
I understood nothing but this dude being so excited about stuff was compelling to watch.
@ChrisGuerra31
@ChrisGuerra31 Рік тому
That was a lot of triple yesses! 😃 Squeezing that paper! 📜
@dieblauebedrohung
@dieblauebedrohung Рік тому
Man the video was interesting, but the way he talked was hard to bear. Having a pause after every second word and then dragging every thrid word for half a second is an interesting rhythm perhaps, but hard to follow without going insane. I do get language barrier, but with a Script and Preparation, this should be possible to overcome. I'm not a native speaker either.
@GeorgeAlexanderTrebek
@GeorgeAlexanderTrebek Рік тому
Yeah reading the comments it seems many non-native English speakers have trouble understanding/listening to his voice. I personally have watched many of his videos and didn't even realize anyone had this problem. For me its extremely clear and understandable, but English is my first language so that may be why. I can maybe see his tone/pauses being confusing or jarring for someone non-native....
@dieblauebedrohung
@dieblauebedrohung Рік тому
@@GeorgeAlexanderTrebek I didn't say confusing, or that I didn't understand it, his Pronounciation was very clear after all; it was simply Torture for the Mind. The Language itself was not the Problem, it's his speech Patterns whose Weirdnesses I attribute to his probable Migration Background / English Unfamiliarity. I did say "hard to follow" but only if you look at the Words in a Vacuum, I said "hard to follow without going insane"^^ So yeah, I learned a lot from the Video, just at the Cost of some Mental Stability that Evening lol If someone was talking like that in my own Language I'd be equally annoyed.
@MichaelPohoreski
@MichaelPohoreski Рік тому
He talks like. William Shatner. from Star Trek. And is. Annoying to. Follow along. With all the. Unnecessary pausing. :-/
@GeorgeAlexanderTrebek
@GeorgeAlexanderTrebek Рік тому
@@dieblauebedrohung hmm interesting, that's fair. I never really said That YOU said confusing or didnt understand, more that many non native English speakers have these issues just by reading the comments. I also said that it might be hard to listen to or be jarring for non native speakers, which is what you mentioned after about him having irregular speak patterns. I watch his videos at 1.1 or 1.2 speed so less noticeable for me anways....
@dieblauebedrohung
@dieblauebedrohung Рік тому
@@GeorgeAlexanderTrebek Okay, fair enough. Wording lul. But yeah, I sped it up as well and that helped a lot.
@luizcosta8122
@luizcosta8122 Рік тому
I'm not mad the papers are taking longer than 2 minutes. This is good.
@The.Pickle
@The.Pickle Рік тому
We have recently seen the rise in A.I generated images; do you think we could achevive procedurally generated enviroments using A.I? Is there any similarity between Ray Tracing software and programs used to simulate particle physics?
@mercyveritas1125
@mercyveritas1125 Рік тому
Can u use a different accent in your text-to-speech program?
@Kapomafioso
@Kapomafioso Рік тому
No need to be dicky. Dude's from Hungary.
@ramonperezsanchez
@ramonperezsanchez Рік тому
Stimpy does one hell of a voiceover in these. Kuddos!
@theLUCKYdude17
@theLUCKYdude17 Рік тому
Loved the enthusiasm. Truly a curious future we have
@sh_4016
@sh_4016 Рік тому
I found out at 13:25 that the new method has some more troubles calculating specular rays but is better with reflection and refraction rays but still it has much less Noise than the previous method but if we also overlay a view of the previous frames it would have even less Noise and also some Motion blur so it looks more natural also NVIDIA could use a seed function to remove even more flickering instead of using a random number each time each frame
@FlareSound
@FlareSound Рік тому
I really like the software advancements NVIDIA is doing, but sometimes I can't help but wonder about their hardware, Especially with the new RTX 40 line. I hope they can address the fact you need a 700W+ PSU just to run a graphics card..
@guillermojperea6355
@guillermojperea6355 Рік тому
No, you don't; it's only on a halo product, and the consumption increase is the new normal if you want to reach that level of processing capabilities. If it could be helped, it would, but limitations in Moore's law and costs aren't helping.
@Wobbothe3rd
@Wobbothe3rd Рік тому
The rtx4090 doesn't actually run at much higher power levels than the 3090.
@JorgetePanete
@JorgetePanete Рік тому
​@@guillermojperea6355 It's Nvidia's fault
@andrasbiro3007
@andrasbiro3007 Рік тому
You are just too young. I remember when we laughed a the Pentium CPU, because it required cooling. It was cooled by a tiny fan, little less than 2 inches. Before that CPUs didn't have cooling at all, not even passive. And that also meant that power consumption was only a few watts. Power and cooling requirements have been increasing ever since. In exchange you get even higher performance. It's possible to turn down the voltage and frequency a bit to use less power, and need less cooling, but you get less performance, and it doesn't reduce the cost of the card significantly, because you are mostly paying for the silicon. Also 1000W+ PSUs aren't that expensive anymore. At least not compared to a 4090.
@drb0mb
@drb0mb Рік тому
that stuff at like 12:45 is comparable to waiting for the DVD logo to bounce perfectly in a corner
@mndlessdrwer
@mndlessdrwer Рік тому
I'd be curious to learn how nVidia's ray-tracing papers and their applications differ from or are related to those which led to AMD and Intel's ray tracing support. nVidia uses their own proprietary tensor cores to handle the vast majority of the ray-tracing calculations, so I'm curious if AMD and Intel have simply adapted their process and algorithms to more abstracted instruction sets for their conventional graphics processor unit modules or if they have had to create their own method for calculating the ray-traced light dynamics for their scenes. It's a really cool subject and I'd love to see a deep-dive video about it sometime with analysis on how these different platforms have managed to get ray-tracing to work for them and how the different iterations of the process have changed how the image is formed.
@hikmetyalcnkaya4162
@hikmetyalcnkaya4162 Рік тому
aAand, sOooo, thEeen
@rb8049
@rb8049 Рік тому
Can’t the standard error be calculated at each point? If the standard error is large, then the Ray is replaced by a diffuse light source. ?
@kylebowles9820
@kylebowles9820 Рік тому
ReSTIR does much more by checking neighbors in space and time as well as a "super bias" fallback like you suggest
@oNovais
@oNovais Рік тому
I just wish the render engines can access these tecnologys, what octane and redshift for example do today already is amazing. But implementing these with so dslrr3 tech would be smazing for animations workflow
@Zettoman
@Zettoman Рік тому
i remember starting blender in 2009 on a c2duo and rendering overnight what now takes seconds. it's just gonna get better.
@cptairwolf
@cptairwolf Рік тому
Why are you re-posting this video? You showed this one months ago now?
@AndreeMarkefors
@AndreeMarkefors Рік тому
Would it be. Possible. Tooo. Narrate this. In a waaay. That is. Human under. Standable? Oh, my! Yes, yes and yes.
@08wolfeyes
@08wolfeyes Рік тому
I love seeing the improvements being done, great work. As an artist myself, these changes will help in so many ways. Question: While Ray-Tracing is impressive, isn't Path-Tracing better, doesn't it give more realistic results? If so then is the reason it's not used to much in real-time due to the time it takes to calculate the light rays??
@theneonbop
@theneonbop Рік тому
Path tracing is often referred to as ray tracing. These papers mentioned are all variants of path tracing.
@456MrPeople
@456MrPeople Рік тому
Ray tracing in the research sense refers to path tracing where the light rays originate from the light source and bounce around the scene until they hit the camera. Ray tracing in the graphics sense refers to light rays originating from the camera that bounce around the scene and end at the light source, which is computationally faster but less accurate than path tracing. In fact rasterization is technically also ray tracing, but with a single ray originating from the camera that ends when it hits an object and doesn't bother finding a light source.
@xygomorphic44
@xygomorphic44 Рік тому
Take a drink every time he says "Wow! Looks at this scene"
@JoneNascimento
@JoneNascimento Рік тому
Caustics is exponentially bouncing lights. There is no way we can bruteforce this, we need to make shortcuts if we want it in real time.
@kylebowles9820
@kylebowles9820 Рік тому
They lost quality on the bunny ears and significant energy in specular reflections. Reflections are smudgy and jumpy over time from undersampling in temporal domain. Eventually I'd like to see image based BRDFs as well so the materials can get some more variety.
@C.I...
@C.I... Рік тому
I would rather see advances looking like camera noise than this blurry mess honestly.
@kylebowles9820
@kylebowles9820 Рік тому
@@C.I... Check out the grain produced by a proper spectral path tracer with a camera model
@freecode.ai-
@freecode.ai- Рік тому
Keep up the great work!
@SianCKB
@SianCKB Рік тому
wait I swear I have seen this video before! is this a reupload? :D great video nonetheless!
@advance512
@advance512 Рік тому
Can't wait till we get sharper, denoised, real time ray traced legs in the metaverse. It will be so immersive!
@Deezsirrer
@Deezsirrer Рік тому
Stop with the reuploading I saw this script 4th times now
@l3nn13
@l3nn13 Рік тому
sheesh you really hype me up with these papers xD
@silvenshadow
@silvenshadow Рік тому
So amazing! Next up, solid modelling for real physics and deformation in real-time!! This is something that I always dreamed about as a kid. Maybe when I'm old it will run on my phone. :)
@Akiyamka
@Akiyamka Рік тому
It was so hard to listen...
@c0xb0x
@c0xb0x Рік тому
He refuses to get a speech therapist. :(
@segamai
@segamai Рік тому
I’m so sorry but the way this guy talks grates on my nerves like nothing else jesus christ
@cohenschellenberg3325
@cohenschellenberg3325 Рік тому
i disagree, his voice isnt bad at all
@10001000101
@10001000101 Рік тому
Each of the newer techniques lost detail from the old version, it's like viewing a game without AA then using FXAA to smooth out the edges.
@ten0re0re
@ten0re0re Рік тому
Awesome, thanks for the video. Do you know to what extend is quantum mechanics utilized in these solutions? The original noisy technique looks to me like it only considers classical paths for all light rays, I wonder how much of this can be replaced with randomness to simulate nature's own optimization technique.
@AkitaMix
@AkitaMix Рік тому
Those light rays are likely generated randomly cause doing them individually and specifically all the time would eat up memory.
@enriquegato
@enriquegato Рік тому
The solution is impressive. The sample scenes are terrible. Unless you understand the value of what you're looking at most of them look like an old videogame.
@spirosfoufoutos6241
@spirosfoufoutos6241 Рік тому
I like these 2 minute papers but I can't stand the hyperbolic tone for EVERY paper. Chill out man..
Is RTX a Total Waste of Money?? - Can we even tell when it's on?
15:10
Linus Tech Tips
Переглядів 3,9 млн
This Is Ray Tracing Supercharged!
8:17
Two Minute Papers
Переглядів 147 тис.
КТО СМОГ ПОБЕДИТЬ?😳
00:36
МЯТНАЯ ФАНТА
Переглядів 987 тис.
How Ray Tracing (Modern CGI) Works And How To Do It 600x Faster
32:06
Josh's Channel
Переглядів 525 тис.
Unreal Engine 5.4: Game Changer!
5:57
Two Minute Papers
Переглядів 95 тис.
Destroying My RTX 4090 With Incredible Minecraft Graphics
8:11
AsianHalfSquat
Переглядів 3,3 млн
Wow, A Simulation That Looks Like Reality! 🤯
11:01
Two Minute Papers
Переглядів 271 тис.
DeepMind AlphaFold 3 - This Will Change Everything!
9:47
Two Minute Papers
Переглядів 20 тис.
NVIDIA Did It: Ray Tracing 10,000 Times Faster!
5:46
Two Minute Papers
Переглядів 338 тис.
NVIDIA REFUSED To Send Us This - NVIDIA A100
23:46
Linus Tech Tips
Переглядів 9 млн
How do Video Game Graphics Work?
21:00
Branch Education
Переглядів 2,9 млн
I made a better Ray-Tracing engine
17:38
NamePointer
Переглядів 239 тис.
EA’s New AI: Next Level Gaming Animations!
6:18
Two Minute Papers
Переглядів 249 тис.
Result of the portable iPhone electrical machine #hacks
1:01
KevKevKiwi
Переглядів 8 млн