NeRF: Neural Radiance Fields

Поділитися
Вставка
  • Опубліковано 1 сер 2024
  • NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis
    Ben Mildenhall*, Pratul P. Srinivasan*, Matthew Tancik*, Jonathan T. Barron, Ravi Ramamoorthi, Ren Ng
    *denotes equal contribution
    Project Page: www.matthewtancik.com/nerf
    Paper: arxiv.org/abs/2003.08934
    Code: github.com/bmild/nerf
  • Фільми й анімація

КОМЕНТАРІ • 124

  • @suharsh96
    @suharsh96 4 роки тому +154

    What a time to be alive !

    • @blackbird8837
      @blackbird8837 4 роки тому +2

      Everyone would think that during the time they live in as only so much as currently released can be drawn from as reference to make this statement.

    • @IRONREBELLION
      @IRONREBELLION 4 роки тому +14

      Dr Károly Zsolnai-Fehér
      ?!?!

    • @EnriquePage91
      @EnriquePage91 4 роки тому +13

      Dear Fellow Scholars !

    • @EnriquePage91
      @EnriquePage91 4 роки тому +1

      IRON REBELLION literally same lol

    • @IRONREBELLION
      @IRONREBELLION 4 роки тому +1

      @@EnriquePage91 I love him so much haha

  • @AndreAckermann
    @AndreAckermann 4 роки тому +17

    This is amazing! The implications for 3d and compositing workflows alone are mind-boggling. Can't wait to see this filter through to mainstream products.

  • @DanielOakfield
    @DanielOakfield 4 роки тому +4

    Absolutely exciting time ahead, thanks for sharing Matthew!

  • @bilawal
    @bilawal 4 роки тому +3

    2:45 🔥 elegant approach and game changing results - handles fine scene detail and view dependent effects exceedingly well

  • @johnford902
    @johnford902 4 роки тому +5

    It’s awesome to live at this time and be participating in history.

  • @Paul-qu4kl
    @Paul-qu4kl 4 роки тому +49

    Progress in photogrammetry has been very slow in recent years, hopefully this approach will yield some new developments. Especially better reflection and translucency capture, which has so far been huge problem, looks like a possibility based on the shown scenes. Very exciting, especially with VR equipment becoming more prevalent. Can't wait to capture real 3D scenes (albeit without motion), rather than stereoscopic video.

    • @technoshaman001
      @technoshaman001 Рік тому +1

      ua-cam.com/video/Pb-opEi_M6k/v-deo.html this is the latest update 3 years later! pretty amazing IMO lol

  • @wo262
    @wo262 4 роки тому

    Super cool. This will be super useful to sythesize light fields. A lot of people are saying this is not practical right now because it takes seconds to infer one POV. But real time visualization of light field data already exist so you could store precalculated inference that way

  • @directorscut4707
    @directorscut4707 Рік тому

    Mind Blowing! Cant wait to have this in google maps or VR implemented and explore the world!

  • @user-tt2qn1cj1x
    @user-tt2qn1cj1x Рік тому

    Thanks for sharing and also mentioning the other contributors to NeRF creation and development.

  • @blender_wiki
    @blender_wiki 4 роки тому

    Outstanding work. Congratulations

  • @5MadMovieMakers
    @5MadMovieMakers 2 роки тому +1

    Looks neat!

  • @parkflyerindonesia
    @parkflyerindonesia 4 роки тому +2

    Matt, this is gorgeous! You guys are genius!!! Stay safe and stay healthy guys 👍

  • @trollenz
    @trollenz 4 роки тому

    Brilliant work ! ❤️

  • @chucktrier
    @chucktrier 4 роки тому +1

    This is insane, really nice work

  • @ScienceAppliedForGood
    @ScienceAppliedForGood 3 роки тому

    This looks very impressive, progress here seems on the same level when GANs where introduced.

  • @jwyliecullick8976
    @jwyliecullick8976 2 роки тому

    Wow. The utility is constrained by the images used to feed the neural network, which may not reflect in varied model environmental factors. If you have images of a flower on a sunny day, rendered in a cloudy day scene, they will look realistic -- for a sunny day. Anything short of raytracing is cartoons on a Cartesian canvas. This is an amazing technique -- super creative application of neural nets to imagery data.

  • @huanqingliu9634
    @huanqingliu9634 2 роки тому

    A seminal work!

  • @marijnstollenga1601
    @marijnstollenga1601 4 роки тому

    Amazing stuff!

  • @letianyu981
    @letianyu981 3 роки тому

    Dear Fellow Scholars !
    This is two minutes paper with Dr Károly Zsolnai-Fehér
    ?!?!
    What a time to be alive !

  • @suricrasia
    @suricrasia 4 роки тому

    this is astounding

  • @Fodi_be
    @Fodi_be 4 роки тому

    Astonishing.

  • @gusthema
    @gusthema 4 роки тому +2

    This is amazing!!! congrats!! are you publishing this model on TF Hub?

  • @ImpMarsSnickers
    @ImpMarsSnickers 4 роки тому

    This thing will allow to create slowmotion from a simple video, and a SUPER SLOWMOTION from a slowmotion!
    And also stabilize a shaking camera.

  • @antonbernad952
    @antonbernad952 Рік тому

    While the hot dogs were spinning at 1:58, I got really hungry and had an unconditional craving for hot dogs. Still nice video, thanks for your upload!!!11OneOneEleven

  • @GdnationNY
    @GdnationNY 4 роки тому +1

    Stunning! Are these vector points, volume rendered point clouds? These fly throughs are image sequences?

    • @hecko-yes
      @hecko-yes 4 роки тому

      from what i can tell it's kinda like a metaball except instead of a simple distance function you have a neural network

  • @raycaputo9564
    @raycaputo9564 4 роки тому

    Amazing!

  • @martinusmagneson
    @martinusmagneson 4 роки тому +13

    Great work! Could this also have an application in photogrammetry?

    • @wandersgion4989
      @wandersgion4989 4 роки тому +6

      martinusmagneson From the look of it, this could be used as a substitute for photogrammetry.

    • @juicysoiriee7376
      @juicysoiriee7376 4 роки тому +2

      I think this is photogrammetry?

    • @Jianju69
      @Jianju69 Рік тому +1

      @@juicysoiriee7376 It is in the sense that it converts a set of photographs to a 3D scene, yet it does *not* create a 3D model in the conventional (polygonal) sense.

  • @AdictiveGaming
    @AdictiveGaming 4 роки тому +6

    Can't believe what I just saw. Is there a chance you will make some feature videos? Like is it a real 3d model inside there? Would it be then possible to be imported to 3d software? How the materials are made? How can everything be so perfectly detailed and sharp?and so on and so on

    • @hydroxoniumionplus
      @hydroxoniumionplus 4 роки тому

      Bro, literally just read the paper if you are interested.

  • @joshkar24
    @joshkar24 4 роки тому

    Can this be used to add head movement interactive option to vr movies? Would require a bunch of video cameras and expensive computer crunching beforehand, and how to store/stream the finished data? Or is that a use case where traditional real-time 3d engines are a better fit? Or some hybrid blend?

  • @BardCanning
    @BardCanning 4 роки тому

    OUT
    STANDING

  • @DanFrederiksen
    @DanFrederiksen 3 роки тому

    Nice. The two input angles are screenspace x y coords? and the x y z is the camera position in the training? how do you extract the depth data from the simple topology then?

  • @erichawkinson
    @erichawkinson 3 роки тому +1

    Can this method be applied to stereoscopic equirectanular images for use in VR headsets?

  • @YTFPV
    @YTFPV Рік тому

    Amazing stuff i need to wrap my head on how the depth is generated at 3:22 with the Christmas tree ? I am working on movie where we had to generate depth from the plate and we use all the tool in book but it's always flickering pretty bad never has nice. How would i use this if it's possible?

  • @ArturoJReal
    @ArturoJReal 4 роки тому +1

    Consider my mind blown.

  • @azimalif266
    @azimalif266 4 роки тому

    This will be Awesome for games.

  • @unithom
    @unithom 4 роки тому

    If the new iPads have LIDAR and a sensitive enough gyroscopic sensor - how long before this method can be used to capture objects? (Within 15’ radius of scenes, anyway)

  • @nhandexitflame8747
    @nhandexitflame8747 3 роки тому +1

    how can i use this? i coudlnt find anything so far. please hjelp!

  • @jensonprabhu7768
    @jensonprabhu7768 4 роки тому +1

    Wow it's cool..say for example, to capture an average human just standing. How many pictures are required to capture enough detail?

  • @damagedtalent
    @damagedtalent 4 роки тому

    Incredible is there anyway I can do this at home?

  • @alfvicente
    @alfvicente 4 роки тому +61

    Sell it to Google so they can render 3d buildings properly

    • @nonameplsno8828
      @nonameplsno8828 4 роки тому +6

      its available for free,they might use it on street view and create a real 3d google planet thing.

    • @Romeo615Videos
      @Romeo615Videos 3 роки тому +2

      @@nonameplsno8828 wish there was a step by step tutorial to try this with

    • @hherpdderp
      @hherpdderp Рік тому

      @@Romeo615Videos there is now but you have compile it yourself.

    • @roshangeoroy
      @roshangeoroy 10 місяців тому

      Check the paper. It's under Google Research.

  • @josipkova5402
    @josipkova5402 Рік тому

    Hi this is really interesting. Can you tell me maybe how much costs one rendering of about 1000 photos? Which program is used for that? Thanks :)

  • @barleyscomputer
    @barleyscomputer 5 місяців тому

    amazing

  • @ZachHixsonTutorials
    @ZachHixsonTutorials 4 роки тому +8

    So is this actual 3D geometry, or is it just the neural network "interpolating," for lack of a better word between the given images?

    • @GdnationNY
      @GdnationNY 4 роки тому

      These depth maps could be the next step to a mesh possibly?

    • @ZachHixsonTutorials
      @ZachHixsonTutorials 4 роки тому

      @@GdnationNY The thing I'm curious about is if there are any ways to translate the material properties to a 3D object. The program seems to understand some sort of material properties, but I'm not sure if there is a way to translate that

    • @TomLieber
      @TomLieber 4 роки тому +6

      It's the "neural network interpolating." Each point in 3-D space is assigned an opacity and view-specific color by a neural network function. It's rendered by integrating that function over each ray. So it doesn't model light sources, their effects are just baked into everything the light touches. You could get a mesh by marching over the function domain and attempting to disentangle the lighting effects, but it'd take some doing.

    • @ZachHixsonTutorials
      @ZachHixsonTutorials 4 роки тому +1

      @@TomLieber That's what I was thinking, but there is that one section of the video where the reflections on the car are moving, but the camera is not. That part kind of made me wonder

    • @TomLieber
      @TomLieber 4 роки тому +2

      ​@@ZachHixsonTutorials You can do that by projecting the rays from your actual camera position, but evaluating the neural network with the direction vector from a different camera viewpoint.

  • @ImpMarsSnickers
    @ImpMarsSnickers 4 роки тому

    Glass, Car window and interiors, Reflections, only 20-50 photos... perfect scan result...
    ... On a second thought, I think I get the idea, I realized camera path walks only between frames shot by camera, and it's not so much about making it in 3d, it's about how light works, and a fast render result. It would be great for movies to erase things in foreground, leaving visible what is behind those things that pass in front of camera! In that case it's a great project :)

  • @thesral96
    @thesral96 4 роки тому +4

    Is there a way to try this with my own inputs?

    • @Jptoutant
      @Jptoutant 4 роки тому +1

      archive.org/details/github.com-bmild-nerf_-_2020-04-10_18-49-32

  • @omegaphoenix9414
    @omegaphoenix9414 4 роки тому +2

    Can we get some kind of tutorial, more in-depth as to how to do this on our own? Can this be implemented into a game engine while still keeping reflections? Is this cheaper on performance than computer generated screen space reflections? I actually shivered watching this from how insane it looks. I've been into photogrammetry for quite some time now (I know it's not the same) and I would love to try and replicate this for myself as soon as possible

    • @omegaphoenix9414
      @omegaphoenix9414 4 роки тому

      I may have misinterpreted what this does, but if so, can this be used in games?

  • @davecardwell3607
    @davecardwell3607 4 роки тому

    Very cool

  • @HonorNecris
    @HonorNecris 2 роки тому +2

    So with NeRF, how does the novel view actually get synthesized? I think there is a lot of confusion lately with these showcases as everyone associates them with photogrammetry, where a 3D mesh is created as a result of the photo processing.
    Is each novel view in NeRF created per-pixel based on an algorithm and you are animating the resulting frames of these slight changes in perspective to show 3 dimensionality (the orbital motion you see), or is a mesh created that you are moving a virtual camera around to create these renders?

    •  2 роки тому +3

      It's the first. No 3D model is created at any moment.
      You have a function of density wrt to X,Y,Z though, so even though everything is implicit, you can recreate the 3D model from it. Think of density as "somethingness" from which we can probably construct voxels. TO get a mesh is highly non-trivial though
      This is kinda what they are doing when showing a depth map, they probably integrate distance with density along the viewing ray.

  • @Nickfies
    @Nickfies 3 місяці тому

    what exactly is theta and phi respectively? is one the rotation around vertical axis and phi the tilt?

  • @russbg1827
    @russbg1827 3 роки тому

    Wow! This means you can get parallax in a VR headset with a 360 video from a real environment. I was sad that wouldn't be possible.

  • @driscollentertainment9410
    @driscollentertainment9410 4 роки тому

    I would love to speak with you about this!

  • @musashidanmcgrath
    @musashidanmcgrath 4 роки тому +6

    Incredible work! I'm sure this will be coming to Blender in the future - I spotted the Blender shader balls. :D I'm assuming your team have been using Blender to extract the geometry, especially considering this is all open source Python.

  • @piotr780
    @piotr780 Рік тому

    3:00 how this animation on the right is produced ?

  • @dewinmoonl
    @dewinmoonl 3 роки тому

    cool research

  • @ONDANOTA
    @ONDANOTA 3 роки тому

    Are radiance fields compatible with 3d editors like Blender?

  • @TheAudioCGMan
    @TheAudioCGMan 4 роки тому

    oh my!

  • @antonbernad952
    @antonbernad952 Рік тому

    Nice video, thanks for your upload!!11OneOneEleven

  • @BOLL7708
    @BOLL7708 4 роки тому +3

    Now I want to see this in VR 😅 Is it performant enough for real time synthesizing at high resolution and frame rate? Sure puts prior light field techniques to shame 😗

    • @dkone165
      @dkone165 4 роки тому +2

      "On an NVIDIA V100, this takes approximately 30 seconds per frame"

    • @BOLL7708
      @BOLL7708 4 роки тому +1

      @@dkone165 Ah, so I guess while not real time, it could be used to generate an image set to be used in a real time environment, although it could turn out just impractical. At some point we'll have an ASIC for this, I'd by an expansion card for it 😅

    • @fidel_soto
      @fidel_soto 4 роки тому

      You can create scenes and then use them in VR

    • @dkone165
      @dkone165 4 роки тому

      "The optimization for a single scene typically take around 100-
      300k iterations to converge on a single NVIDIA V100 GPU (about 1-2 days)"

    • @Jianju69
      @Jianju69 Рік тому +1

      @@fidel_soto These "scenes" are not conventional 3D-model hierarchies. Rather, they are 4D-voxel sets from which arbitrary views can be derived within around 30 seconds per frame. Though some work has been done to "bake" these scenes for real time viewing, the performance still falls far short of being suitable for VR. Perhaps a robust means of converting these NeRF-scenes to high-quality 3D models will become available, yet we already have photogrammetry for that task.

  • @WhiteDragon103
    @WhiteDragon103 4 роки тому +2

    DUDE WHAT

  • @Jigglypoof
    @Jigglypoof 4 роки тому +18

    So I can finally make my 2D waifus into 3D ones?

    • @loofers
      @loofers 3 роки тому +1

      ​@Hajar Babakhani i don't think you understand. this would be able to take a drawing, if done well enough with proper shading, and extract full 3d elements. just google "3d cartoon art" or "3d drawings" and this would theoretically be able to render or paint an alternate angle of an object based on the existing, handdrawn view of that object. obviously with only 1 view, hidden features would be just that; hidden; but again here ai could be used to fill in "possible options" for optional details. AI is definitely advancing insanely fast, so fast that if an ai can piece together the pieces to gain understanding of those objects it is seeing, understand interrelationships between those objects, words, etc, we might actually see general AI within 5 years. i personally think that ai just needs a sort of secondary (or tertiary, i suppose) "overviewer ai" which would be watching the process of the ai adjusting its gan settings, and would tweak them to improve itself/match what it observes in the real world, and which could comment to itself on its own code, change its own code. i think we may need to be very very careful in these next few years in terms of restricting datacenter-level ai research. cause all it would take is an ai getting some money and access to remote mineshaft with robots that can weld and assemble small parts, and you have some real potential issues. it ships some cpus to those mines, and bam, we're in terminator 2 territory :p (plz be kind to me ai, i'm chill with you being god of the universe!) we've already given ai the ability to see, we're starting to give it the ability to draw things it saw, next we just need to get it the ability to dream/imagine/adjust itself and we're in scifi territory. i think a really good example of where this kind of view synthesis will be maybe applied in the future, is in both video compression and upsampling of older or lower quality video, and 3d scene extraction from movies and the like.
      take a look at this:: ua-cam.com/video/AwmvwTopbas/v-deo.html . here you see a sample of some 8k video upsamples that show real qualitative detail improvement; but you can also see the dot/spec/low image quality of the lich king model/character (at the time index in the video and a few seconds after that time index). if an ai would be able to grasp the semantics of the scene, the possiblity of object permanence etc, it could infer from the earlier content in the video that the dot there in the image is the character (the lich king, in this world of warcraft cinematic example) - through rules of object permanence, it could estimate that the spec/smudge in the scene is the lich king, simply viewed from another angle/distance, and thus convert the entire scene drawn herein into a fully 3d rendered construction, in higher resolution, or approximating the resolution of the source cad/blender file data.
      it could at least do that after seeing the clip in its entirety, or with human intervention - but i think ai will get to where it can do it completely unaided, and probably quite soon (sub 3 years).
      while this kinda sounds a bit scifi, 2-3 years ago i'd have said that the stuff we are looking at now in this NERF video is potentially doable by ai with extremely intelligent programmers/mathematicians steering the process, and look at where we are.
      Matthew, you guys are literally creating technological magic. amazing work.

    • @Binary_Omlet
      @Binary_Omlet 3 роки тому +2

      Yes, but she still won't love you.

  • @hherpdderp
    @hherpdderp Рік тому

    Am I understanding correctly that what you are doing here is rendering the nodes of neural network in 3d? If so I wonder if it could have non CG uses?

  • @spider853
    @spider853 Рік тому

    how was it train?

  • @maged.william
    @maged.william 4 роки тому +2

    @ 3:59 How the hell did it model glass!!

    • @ImpMarsSnickers
      @ImpMarsSnickers 4 роки тому +1

      I was thinking the same, also car windows with interior, reflections, and 1:58 where the hell they scanned real life Blender's Material Spheres?

    • @TomLieber
      @TomLieber 4 роки тому +2

      I'd love to know! I wish we could see where in space the reflections are being placed. Ideally, glass would be modeled as transparent at most angles except those with specular reflections, but the paper says that they constrained density to be a function of position only, so does that mean that in the model, glass is opaque and has the view through the glass painted onto it?

  • @Den-zf4eg
    @Den-zf4eg 6 місяців тому +1

    Якою програмою можна це зробить?

  • @romannavratilid
    @romannavratilid Рік тому

    Hm... so its basically something like photogrammetry...?
    This could also help photogrammetry right...? Like i capture only lets say 30 photos... But the resulting mesh and texture might look like it was made from i dont know... 100+ photos...? do i understand this correctly?

  • @ak_fx
    @ak_fx 3 роки тому

    Can we export 3d model?

  • @IRONREBELLION
    @IRONREBELLION 4 роки тому

    Hello. this NOT Dr. Károly Zsolnai-Fehér

  • @THEMATT222
    @THEMATT222 10 місяців тому

    Noice 👍

  • @linshuang
    @linshuang 9 місяців тому

    Pretty fucking cool

  • @Jptoutant
    @Jptoutant 4 роки тому

    been trying for a month to run the example scenes, anyone got thru ?

  • @ArnoldVeeman
    @ArnoldVeeman 3 роки тому

    That's photogrammetry... 😐 (Edit) Except, it isn't... It's a thing I dreamt of for years

  • @unavidamas4864
    @unavidamas4864 4 роки тому

    UPDT

  • @vinesthemonkey
    @vinesthemonkey 4 роки тому

    It's NeRF or Nothing

  • @DalaiFelinto
    @DalaiFelinto 4 роки тому

    I believe the link in the video description is wrong, but I found the page here: www.matthewtancik.com/nerf

  • @shortuts
    @shortuts 4 роки тому

    holy s**t.

  • @blackbird8837
    @blackbird8837 4 роки тому +1

    Doesn't look like anything to me

    • @kobilica999
      @kobilica999 4 роки тому

      Because its so realistic that u just dont get how big deal that is

    • @blackbird8837
      @blackbird8837 4 роки тому +1

      @@kobilica999 I was referencing Westworld

    • @kobilica999
      @kobilica999 4 роки тому

      @@blackbird8837 ooo that is different story xD

    • @holotwin7917
      @holotwin7917 3 роки тому

      Bernard?

  • @alfcnz
    @alfcnz 4 роки тому +3

    SPECTACULAR RESULTS and 3D animation!
    If interested, feedback follows.
    1. You start at 0:00 with a completely white screen. No good.
    2. The project title does not get any attention by the audience, given that EVERYTHING moves in the lower half.
    3. At 0:39 you are explaining what a hypernetwork is without using its name.
    4. 1:58 6 objects spinning like tops. It's hard to focus on so many moving things at once.

  • @adeliasilva409
    @adeliasilva409 4 роки тому

    ps5 grafics

  • @AndreAckermann
    @AndreAckermann 4 роки тому +1

    This is amazing! The implications for 3d and compositing workflows alone are mind-boggling. Can't wait to see this filter through to mainstream products.

  • @AndreAckermann
    @AndreAckermann 4 роки тому

    This is amazing! The implications for 3d and compositing workflows alone are mind-boggling. Can't wait to see this filter through to mainstream products.