Amazing AR Effects Are Coming!

Поділитися
Вставка
  • Опубліковано 30 чер 2020
  • ❤️ Check out Weights & Biases and sign up for a free demo here: www.wandb.com/papers
    Their mentioned post is available here:
    app.wandb.ai/latentspace/publ...
    📝 The paper "Consistent Video Depth Estimation" is available here:
    roxanneluo.github.io/Consiste...
    🙏 We would like to thank our generous Patreon supporters who make Two Minute Papers possible:
    Aleksandr Mashrabov, Alex Haro, Alex Paden, Andrew Melnychuk, Angelos Evripiotis, Benji Rabhan, Bruno Mikuš, Bryan Learn, Christian Ahlin, Daniel Hasegan, Eric Haddad, Eric Martel, Gordon Child, Javier Bustamante, Lorin Atzberger, Lukas Biewald, Michael Albrecht, Nader S., Nikhil Velpanur, Owen Campbell-Moore, Owen Skarpness, Robin Graham, Steef, Sunil Kim, Taras Bobrovytsky, Thomas Krcmar, Torsten Reil, Tybie Fitzhugh.
    More info if you would like to appear here: / twominutepapers
    Károly Zsolnai-Fehér's links:
    Instagram: / twominutepapers
    Twitter: / twominutepapers
    Web: cg.tuwien.ac.at/~zsolnai/
  • Наука та технологія

КОМЕНТАРІ • 459

  • @youtube_fantastic
    @youtube_fantastic 4 роки тому +818

    Nothing like smell of a fresh 2 minute paper!

    • @chrisray1567
      @chrisray1567 4 роки тому +7

      I love the smell of wood pulp in the morning!

    • @NubNublet
      @NubNublet 4 роки тому +8

      Especially one that isn’t more fluid simulation lol

    • @ZorgKirill
      @ZorgKirill 4 роки тому +1

      that runs for 5 minutes xD

    • @DwAboutItManFr
      @DwAboutItManFr 4 роки тому +4

      Image the smell two papers down the line?

    • @phiro4305
      @phiro4305 4 роки тому +1

      Nothing like the smell of a depth-mapped cat

  • @blenderguru
    @blenderguru 3 роки тому +61

    Holy cow! Realtime depth information could be used for overlaying information for all sorts of displays: classroom learning, live events, concerts, task specific assistance. Might be a big leap forward for AR. Exciting tech.

  • @sqworm5397
    @sqworm5397 4 роки тому +419

    That "dear fellow scholars" hits hard everytime

    • @joancaic2853
      @joancaic2853 4 роки тому +37

      What a time to be alive!

    • @tempahp
      @tempahp 4 роки тому +24

      Hold on to your papers!

    • @aliasd5423
      @aliasd5423 4 роки тому +30

      *happy fellow scholar noises

    • @CellularInterceptor
      @CellularInterceptor 4 роки тому +1

      it's their trademark

    • @peskarr
      @peskarr 4 роки тому +1

      Lets add one more paper to this little one, becouse everybody needs a friend. Little neural network has to have a place to set there. There he goes...

  • @b3nsu
    @b3nsu 4 роки тому +512

    the 2010s were the years of the smartphone
    the 2020s will be the decade of AR and VR

    • @tiavor
      @tiavor 4 роки тому +39

      @nasolem we can only hope and wait for a real SAO.

    • @laurenz1337_
      @laurenz1337_ 4 роки тому +14

      And the 2030s will be the decade of neural link and brain interfaces

    • @Vaeldarg
      @Vaeldarg 4 роки тому +20

      @@cybercrazy1059 Hate to break it to you, but "mind upload" doesn't mean what you think it means. When you upload a file, it doesn't disappear from your computer. It simply creates a new file at the new location that is a copy of the original information. However, full-sensory simulation for virtual environments just need how the brain uses neural transmitters/inhibitors to be figured out for that to be achieved.

    • @turkosicsaba
      @turkosicsaba 4 роки тому +6

      As we can see in the video above, we will always find ways to add cats to our new ARs and VRs.

    • @Vaeldarg
      @Vaeldarg 4 роки тому

      @@nagualdesign Only an actual baby would be bothered by such an obvious troll.

  • @WangleLine
    @WangleLine 4 роки тому +115

    *excited paper holding*

    • @keyboard1819
      @keyboard1819 4 роки тому

      Hey, you are a Jonas tyroller viewer. RIGHT!!?

  • @DuckTheFinn
    @DuckTheFinn 4 роки тому +258

    You can't expect us to keep a grip on these papers if you keep showing stuff like this!

    • @alexcrespo3252
      @alexcrespo3252 4 роки тому +9

      My papers all emigrated to nigeria to start a new life

    • @Metaloid-wv4kz
      @Metaloid-wv4kz 4 роки тому +2

      I had to order in a dump truck.

  • @Jackp2003
    @Jackp2003 4 роки тому +13

    I get so excited whenever you upload!

  • @georgianfishbowl170
    @georgianfishbowl170 4 роки тому

    every video you make blows me away! these advances are crazy and its so great to see these thing happening RIGHT NOW! Thanks for making these videos and bringing it to our attention

  • @tobi6758
    @tobi6758 4 роки тому +145

    The next iPhone is expected to have a "LiDAR" sensor. Wonder if that will give that "perfect" AR effects.

    • @JBB685
      @JBB685 4 роки тому +15

      We’re getting into some really exciting territory here!

    • @Navhkrin
      @Navhkrin 4 роки тому +29

      It most certainly should, unless Apple messes up somehow. In my opinion, future is going to be about combining a very-cheap (sparse) but accurate lidar with depth estimator for filling in the gaps, it gives us precise values to work with. Lack of reference points is the biggest problem with monocular depth estimation.

    • @cunty
      @cunty 4 роки тому +6

      doubt it. the ipad's lidar sensor, which i'm assuming is the same going into the iphone, is not nearly as fine on the details as faceid. the dots the lidar projector projects are spaced pretty far apart, which leaves the ipad to have to fill in the rest of the info.

    • @DamianReloaded
      @DamianReloaded 4 роки тому +9

      @@cunty I'm confident higher resolution cameras and stronger parallel processors will enable NNs to do this from video only.

    • @dykam
      @dykam 4 роки тому

      @SCUUBE Got reminded of the same. My tries with arcore didn't have flickering, tho it did sometimes take some time for it to discover the depth in certain areas. There's an arcore labs app in the store to try it out.

  • @SamDutter
    @SamDutter 4 роки тому +7

    Amazing!!! This probably has tons of application for photogrammetry as well.

  • @jameshughes3014
    @jameshughes3014 4 роки тому

    how do your videos so consistently blow my mind? You have a true talent for presenting dry academic information in a way that makes it exciting and understandable. Thank you for what you do.

  • @rasp1628
    @rasp1628 4 роки тому +69

    Maybe google street view could use this to make more detailed 3D buildings?

    • @luck3949
      @luck3949 4 роки тому +10

      This isn't something radically new, it's an improvement of existing methods. Making 3d models out of images is possible since 2009, find "building Rome in a day" paper and video. So I guess google street view isn't 3d not for technical reasons, but because their management don't want to bother making it in 3d.

    • @rasp1628
      @rasp1628 4 роки тому +1

      @@luck3949 that's cool, I wonder what's possible with todays technology🤔.

    • @luck3949
      @luck3949 4 роки тому +1

      @@rasp1628 I don't know, as I don't monitor this field. Look at LSD SLAM if you want to see something impressive, but it's 5 years old.

    • @tendermoisturized4199
      @tendermoisturized4199 4 роки тому

      @@luck3949 Well, yes but improvements in the process could potentially make it more cost-effective and worth it for google, specially if you can teach an AI to just run through the preexisting library of images and produce decent results on its own.

    • @MrRobotrax
      @MrRobotrax 4 роки тому +1

      @@luck3949 it's already 3d

  • @luis96xd
    @luis96xd 4 роки тому

    Wow, this effects are so Amazing! Thanks for sharing this with us 😄

  • @JoakimMoesgaard
    @JoakimMoesgaard 4 роки тому

    I love keeping myself updated with your videos. Thank you.

  • @pladselsker8340
    @pladselsker8340 4 роки тому

    very very very cool! Can't wait to see the follow up papers

  • @DriesduPreez
    @DriesduPreez 4 роки тому

    No way? That depth solve is so crisp and consistent, and that's all without any depth sensor or additional camera? Man I can only imagine what this means for VR and AR.

  • @thedrunknmunky6571
    @thedrunknmunky6571 4 роки тому +1

    You read my mind! Just a few hours ago (I swear) I was thinking of having to search online to find an algorithm to get depth maps from a video stream. Although I can't use it yet for my project (as the depth maps are not detailed enough yet), imagine how detailed and fast it will be a few papers down the line! I can't wait!

  • @NosAltarion
    @NosAltarion 4 роки тому

    I found your channel by pure luck and god damn. That's been the best subscription I ever made to a youtube channel.
    I can't express how most of your video just blew my mind.
    And the presentation is perfect. Bite sized yet explained so clearly.
    Thank you for your content.

  • @elammertsma
    @elammertsma 4 роки тому

    It'll be exciting to see this get sped up to support real-time depth mapping. From the paper, it took 40 minutes to process a video of 244 frames (approx a 9-second video on most devices) so there's quite a bit of work necessary to get to the AR-stage, but these results are already incredibly impressive. Now it's going to be all about speeeeeed!

  • @emrahyalcin
    @emrahyalcin 4 роки тому

    thank you W&B. because of you, we can watch this channel. Thank you really.

  • @fureversalty
    @fureversalty 4 роки тому

    2:34 At first I was kind of skeptical about the accuracy of the depth effect, until I saw that. The way the water refracts the fire hydrant is so cool.

  • @moby_vyk
    @moby_vyk 4 роки тому +116

    I miss the times where you were also explaining the papers :(
    From what I understand from their video: ua-cam.com/video/5Tia2oblJAg/v-deo.html , it's not real-time, it takes a video as input, and by taking from 2 random frames a pixel and trying to estimate it's depth, you'll get 2 approximations. The difference between these 2 is the error that is then backpropagated through a network that, after doing this a lot of times, will end up giving a way better and consistent approximation.

    • @HarryHeck2020
      @HarryHeck2020 4 роки тому +12

      two papers down the line...

    • @MOOBBreezy
      @MOOBBreezy 4 роки тому +10

      Yeah, that's what I starting thinking. I noticed that none of the examples shown was real-time, so this wouldn't work as well in driving ai.

    • @SillyMakesVids
      @SillyMakesVids 4 роки тому +1

      What happens for moving objects?

    • @user-wq3hc4ze3n
      @user-wq3hc4ze3n 4 роки тому

      @@MOOBBreezy I think you still would be able to train an ai on the results.

    • @martinsmolik2449
      @martinsmolik2449 4 роки тому +2

      I still think that this will be the future of CGI tho

  • @angledcoathanger
    @angledcoathanger 4 роки тому

    That's amazing. I'm glad I was holding onto my papers.

  • @benshakespeare268
    @benshakespeare268 4 роки тому

    That's amazing. I work with images everyday and I can't imagine improving the result any further ... without including extra data sources, that is.

  • @eenvleugjegoeiegames
    @eenvleugjegoeiegames 4 роки тому +11

    Your videos always give me the motivation to keep working on my AI degree with enthousiasm, so a huge thanks for that!

    • @TwoMinutePapers
      @TwoMinutePapers  4 роки тому +2

      Absolutely amazing, kind thanks for sharing this!

    • @boujeejams3086
      @boujeejams3086 4 роки тому +1

      Could you share some Al resources for people who aspire to do it too please

  • @dissonanceparadiddle
    @dissonanceparadiddle 4 роки тому

    YES!!! you humans are doing so good! Keep going

  • @neoqueto
    @neoqueto 4 роки тому

    The glowing particles effect... holy crap, this method comes with re-lighting FREE OF CHARGE!

  • @mcantisnake
    @mcantisnake 4 роки тому

    Nice video! wish you success, keep it simple like you do and people will come back!

  • @sanboxengine
    @sanboxengine 3 роки тому

    I've been following the Two Minute Papers channel for quite some time, I think it's been a few years since I found the channel and got in love with the content. Even if I know the intro says "Dear fellow scholars, this is Two Minute Papers with Dr. Károly Zsolnai-Fehér.", because Ive heard and read it so many times, even then I still hear the intro like if it says "Dear fellow scholars, this is too many papers with Carlos Jonas Ifahir". Please, tell me I'm not the only one xD

  • @RavenDuran231
    @RavenDuran231 4 роки тому +6

    "What a time to be alive!" - I always loved that line. I can feel the sheer passion! :D

  • @somethingthatpops
    @somethingthatpops 4 роки тому

    Imagine how this could be used with art sculpting software, where you can reach your hands inside an object and have them be properly occluded (or not occluded) based on this depth map. I had thought this would only be possible using something like lidar or a lytro camera, but it looks like itll work for any camera soon enough! AR is awesome

  • @notnullptr
    @notnullptr 4 роки тому +3

    What a time to be alive!

  • @sabofx
    @sabofx 4 роки тому

    Excellent! I cannot wait for somebody to code this into Adobe After Effects.

  • @anran6425
    @anran6425 4 роки тому

    The AR effects look amazing!

  • @HarhaMedia
    @HarhaMedia 4 роки тому

    Having worked with AR stuff, I can say that this looks amazingly promising.

  • @MrMCKlebeband
    @MrMCKlebeband 4 роки тому

    that this works so well is beyond nutty.

  • @debangan
    @debangan 4 роки тому

    "Hold on to your papers"
    Damn! That was a nice one

  • @abrikos1100
    @abrikos1100 4 роки тому

    It's amazing! Thanks!

  • @degiguess
    @degiguess 4 роки тому

    one of the things I'm most excited for with these depth algorithms is the idea of 360 degree video being able to have depth information so it can be displayed properly in VR. Imagine being able to watch live events in proper 3D VR like you're actually there.

  • @AK-km5tj
    @AK-km5tj 4 роки тому +1

    Thank you!

  • @nolifeonearth9046
    @nolifeonearth9046 4 роки тому

    a new era for "here in my garage"-style videos.

  • @axiom1650
    @axiom1650 4 роки тому +1

    The flickering ball on the table looks more like z-fighting to me. Can't wait for your take on GPT-3!

  • @DavidMcCoul
    @DavidMcCoul 4 роки тому

    What an exciting time to be alive!

  • @bobsmithers4895
    @bobsmithers4895 4 роки тому +1

    This is so fricking cool!

  • @KuraIthys
    @KuraIthys 4 роки тому

    Well, this is impressive...
    This should be very interesting in future...

  • @DamianReloaded
    @DamianReloaded 4 роки тому

    This is really outstanding.

  • @EddyKorgo
    @EddyKorgo 4 роки тому +1

    2:41 neural RGB - thats how i see images in my head when i have headache. That flickering.

  • @majorjohnson8001
    @majorjohnson8001 4 роки тому

    This is exactly what I want/expect from a device like the Microsoft Hololens.
    Mind, the dev kit I worked with back in 2016 was really pretty good for the form factor (it would scan space around it and produce a mesh from a voxelized space using 5cm voxels) but it couldn't handle real time changes to that environment. Things that didn't move (floor, ceiling, walls, furniture) would be generally stable, but if you were lucky enough to have a person walk by, you'd have a physics ghost of them for about five minutes before that region was rechecked.
    But it could not see glass at all (and I suspect that will always be a problem) and I never got a chance to see how it dealt with mirrors.

  • @MichaelJOneill333
    @MichaelJOneill333 4 роки тому +1

    This. Is. AMAZING!

  • @manaarikicarpentier6038
    @manaarikicarpentier6038 4 роки тому

    I really love your Channel, merci beaucoup

  • @YusiDJordan
    @YusiDJordan 4 роки тому

    This will be incredible for VFX artists like myself. God, the future looks beautiful.

  • @connormichalec
    @connormichalec 2 роки тому

    Absolutely incredible

  • @un1b4ll
    @un1b4ll 4 роки тому

    I'd love to see a video about your deal with W/B and learn about how partnerships and ownership circularly benefit the content creation, W/B, yourself, and your viewers.

  • @BrownHuman
    @BrownHuman 4 роки тому

    BUt does it need a high end scanner like the one in the new iPad pro Lidar scanner ? or does it work with all kinds of general smartphone cams

  • @blacksages
    @blacksages 4 роки тому

    Incredible !

  • @Scio_
    @Scio_ 4 роки тому

    I want this in ARCore, last week!

  • @Frostyflytrap
    @Frostyflytrap 4 роки тому +1

    I'm curious how this handles light refracting through transparent/translucent objects.

  • @Axewayboy
    @Axewayboy 4 роки тому

    Beautiful. Add some sensors (optional, so it can run without them) to the input of that and create a perfect 3D world for a car or machine. Imagine know mixing 2 cameras for a robot, cheap 3D world.

  • @misaalanshori
    @misaalanshori 4 роки тому

    I'm really excited for this, not really for the AR effects, but an accurate depth maps from videos means smartphone cameras could finally record videos with bokeh, there is portrait mode for photos, but there is no good version for videos yet (i think)

  • @besknighter
    @besknighter 4 роки тому

    IMHO, the quality of their results are already pretty usable for almost all consumer products!

  • @mm-rj3vo
    @mm-rj3vo 4 роки тому

    Holy CRAP, I cannot WAIT for this to be implemented into XR, AND driverless cars!

  • @foldinmusic8810
    @foldinmusic8810 4 роки тому

    this it the thing ive been waiting for years

  • @kingpet
    @kingpet 4 роки тому

    shout out to the cat being the patient subject of this video!

  • @Bos_Meong
    @Bos_Meong 4 роки тому +1

    that cat is so adorable by the way

  • @Uhfgood
    @Uhfgood 4 роки тому

    I don't really understand half this stuff, but it's really cool to watch. It means automatic rotoscoping so then vfx artists don't have to do so much work by hand roto-ing stuff.

  • @thevivekmathema
    @thevivekmathema 4 роки тому

    can anyone help me solve a problem of matching a head video of a person on the neck of other person when the portion of the other is not available. i.e cropped out video?

  • @ProfessionalTycoons
    @ProfessionalTycoons 4 роки тому

    this is such a breakthrough

  • @kevalan1042
    @kevalan1042 4 роки тому +1

    Dr! First time I hear the title, congrats (maybe it was a long time ago?)

  • @jascrandom9855
    @jascrandom9855 4 роки тому

    This would be Amazing for VFX.
    Blender so needs this.

  • @imaUFO672
    @imaUFO672 4 роки тому

    This could speed up the process of adding visual effects in movies drastically

  • @Tailslol
    @Tailslol 4 роки тому

    binocular camera and temporal frame by frame cleaning would help this a lot.

  • @muizzy
    @muizzy 4 роки тому

    Hi Károly,
    I've been following this channel and the general AI field for a few years now, but since I finished school and started working, I've really started to do something with it in my free time and apply the concepts I've been hearing about so much.
    I would love your insight on where to go to continue my learning. Especially looking towards network architectures.

  • @vestlen
    @vestlen 4 роки тому

    I love your videos! Could you do an update to your How To Get Started With Machine Learning video? It's been 4 years and so much has changed!

  • @Tondadrd
    @Tondadrd 4 роки тому

    2^7th!!
    This paper blew my mind so hard. I attended class about computer vision for a semester and the vision is so hard and complex (not to learn, to do at all)! Even with depth cameras!
    I don't really care about video effects, it would be so amazingly useful without them alltogether!

  • @ayandas8757
    @ayandas8757 4 роки тому

    Just awesome 😮😮

  • @Veptis
    @Veptis 4 роки тому

    The issue I am having is that there is no hard edges. It's too smoothed. And it's coarse. The resolution is like 1/9 for G vs D.
    And if you have multiple sensors, this should vastly improve due to stereo vision. Or even tof sensors build into phones now.
    I would use this method in combination with a low resolution thermal camera to do a 3D reprojection for higher resolution and accuracy along usual photogrammetry solutions. I sketched it out a long time ago - but never moved on.
    I have seen this paper a few weeks ago and looked into it a little. Can't wait to study this type of stuff at an academic level.

  • @kylebowles9820
    @kylebowles9820 4 роки тому

    Wow, that's good depth quality; better than stereo depth sensors! Watch out Google Depth API :)

  • @GrimOfDonuts
    @GrimOfDonuts 4 роки тому

    So, you need an AI to run first to create a sharper image to remove movement blur to get accurate object edges and another to apply the movement blurs to the areas that got sharper to keep camera movement consistency?

  • @Dugiedugdug
    @Dugiedugdug 4 роки тому

    Interested that most of these clips seem to have a moving camera? Is this necessary? Obviously that would give more information on the depth of an object by comparing each frame to the last, does this method work on static images?

  • @willinton06
    @willinton06 4 роки тому

    Oh Apples’ Glass team is going to enjoy this paper

  • @osakanone
    @osakanone 4 роки тому

    I wonder what happens when you pair the colour and depth data with the past data of previous states and a classifier?

  • @fulanodetaldoorkut
    @fulanodetaldoorkut 4 роки тому

    Now I want to see what that can do for selfdriving cars since it is much better than previous method.

  • @JamesJazzz
    @JamesJazzz 4 роки тому

    This really reminds me of that 2012 game Kinect Party. Just makes you wonder how these experiences will look with current tech.

  • @VaradMahashabde
    @VaradMahashabde 4 роки тому +1

    I thought this is how AR worked already! Any news to learn about current AR implementations?

  • @FurryEskimo
    @FurryEskimo 4 роки тому

    I’ve seen a couple programs like this shown on your channel, where an input image or video can suddenly generate a better zooming effect or useful depth map, but do none of these have demos available?

  • @monad_tcp
    @monad_tcp 4 роки тому +1

    Two Minute Papers, I almost hear "too many papers". of course, there's never too many papers !!

  • @serta5727
    @serta5727 4 роки тому +1

    Just thinking about combining the algorithm to two cameras. Them seeing what we see, making a good 3d scanner

  • @adrianvasquez4351
    @adrianvasquez4351 4 роки тому

    That cat is adorable!

  • @multimediahelp1
    @multimediahelp1 4 роки тому

    Could this be used for 3D scanning objects with your phone better and more accurate?

  • @arnavrawat9864
    @arnavrawat9864 4 роки тому

    This was amazing

  • @FlyingBanana78
    @FlyingBanana78 3 роки тому

    There is another app out there that is relatively new called camtrackAR on the apple store that is free and auto tracks footage recorded from iPad or iPhone and creates a camera solve that can be used in blender. Free version allows one point to be used and paid version allows for more than 1 but one can still get some real nice results.

  • @antlu65
    @antlu65 4 роки тому

    One of my fav parts of each video is the very beginning. I listen to the vid author pronounce his own name, and it sounds amazing. Then i try but its no good lol.

  • @finnaustin4002
    @finnaustin4002 4 роки тому

    I wonder if this could be combined with an actual depth camera/stereoscopic cameras for higher performance

  • @qu765
    @qu765 4 роки тому

    I feel like the water had a bit too high IOR. Great video.

  • @alexpaden8544
    @alexpaden8544 4 роки тому

    looks great

  • @confuseatronica
    @confuseatronica 4 роки тому

    2 more cats down the line, the picture will be even fuzzier

  • @EpicVideoClips101
    @EpicVideoClips101 4 роки тому +43

    Snapchat: I own this now

  • @SHCreeper
    @SHCreeper 4 роки тому

    I think one important aspect that was left out is the time it takes to calculate these depth maps. If I remember correctly, it was around 20 minutes per frame.

  • @beaconofwierd1883
    @beaconofwierd1883 4 роки тому

    This is going to completely replace bluescrren in the film industry in a few years. No more blue shining lights, no expensive LED background screens, just put up a gray neutral screen as background, or just use any old room. 1000 budget home made films could have the same visual effects as todays block busters :O

  • @mascuudsaid9791
    @mascuudsaid9791 3 роки тому

    What a time to be alive is in two minute paper vedios