NeRF 3D Capture With Luma AI Is MIND BLOWING !

Поділитися
Вставка
  • Опубліковано 8 вер 2024

КОМЕНТАРІ • 65

  • @fsstudiodev
    @fsstudiodev  Рік тому +5

    📌Would you consider using Luma AI for your 3D assets? What are your thoughts about using this technology for early prototyping phases ?

    • @SirenheadRizzler
      @SirenheadRizzler Рік тому

      I believe this could have true potential considering this is such early days. Luma AI requires an invite code or I would be trying it out myself!

  • @DJ_JJ
    @DJ_JJ Рік тому +16

    The sick part is, Luma is partnering with Poly cam, meaning we will get incredible photogramitory for geometry and crazy radiant feilds with reflections, transperency, roughness, etc

    • @jimj2683
      @jimj2683 Рік тому +3

      In 10 years I think 3d capturing will be so good that Google Street View will be converted to 3d (with a driving simulator on top).

    • @wspek
      @wspek Рік тому

      Do you have a resource for this?

    • @bolayer
      @bolayer Рік тому

      @@jimj2683 They're already getting ready to turn it into a Nerf now, not in a decade.

  • @kwea123
    @kwea123 Рік тому +3

    good to know that it takes 20-60m to complete. Guess it still requires colmap camera pose estimation behind the scene, there would be a huge speedup if this difficulity is cleared.

  • @amerfilmstudios9292
    @amerfilmstudios9292 Рік тому +2

    Get a rotational display ( usb cable ) and place any object on it. Place lock the camera on a height and angle adjustable tripod. Point the camera to a direction only with same background for all 360° images. This also applied for green or blue background for instant chroma key editing of an object.
    Additional : play with some studio light point to the object to match with the final background for such video/movie scenes 👌🏻
    U do less hard work with much better and clear 360° images. Try it 👌🏻

  • @codeFriendlyART
    @codeFriendlyART Рік тому +3

    Thank you for showing us this pretty cool tech! I would totally like to try it out when available and I'm positive it will be widely adopted, provided enough marketing/educational efforts like this! Keep up the great work guys and congratulations!

    • @fsstudiodev
      @fsstudiodev  Рік тому

      You are very welcome and thank you for your kind message, best to you as always !
      Dilmer

  • @TGSuineg
    @TGSuineg Рік тому +3

    How does it handle transparency and specularity compared to photogrammetry and does it create textures besides diffuse (such as metallic, roughness, etc.)?

    • @DJ_JJ
      @DJ_JJ Рік тому +1

      it uses neural networks and does yes. You could tell if you looked at a reflective material such as a tv while off or a chrome ball which Corridoor Crew did in their video.

  • @PrimeStudios
    @PrimeStudios Рік тому

    Nice! Looking forward to trying it out myself.

  • @FPChris
    @FPChris Рік тому

    Unreal’s nanite helps make the poly count less of an issue

  • @PeterLunk
    @PeterLunk Рік тому +2

    can't wait to try...

  • @williamcousert
    @williamcousert Рік тому +1

    Is this useful for scanning a room that could be used as a set in a digital production?

  • @Zouvi2001
    @Zouvi2001 Рік тому

    How would you export the scan animation (in Luma app after scan) as an MP4? Is that possible? If not is there another way to do that? Great video

  • @RyDaCol
    @RyDaCol Рік тому

    You are insanely underrated!

  • @TyKix
    @TyKix Рік тому +1

    I exported my scans and I have a lot more of the scene than I really wanted. Have you encountered this with other scans you have exported to use in a processing application. I uploaded the OBJ straight to Sketchfab and really wish I was more experienced at removing areas I don’t want.

    • @fsstudiodev
      @fsstudiodev  Рік тому +2

      Yes that’s normal since they currently don’t provide a slicing feature, you would need to bring the asset into Blender, Maya, or other similar tool to clean it up. I did submit a feature request to add a slicing type feature to their app but no ETA yet.
      Great question and thanks for watching !

  • @jeffreyeiyike122
    @jeffreyeiyike122 Рік тому

    i am currently working on nerf and synthesis in robotics

  • @gu9838
    @gu9838 Рік тому

    awww baby yoda lol

  • @imrenagy8122
    @imrenagy8122 Рік тому +1

    Amazing.

  • @vancandan1
    @vancandan1 Рік тому +3

    This is just NVIDA NeRF which is open sourced can capture reflections but not very good for photogrammetry. Why is LUMA not releasing this? Something feels fishy here

  • @user-nr2vy5ng9s
    @user-nr2vy5ng9s Рік тому

    Hi,
    Is it possible to export in STL format and than upload it to 3d printer?

  • @fgv_producciones
    @fgv_producciones Рік тому

    Great video and can you animate the model?

  • @RyDaCol
    @RyDaCol Рік тому

    Love it

  • @armandsfreibergs4069
    @armandsfreibergs4069 Рік тому

    so whats the new thing here or better? its just photogrammetry?

  • @lucaspedrajas5622
    @lucaspedrajas5622 Рік тому

    how can we use this for volumetric video?

  • @Igoreshkin
    @Igoreshkin Рік тому

    I have problems with finding answers for my quastions on their website(maybe someone will help). Is it only for iOS? What about pricing in the future?

    • @fsstudiodev
      @fsstudiodev  Рік тому +1

      Yes this is currently only available for iOS, not sure about the pricing model.

  • @DommageCollateral
    @DommageCollateral Рік тому

    how many verts does the model have? - 2 million?

  • @barbarusonline
    @barbarusonline Рік тому

    you did not show HOW to download the mesh from Luma AI

  • @elimindell
    @elimindell Рік тому

    this is crazy

  • @Ludifant
    @Ludifant Рік тому +1

    So how is this different from photogrammatry?

  • @user-ti3iv9lu9r
    @user-ti3iv9lu9r Рік тому

    😊

  • @igorchak48
    @igorchak48 Рік тому

    NeRF is the future! How can I get an Invite code?

  • @sifu2u_now
    @sifu2u_now Рік тому

    Is it only available on Mac?

  • @mikeillgaming4224
    @mikeillgaming4224 Рік тому +1

    Takes a long time to process

    • @fsstudiodev
      @fsstudiodev  Рік тому

      Yes we are hoping it improves with within the next few years to a level that it is instant !

  • @carlocarnevali7790
    @carlocarnevali7790 Рік тому

    Has anyone got invite codes for directly using it?

  • @MisterWealth
    @MisterWealth Рік тому

    Is there an Android app for this?

  • @amRHA
    @amRHA Рік тому

    I downloaded the app and I haven't received socalled invitation for three days
    it stinks a bit to me

    • @fsstudiodev
      @fsstudiodev  Рік тому

      I recommend engaging with them in twitter, that’s how they are currently inviting people.

    • @amRHA
      @amRHA Рік тому

      @@fsstudiodev thank you ..but I really hoped that the app was not this good

  • @farusiu
    @farusiu Рік тому

    is this ios only?

  • @walbollox1993
    @walbollox1993 Рік тому

    Only for mac monkeys?,,,,will just watch from my window

  • @DruidCyborgPuppy
    @DruidCyborgPuppy Рік тому +9

    I don't get this, other older photogrammetry software seems like a better choice. Why capture backgrounds? Is it able to capture other than small objects? Is it able to fill the gaps in models? I mean this was possible 8-10 years ago, luma software isn't out yet and it seems like it has fewer features, why not use software that's already there. Major game franchise already use photogrammetry (yes, it's not that simple just to take some photos with your phone), but I'm curious, why bother with this toy?

    • @RoyMagnuson
      @RoyMagnuson Рік тому +14

      it captures the lighting of the space much more accurately.

    • @perspectivex
      @perspectivex Рік тому +9

      As far as I can tell from having done photogrammetry and looked at videos on what "NeRF" is, NeRF seems to let you generate a really-low-quality-compared-to-properly-done photogrammetry 3d model with just a few pictures (I'm sure it'll improve and maybe match photogrammetry at some point but it's still far from it if this video is anything to go by). As a wild guess, maybe with NeRF you can do 5 or 10 pics? to capture a complete object vs. say 70-100 for basic photogrammetry. The AI model fills in the gaps. As to the lighting Roy mentions, I've no clue. For engineering and 3d printing purposes I care about capturing high-resolution, surface-accurate meshes (which I find hard to do really well...maybe AI will help there at some point too). I'm glad this video creator showed the actual meshes. Too many photogrammetry videos show a final model with the texture/source images mapped to it and it looks fantastic but it hides the mesh underneath that's looks like it's been worked over with an ugly stick and could outdo oatmeal for being coarse and lumpy. I guess for game assets it's ok since you anyway reduce polygon count as much as feasible and only care how the object looks in the game?

    • @dissidents1542
      @dissidents1542 Рік тому

      @@perspectivex thanks for this bro

    • @dissidents1542
      @dissidents1542 Рік тому

      @@perspectivex great summary

    • @liam9519
      @liam9519 Рік тому +1

      @@perspectivex for good results, NeRF generally still requires 100s of images