Scan yourself into Unreal 5 | Photogrammetry & Mesh to Metahuman Tutorial
Вставка
- Опубліковано 28 тра 2024
- Using Zephyr (www.3dflow.net/3df-zephyr-free/), Metahumans and Unreal we'll learn to create custom scans of our head to rig with Metahumans for animation, games or whatever you like!
Support us on Patreon for assets:
/ smallrobotstudio
Get assets on our Gumroad Store:
www.gumroad.com/smallrobotstudio
Check out more Small Robot Studio:
/ smallrobotstudio
smallrobotstudio.com
Check out more or Érika’s Work:
www.erikacasab.com
/ erikacasab
Stay up to date on Facebook:
/ smallrobotstudio
Timestamps
00:00 Zephyr
01:58 Scan Results
02:41 Cropping Scan
04:09 Generating High Res Scan
05:39 Export Settings
06:21 Cleanup
09:44 Importing to Unreal
12:41 Exporting to Metahumans
Most complete video on scanning yourself that I've come across. Appreciate you putting this together, thank you.
this plugin is a game changer. the fact that you can do so with such a low quality scan even more so
Great video. It's tough to take the perfect images, angles and even lighting makes a big difference in my experience
Thanks for the tutorial. I just literally discovered your channel through my feed tonight. Really thoughtful walkthrough. Im definitely going to give this a try sometime.
It's really cool I've been at it all weekend improving my scan data and getting better likeness!
Great tutorial! Very helpful.
It looks like a good work flow for quick/cheap character face creation. Seems like a good result for a low effort process and still get mesh that resembles original subject
I agree, with some extra effort, which I'll go into in a separate video I'll show how you can get the likeness a bit closer
Nailed it! Thx for the advice!
Cheers!
Great video, thank you!
However, I don't understand where the 50 photos come from? 7 photos (top) x 5 photos (side) = 35 photos... Did I miss something?
The rusult is rather good! Thank U for lesson! I hope in some time some UE character will appear)
Thanks it worked
Awesome 😮
AMAZING ! Thx
really awesome, I have a photogrammetry scanner with 300 cameras, and that helps me a lot😉😉
Lol! 300 will do it!
@@SmallRobotStudio 😅
This is awesome. Thank you. Any ideas on how to draw the facial features manually? I'm playing around with various head scans and some aren't being recognized when using the "autotracking" feature.
Check the docs and make sure the eyes and various pieces the scan needs to identify facial features have good contrast. I have manually corrected some tracks but not had the issue of it not being able to track
@@SmallRobotStudio Will do. Thank you!
nice! How are you animating your mouth at the beginning? Is it manually keyframed, arkit or livelinked? I ask cos this video is one of the only ones I've come across where the mouth is being animated without the metahuman mesh mouth & lips getting distorted!
ua-cam.com/video/AjpuBW2RXHI/v-deo.html for the mouth animation and just import the control rig as an .FBX using this method ua-cam.com/video/Z-Mjr3rz4rs/v-deo.html
@@SmallRobotStudio Thanks! will check it. I prefer fbx workflow too so will be interesting to see if I can jerry-rig it using an fbx out of iClone. Fingers-crossed
Excellent work, Mister.
So as an .obj file I'd be able to export that into Blender, wouldn't I ?
Thanks.
Yep you can
Thanks so much for this video man. I’ve been trying to get this thing to work for a while, but couldn’t find much help. I’ve been facing an issue with Zephyr not taking a bunch of my photos, do you have any idea why that might be?
Afraid I haven't had any issues like that so can't really say. My guess is either the photos are too low quality/res or the background doesn't give the software enough depth to interpret distance from the subject
@@SmallRobotStudio Okay thanks! If it’s the depth issue, how would I solve that? I don’t think it’s low res issue as I’ve been using iphone pics for this. Will trying with the ultrawide be any help?
@@the_good_p I think (and I'm no expert in Zephyr so maybe check their documentation) the more paralax each photo detects by comparison the better your results will be. I doubt focal length play too big a part though I wouldn't go lower than say a 28mm as the added lens distortion may be an issue
@@SmallRobotStudio thanks mate. Let me see if I can sort it out. Will update if I figure it out.
Hey late to the metahuman but im using 5.2 and when i go to import my meshes into unreal engine its taking a long time, been waiting about 40 mins and i got a pretty good computer. Is that normal to be waiting that long?
Thank you! Helpful! Can I know how to use a face from a picture to create a metahuman?
As far as I'm aware there's no software that can do that.
cool, thx :-)
I mean, this tech is impressive but as I don't know you it takes some time to recognize you when seeing this metahuman
is it possible to change the FOV of the metahuman editor camera? i have a feeling it is quite different than what you might have in the UE default editor view
Metahumans editor is between 80-100mm (best guess) and unfortunately no there is no way currently to change it - I wish there was though!
hi, im completely stuck at the part where you have to scale your mesh @ 12:05 you recommend to use another software. What kind of software are you talking about. Thanks
If you're having holes in your mesh on import use a modelling software like Maya or Blender to scale the geometry up and reimport it to Unreal
Still a massive Uncanny Valley effect.
mostly cause its still Realtime, any 3d artist can make this look closer with a path tracer
sir, pc configaration for 3d animetion please!
why i can't download the Zephyr ? could you help me?
I don’t know for some reason it didn’t pick up any of my head only the background .. lol. How should lighting be set up?
Indirect, diffuse lighting is best where it's bright enough to see your features. Ideally there will be some things in the background for the track to be able to judge distance. Also you need to stay dead still while the photos are taken
i imported my static mesh but the head was laying horizontally, i couldn't rotate as the transforms were greyed out...very frustrating
meta uman plugin is not showing up in search. how to fix?
How to add this type of animation in flutter apps
gracias, buen tutorial
This isn't working for me because im on a laptop
it keeps on crashing when i do the tracking part and it after it stopped, it closed 3 seconds afterwards
the metahuman plugin is not available for unreal engine 5.2
Do you have to scan yourself or can you import any humanoid character model?
Your own character should work using the same method - I'm going to try this weekend with some stylized characters. The important thing is that it can recognize landmarks and it needs some textures to identify that (eyes etc) - the docs go into this in more detail.
@@SmallRobotStudio I'm guessing as long as your character has spherical eyes it will be able to move properly.
@@JiReyAnimation If the eyes don't track properly you can always adjust them manually
But you didnt assign texture of the scan to the metahuman face. Is it possible to do it?
That's not really part of this workflow. If you want to do that you could take your final metahuman and your scan into ZBrush and project the textures from one to the other and bake those textures out to Unreal. Not sure what other software would do something similar but that's what I'd do.
@@SmallRobotStudio ok, thanks for info
Is it possible to export metahuman character, animate it and insert into an real life video?
You'd need to composite the animation over live action footage, so yeah possible but requires a bit of work
স্যার, আমার বাজেট কম তাই i512400, Gigabite B660m ddr4, 8gd ram, 1tb হার্ডডিক্স, গিগাসনিক 17" মনিটর 45k দিয়ে পিসি বিল্ড করলে ভিডিও এডিটিং + ব্লেন্ডার শেখা যাবে? ১ বছর পরে গ্রাফিক্স + এসএসডি ব্যবহার করবো কেমন হবে জানাবেন, আপনার সিদ্ধান্তেই আমি পিসি কিনবো, ভাল থাকবেন সবসময় শুভ কামনা নিরন্তর।
Could you export from Zephyr into Unreal without taking it in Zbrush
As I said you don't need to cleanup if your scan is good, I just wanted to correct the hair
@@SmallRobotStudio Okay thanks
not sure why my material doesnt come into the metahuman indentifier? does that even matter? assume meta human disregards the texture?
Hey, sorry for the late reply. Open up the Blueprint for your imported model and apply and save your material there and that should fix it.
@@SmallRobotStudio ah thanks!
wow great video thanks! so i couldn't understand... Did you bring back a texture image of your skin into unreal for the metahuman? Or are we just moving the sliders close?
So the capture was done by only one camera...? (someone else captured you)...?
Yes
Hello, Im not able to see my custom Metahuman in Bridge (Inside Unreal 5.0.2) can anybody help? Thx :)
I’m having the same issue!
@@joshuacadogan5174 hello I manage to make it work, I deleted all my epic games and quixel (app data files) then I made myself a new account (same epic games account for Metahuman creator app, Bridge and Unreal Engine) then I isntalled Bridge stand alone and Bridge plug in for Unreal 5.0.2 and after having the same library set up for Bridge stand alone and the plug in version, pretty painful but it worked for me haha
@@danelokikischdesign thank you so much for the update! I was going to let you know, it turned out I had two epic games accounts. Bridge automatically sent my mete humans to one but creator opened in another. I deleted one of them and made a creator account with the same bridge account. Crazy!
@@danelokikischdesign I also had to clear my chrome data, it would just fail connection otherwise as it tried to find my deleted account.
Not working for me. I've sculpted a head in zbrush and when I import to unreal engine it says unable to triangulate mesh. Followed this guide to the T. Do you need to import textures this part is confusing, as metahuman has it's own skin
Main thing is that the nasal labial blades, lips and eyes have good contrast so a texture map would help for the track
@@SmallRobotStudio I had the same problem but managed to work it out, I had to figure out how to texture my model because I wasn't using photogrammatery i was using zbrush and normally just model, and when you go to unreal engine you only show the model and one jpeg texture, so its a bit confusing because when you normally texture you have like base/metal/roughness etc, anyway some how managed it and was able to make this Schwarzenegger metahuman thanks to your video, so thanks. Anychance you can do a strict zbrush modelling to unreal process, im sure im doing it the long way, thanks again.
ua-cam.com/video/Kx2N7HhUDcY/v-deo.html
@@digitalwizard1608 Nice job on the Terminator! Yep I'm working on a video about how I'm using a sculpt from Zbrush for Mesh to Metahumans, hopefully that'll be out soon.
@@SmallRobotStudio awesome man thanks for all your work you do
Can you do that with 1 image only instead of 50?
For photogrammetry you require images from multiple angles of the subject. The more the better, one won't be sufficient.
@@SmallRobotStudio How many pictures do you think should be enough?
@@UE5withTallal I'd want at least 30
no why would you think that you could? dont vote or have kids if you lack common sense that badly
@@SmallRobotStudio Oh okay, I think I can make atleast 30 photos of mine what about if I needed to make a meta human of a celebrity etc, How will i be able to do that is there any way? Thank you for your replies btw
Software name please
Unreal 5 and Metahumans. Zephyr for the scan, ZBrush for the edits.
13:05 annoying unreal thing. model was selected when you created that thing then need still select it even choose some other..and still all this should be in metahuman app/cloud load mesh thats that then import meta human to unreal not jump between lol
what are some Mac-friendly alternatives to Zephyr?
Don't own a Mac so I couldn't say
Anyone else have it crashing while autotracking?
.JPG, PNG but still not works.
Is this a different workflow to your other video? It seems less stressful, the other one you spent ages cleaning up in zbrush, here you didn't even make a mouth bag or eye bags.
This is a brand new workflow, this plugin was released yesterday
UA-cam is certainly paying very VERY close attention to what I'm doing every single microsecond of my life. The algorithm strikes yet again. BTW I've boycotted all advertisers before watching this.
Okey dokey?
0
This is scary no?
Why?
Acts 2:17-21
17 “‘In the last days, God says,
I will pour out my Spirit on all people.
Your sons and daughters and ITS will prophesy,
your young men will see visions,
your old men will dream dreams.
18 Even on my servants, both men and women,
I will pour out my Spirit in those days,
and they will prophesy.
19 I will show wonders in the heavens above
and signs on the earth below,
blood and fire and billows of smoke.
20 The sun will be turned to darkness
and the moon to blood
before the coming of the great and glorious day of the Lord.
21 And everyone who calls
on the name of the Lord will be saved.
Notice I added the (ITS) part cause you know some beings are not a male or female or some are just asexuals non-binaries and God would understand that!
Galatians 5:22-23
22 But the fruit of the Spirit is love, joy, peace, forbearance, kindness, goodness, faithfulness, 23 gentleness and self-control. Against such things there is no law.
Matthew 22:36-40
36 “Teacher, which is the greatest commandment in the Law?”
37 Jesus replied: “‘Love the Lord your God with all your heart and with all your soul and with all your mind.’ 38 This is the first and greatest commandment. 39 And the second is like it: ‘Love your neighbor as yourself.’ 40 All the Law and the Prophets hang on these two commandments.”
Matthew 7:13-14
The Narrow and Wide Gates
13 “Enter through the narrow gate. For wide is the gate and broad is the road that leads to destruction, and many enter through it. 14 But small is the gate and narrow the road that leads to life, and only a few find it.
Revelation 21:6-8
6 He said to me: “It is done. I am the Alpha and the Omega, the Beginning and the End. To the thirsty I will give water without cost from the spring of the water of life. 7 Those who are victorious will inherit all this, and I will be their God and they will be my children. 8 But the cowardly, the unbelieving, the vile, the murderers, the sexually immoral, those who practice magic arts, the idolaters and all liars-they will be consigned to the fiery lake of burning sulfur. This is the second death.”
Matthew 5:1-12
Introduction to the Sermon on the Mount
5 Now when Jesus saw the crowds, he went up on a mountainside and sat down. His disciples came to him, 2 and he began to teach them.
The Beatitudes
He said:
3 “Blessed are the poor in spirit,
for theirs is the kingdom of heaven.
4 Blessed are those who mourn,
for they will be comforted.
5 Blessed are the meek,
for they will inherit the earth.
6 Blessed are those who hunger and thirst for righteousness,
for they will be filled.
7 Blessed are the merciful,
for they will be shown mercy.
8 Blessed are the pure in heart,
for they will see God.
9 Blessed are the peacemakers,
for they will be called children of God.
10 Blessed are those who are persecuted because of righteousness,
for theirs is the kingdom of heaven.
11 “Blessed are you when people insult you, persecute you and falsely say all kinds of evil against you because of me. 12 Rejoice and be glad, because great is your reward in heaven, for in the same way they persecuted the prophets who were before you.
This is cheating. I prefer to sculpt my assets.
I've also sculpted myself then used the mesh to Metahumans plugin to upload the results. Likeness is a little better too so if that's your jam I recommend it.
@@SmallRobotStudio nah, I do likenesses through sculpting. Sculpting is more fun and rewarding.
@@Daakusutepsu1 no I mean use your sculpt and just use mesh to Metahumans for the rig - best of both worlds. Unless you want to do all the blendshapes, cavity maps etc. If you do, more power to you!
@@SmallRobotStudio oooh, that does sound interesting. My mistake.
@@Daakusutepsu1 all good, try it out and see if it works for your stuff!