For better results, I think it's better that you don't put in your prompt any "talking" "Saying" (mouth/lips movements) anything for the base video, so the mouth and face do not be doing any expression as base face video, so then will be modified from a "blank default/idle expression" of the face by the face controlled video, so the modification will happen to the same face and not a face that it's already doing/changing over time.
THANK YOU!!! When you opened Comfy I almost clicked away. Its so so complex and I just wanna hire people to collab with at that point. But you broke it down SO well, I understood and it made so much more accessible thanks
Thanks for the tutorial. These lip-sync features are so important to advance the control over consistent characters. I hope more A.I. video platforms will implement this under the hood, without all the incredible complex Comfy parameters.
I respect the flexibility of ComfyUI, but only super geeks are going to even attempt to grok it. Higher level tools (even if Comfy is under the hood) will be needed for wide spread usage, IMO. Awesome presentation. You are top notch, sir.
I'm a horrible actor and I don't want to record myself. I know this is going to sound incredibly lazy but I hope the option to prompt different emotions becomes available. I don't need Oscar worthy performances for the projects that I'm working on. It's great that's there's a lot option for controlling faces. Now only if we could have 3D poseable consistent characters we'll finally be able to have AI actors?
Thanks for tutorial. You used close-up image and the result was almost good and acceptable. My question is, can the same result be obtained with medium or wide shot images? Have you ever tried it?
Can you also do a video about ‘Crew AI’? I think ‘Crew AI’ seems like a useful AI tool because it gives users a lot of control over their AI character’s facial expressions, just by moving some sliders around. It doesn’t really create animations, but it can change the facial expressions of characters in still images. I need to work with both still images and animations, so I like that there’s a tool that can modify characters in still images like this. Also, in terms of animation, I can see a tool like this being useful for creating a ‘start frame’ and ‘end frame’ for an AI animation.
So if you wanted to sync up with audio, I'd imagine you'd have to slow the final output by 25%? If that's the case you'd have to increase the fps by 25% (around 30fps) if your final final output is 24fps with synced video?
I found that as long as I had the same frame rate on each video, I didn't need to increase the speed of the driving video by 125% initially. Also, I got better results, prompting for subtle head movements only and without prompting the character to speak into the camera.
@@EmilyNilsen Yup. Grab the fps from the driving video via the Video Info VHS node and plug that into the Video Combine VHS node. You shouldnt have to do any speed changing.
Hey! Great video. I have an unrelated question, kind of an odd one. What's your background? Are you using a greenscreen? If so why can't I see any green in your glasses? Is it not a green screen and just a straight up tv screen? Your shot is lit very nicely, and it matches the background very well! I just had that question lol
Just to clarify if I've got this correct: 1. If you're uploading a control video to animate an image, you can move your head all over the place and the resulting video will mimic those movements. 2. But if you're uploading a control video to animate a video, you need to keep your head still as it will only apply your face to the video character that is already moving their head. Is it correct to say the control video head movements will not affect the existing video character's head movements?
would it not make more sense to also record the reference footage in front of a white wall or plain background vs clutter in the background? Surely that makes a difference?
Can't get this to work. I've followed the tutorial step by step, and used the specified assets but the video combine node always appears EMPTY and I cannot render. How to fix this? Thanks in advance.
@@curiousrefuge i mean the comfyui's workflow, simple json file, of course the models and everything else is separate download, i understand that. But why not workflow file?
What? What are you talking about? Is there an alternative to do what was shown in this video without comfy UI and doesn't involve installing Live portrait on the PC? (Which I had issues with)
@@adarwinterdror7245 a it re pr en eu r made a tutorial about this topic. It’s his second most recent video. I had to write spaces in between. As my comment disappears when writing his channel name
If you want easy, Runway has Act-One for talking heads. This however looks like it might be beneficial if you were doing a walk and talk video from an image to video and you needed a better facial performance on the character. Is it easy? No. Is it less expensive and easier than hiring a location, hiring actors, and doing a full fledged on location shoot to get essentially a similar shot? Yeah, and what's more you can create the perfect version of your character if you iterate to get what you want in the original image generation and take the time to train proper LORAs.
Very rubbery. No regard for underlying bone or muscle structure. But hey, its early days, its only gonna get better. Good enough for memes and amateur work, I guess.
I think it is over. All music and film was about relating to fellow human beings. This AI is making everything oversaturated and quite frankly pointless. Doesn't make me feel like watching anything. It is just a pile of plastic straws at this point.
@@curiousrefuge Great. It opened a market place with loads of opportunities and jobs for people. If you want people to use AI to replace photographers, filmmakers, writers musicians then It will basically kill the whole collaborative market. AI content was not made to improve anything but rather to destroy content creation economy and alienate people form each other. Soon AI won't even need people to write a prompt. I personally think they should ban AI generated content monetization to save content creator economy.
Not bad! But my technique is better. As I don't need any input videos, just my imagination. And I can get multiple performances, so I can actually direct the actors by choosing the performance that works best. /watch?v=iMXbSXSJqqs We're at the point now where the tech does not matter so much as how we use it.
It might have been better if you would've laid your original vocals over the final clip so we could see how well the lip sync holds up through the processing. No one wants to go through this whole workflow just to make a silent film of people flappin' their lips. This reminds me of all the face capture demos -- and I've watched 'em all -- no one demonstrating the tech ever records any dialogue or normal scenes of simple talking, they just pull crazy funny faces like that's supposed to be useful.
Best tutorial on Live Portrait I've seen yet, thanks for posting.
@@DrZaious Agreed. You can do this for free using comfyUI and Live portrait workflow.
Wow, thanks!
Thanks for the tutorial. it's the most simple solution out there for animating faces that I've seen!
For better results, I think it's better that you don't put in your prompt any "talking" "Saying" (mouth/lips movements) anything for the base video, so the mouth and face do not be doing any expression as base face video, so then will be modified from a "blank default/idle expression" of the face by the face controlled video, so the modification will happen to the same face and not a face that it's already doing/changing over time.
Did you have a stroke while writing this comment?
True!
THANK YOU!!! When you opened Comfy I almost clicked away. Its so so complex and I just wanna hire people to collab with at that point. But you broke it down SO well, I understood and it made so much more accessible thanks
Glad I could help!
Thanks for the tutorial. These lip-sync features are so important to advance the control over consistent characters. I hope more A.I. video platforms will implement this under the hood, without all the incredible complex Comfy parameters.
Glad it was helpful!
With the local version and adding an upscale node, you should be able to have it in a bigger resolution 😊
Great tip! Thanks friend!
Glad that I found your channel and tutorials, great stuff
Glad you enjoyed this!
That sweater on that background is ❤🔥
word.
Glad you liked it!
Thank you, Curious Refuge!
Our pleasure!
i would love to see your studio...how you setup and how you get the ring animation in the background! 🙂
Maybe we'll do a desk breakdown someday :)
Great tutorial! Thanks you!
You are welcome!
You were the best sweaters lol
Thanks for the info, some useful tips. Wondering how do you get around videos that need lip syncing?
Also, where can I buy a sweater like yours? :)
I respect the flexibility of ComfyUI, but only super geeks are going to even attempt to grok it. Higher level tools (even if Comfy is under the hood) will be needed for wide spread usage, IMO.
Awesome presentation. You are top notch, sir.
Thanks for watching!
thank you, this is good to try . I upload some AI films, 8 min, but will try this way also
Glad it was useful!
Many Thanks for all your help, Bro.
Always welcome
I'm a horrible actor and I don't want to record myself. I know this is going to sound incredibly lazy but I hope the option to prompt different emotions becomes available. I don't need Oscar worthy performances for the projects that I'm working on. It's great that's there's a lot option for controlling faces. Now only if we could have 3D poseable consistent characters we'll finally be able to have AI actors?
Metahuman with unreal.
Hedra AI - check it out.
@SykoActiveStudios thanks. Tell me more. :)
You can certainly try other things like Hedra!
Thank you! Can you add how to deal with the actual audio in the performance?
You'll have to bring it into your editor :)
Thanks for tutorial. You used close-up image and the result was almost good and acceptable. My question is, can the same result be obtained with medium or wide shot images? Have you ever tried it?
We will have to do some testing! But close up is always best results
Awesome
this is really cool
Glad you enjoyed this!
Can you also do a video about ‘Crew AI’? I think ‘Crew AI’ seems like a useful AI tool because it gives users a lot of control over their AI character’s facial expressions, just by moving some sliders around. It doesn’t really create animations, but it can change the facial expressions of characters in still images. I need to work with both still images and animations, so I like that there’s a tool that can modify characters in still images like this. Also, in terms of animation, I can see a tool like this being useful for creating a ‘start frame’ and ‘end frame’ for an AI animation.
We'll check it out!
So if you wanted to sync up with audio, I'd imagine you'd have to slow the final output by 25%? If that's the case you'd have to increase the fps by 25% (around 30fps) if your final final output is 24fps with synced video?
I just went through a nightmare like this yesterday.
I found that as long as I had the same frame rate on each video, I didn't need to increase the speed of the driving video by 125% initially. Also, I got better results, prompting for subtle head movements only and without prompting the character to speak into the camera.
Would def take some finesse in the editor!
@@EmilyNilsen Yup. Grab the fps from the driving video via the Video Info VHS node and plug that into the Video Combine VHS node. You shouldnt have to do any speed changing.
Thanks, great video!
Glad you liked it!
Hey! Great video. I have an unrelated question, kind of an odd one. What's your background? Are you using a greenscreen? If so why can't I see any green in your glasses? Is it not a green screen and just a straight up tv screen? Your shot is lit very nicely, and it matches the background very well! I just had that question lol
It's a projector :) Glad you enjoy the look!
Hey I want to buy your sweater asap, where can I buy it?
I like the new background!!!
Just to clarify if I've got this correct:
1. If you're uploading a control video to animate an image, you can move your head all over the place and the resulting video will mimic those movements.
2. But if you're uploading a control video to animate a video, you need to keep your head still as it will only apply your face to the video character that is already moving their head.
Is it correct to say the control video head movements will not affect the existing video character's head movements?
Curiois about this as well
Jump in our discord and we can help clarify!
Love that sweater!
Thanks!
Great video thank you
Glad you enjoyed it
would it not make more sense to also record the reference footage in front of a white wall or plain background vs clutter in the background? Surely that makes a difference?
It would likely help! But not totally necessary
@@curiousrefuge fair enough! 🤝
❤❤❤
Any site with face portrait videos that we can download ?
You can use Midjourney to make thme!
Thank you, ComfyUI is pretty much a game changer these days. Every serious AI video creator should consider learning it.
True! It's a big deal!
Can't get this to work. I've followed the tutorial step by step, and used the specified assets but the video combine node always appears EMPTY and I cannot render. How to fix this? Thanks in advance.
Feel free to jump in our discord and we can try and help you troubleshoot!
@@curiousrefuge Thanks.
What background are you using?
A projector :)
why there is no workflow downloadable?
It's not something to download!
@@curiousrefuge i mean the comfyui's workflow, simple json file, of course the models and everything else is separate download, i understand that. But why not workflow file?
What we need is faces at an angle and people who are moving
True!
How do I get the output video to be longer than 8s?
Editing software and splice clips together :)
can we run it locally on mac or windows without paying? how is this done?
Yes but we prefer running it online as it takes quite a bit of cpu power
What's the GPU requirement to run Live-portrait locally?
Good question...we'll have to check!
Why are people so into the sweater?!
Haha we have no idea!
Wowwwwwwwwwwwww
thanks for watching!
can we only do lipsync ? meaning, the app will ignore the eye.
It's not as refined for other facial features
Why does my output come out so damn blurry? Its so blurry i can barely see lip movement
Let us know in our discord and we'll try to help!
Tell the actor that you're going to use their single performance many times in different projects before you record them. :)
Hah, it's a MJ image!
Gen 3 Alpha is not free 😔
True! Not free for Runway!
i dont like ComfyUI, that's why I like to use LivePortrait in "LivePortrait: Efficient Portrait Animation with Stitching and Retargeting Control".
What? What are you talking about?
Is there an alternative to do what was shown in this video without comfy UI and doesn't involve installing Live portrait on the PC? (Which I had issues with)
@@adarwinterdror7245 aitrepreneur made a tutorial about this topic. It’s his second most recent video
@@adarwinterdror7245 a it re pr en eu r made a tutorial about this topic. It’s his second most recent video.
I had to write spaces in between. As my comment disappears when writing his channel name
@@adarwinterdror7245 Find the link in my post (as UA-cam comments get deleted all the time)
The post will expire in 24 hours
@@adarwinterdror7245find the name in my post
AI gfs are gonna be f'in wild
Hahha, wow!
Im not sure if my scrept needs that. Lol
Maybe one day it will :)
Still not perfect though need another 6 months to a year i think
Indeed! We agree!
Its not working for me when i upload an image of an ai created cat. It doesn’t recognize it.... so ai is still behind
Hmm, jump in our discord and we can try to help!
AI is really good at generating non-ugly women
True! It's trained on a bunch of models it seems for all genders!
Not easy at all lol
We appreciate you trying!
Every AI tool is the best!
If you want easy, Runway has Act-One for talking heads. This however looks like it might be beneficial if you were doing a walk and talk video from an image to video and you needed a better facial performance on the character. Is it easy? No. Is it less expensive and easier than hiring a location, hiring actors, and doing a full fledged on location shoot to get essentially a similar shot? Yeah, and what's more you can create the perfect version of your character if you iterate to get what you want in the original image generation and take the time to train proper LORAs.
Very rubbery. No regard for underlying bone or muscle structure. But hey, its early days, its only gonna get better. Good enough for memes and amateur work, I guess.
It's certainly still developing!
Lora dataset augmentation. Great for that
Looks too artificial, maybe in w few next versions they will fix it.
We def get better!
Still not good enough for my taste, but one or two more papers down the line, or if the EMO paper gets put to use I'm game.
True, still needs a little work!
I think it is over. All music and film was about relating to fellow human beings. This AI is making everything oversaturated and quite frankly pointless. Doesn't make me feel like watching anything. It is just a pile of plastic straws at this point.
Nah. That was done away with when it costs so much to create a vision. Money made us lose the connection.
Before AI, how did you feel about online content?
@@curiousrefuge Great. It opened a market place with loads of opportunities and jobs for people. If you want people to use AI to replace photographers, filmmakers, writers musicians then It will basically kill the whole collaborative market. AI content was not made to improve anything but rather to destroy content creation economy and alienate people form each other. Soon AI won't even need people to write a prompt. I personally think they should ban AI generated content monetization to save content creator economy.
Not bad!
But my technique is better. As I don't need any input videos, just my imagination.
And I can get multiple performances, so I can actually direct the actors by choosing the performance that works best.
/watch?v=iMXbSXSJqqs
We're at the point now where the tech does not matter so much as how we use it.
Thanks for sharing!
This is bad.
It's come a long way!
It might have been better if you would've laid your original vocals over the final clip so we could see how well the lip sync holds up through the processing. No one wants to go through this whole workflow just to make a silent film of people flappin' their lips.
This reminds me of all the face capture demos -- and I've watched 'em all -- no one demonstrating the tech ever records any dialogue or normal scenes of simple talking, they just pull crazy funny faces like that's supposed to be useful.
You've watched them all? We've seen quite a few of people using their real voice (and typically adding some flavor with elevenalbs with v2v)