Thank you for such a comprehensive tutorial. Would also love to see the improved version update that you mentioned @13:10 And is there any other tool to transfer facial expressions from one talent to other, rather than just giving the audio to produce those facial expressions?
Great Video. Do you have a video on running it in your local machine? Also do you have a video on using an API liek you mentioned @ 9:53 in your video?
Interesting! Thanks. I had tried the Collab several times and failed. I wonder if that's because i didn't copy it to my Google drive? It wouldn't notice the images nor audio I added.
can we use this generated video on social media or personal youtube channel...? is this alternative of what you showed yoyo-nb/thin-plate-spline-motion-model
Thanks for the video. Have subbed and liked vid. I have a Q, for Google Colla free plan, what's the maximum amount of audio it can process. For example if I have a 1-hour audio, can I still use the free plan?
The licence says opensource and free for commercial use too. Why we should read all the content of the code searching for a disclaimer that say opposite? In my opinion when I use a code is my responsibility only to read the licence and don't violate it. It is not my responsibility to read all the content of the code. Why do you think that this disclaimer putted somewhere into the code package has some kind of legal value?
Can the SadTalker software work in real time? That is to say, if the sadTalker were to receive a string of text, could it manipulate the face in real time, or does the text string need to be processed on-line by remote server? Please advise.
Anyone knowing how much time it takes to generate that kind of video approximately (assuming you have a decent CPU and GPU, like i7 and RTX2070 for example) ? I can't find this anywhere, and as I'm testing several solutions, I'd rather find the information than trying all of them 😅
Can we make it making the video faster so it will be suitable for live applications? Is there is any alternative model that can do the same in real time?
@@engineerprompt i run it on colab for short audio but it took around 38 second to produce the first video(low quality) and took around 4 minutes to make it high quality, please can you refer me to any resource to learn how to make it faster (maybe make them model run on multiple gpus!)
@@msalama I mentioned in the video that its going to be slow if you are running it on FREE collab. In order to run it faster, you will run it high end GPUs locally.
How do you move the shoulders, only moving the face is very unnatural, very often I bounce when I see an avatar only taks with face without moving the body
I created a copy for Drive and the first video I made was successful, about 1 and a half minutes long. However, even repeating all the processes, I was no longer able to generate even a 30-second video. The error that appears to me is when executing the fifth cell, when I want to see the preview and download it. It simply says the following: "IndexError: list index out of range". I don't know what else to do, if you can help me I would appreciate it in advance.
cool BUT the only runtime option a got was python 3, and after I selected the image I wanted to use, the animation script errored out with "/bin/bash: line 1: python3.8: command not found"
Just saw this tutorial is 2 months old... in AI terms that is 2 years old... Sadly outdated as most AI tutorials become if you blink haha... sure worked great a month ago or so... currently all local running methods are non-functional...
@@ajmalkhan8013 I was finally able to get it running locally direct via their gradio... I was hoping to get the automatic1111 extension functional but currently it is inoperable... I'm sure there is a work around but after a few tests using sadtalker i realized it wasn't what I was looking for... I will wait another month to see where it goes hahah... but as a FREE option it is pretty decent...