-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
take a look at this #31
Comments
I did see it - I'll add to readme. seems a complete rip off from MooreThread animateAnyone - Poseguider https://github.com/MooreThreads/Moore-AnimateAnyone/blob/master/train_stage_1.py#L54 |
maybe you can refer to their model code,at least we have that,training code seems just like animate anyone |
looking at the saved models looks less scary than this mess can probably just load these into simpler architecture. I think the reader /writer is just to have the paralllel unet (to dig out the features from referencenet - reader - and throw to backbone) |
yes,i also think reader /writer is used to implement The image of the target character is inputted into the ReferenceNet to extract the reference feature maps outputs from the self-attention layers. During the Backbone denoising procedure, the features of corresponding layers undergo a reference-attention layers with the extracted feature maps.(from paper's description)
and the second link seems just like pretrained weights from Moore-Animateanyone (another talking head generation repo).Now we're also trying to implement emo,we refer to your existing repo. and I'm finishing Face Locator today.Thanks!
------------------ 原始邮件 ------------------
发件人: "johndpope/Emote-hack" ***@***.***>;
发送时间: 2024年3月27日(星期三) 上午9:55
***@***.***>;
***@***.******@***.***>;
主题: Re: [johndpope/Emote-hack] take a look at this (Issue #31)
looking at the saved models looks less scary than this mess
MStypulkowski/diffused-heads#21Screenshot.from.2024-03-27.12-51-48.png (view on web)
can probably just load these into simpler architecture.
I think the reader /writer is just to have the paralllel unet (to dig out the features from referencenet - reader - and throw to backbone)
https://github.com/Zejun-Yang/AniPortrait/blob/main/train_stage_1.py#L53
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you authored the thread.Message ID: ***@***.***>
|
did you see this? #28 - I think we can just piggy back off the Alibaba pretrained unet model - |
ok ,i'll take a look. we can use pretrained model,in fact Alibaba also uses pretrained model (from hugging face Stable Diffusion v1.5) reference-net and backbone inherit weights from the original SD UNet,only attention layers were changed
…------------------ 原始邮件 ------------------
发件人: "johndpope/Emote-hack" ***@***.***>;
发送时间: 2024年3月27日(星期三) 上午10:16
***@***.***>;
***@***.******@***.***>;
主题: Re: [johndpope/Emote-hack] take a look at this (Issue #31)
did you see this? #28 - I think we can just piggy off the Alibaba pretrained unet model -
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you authored the thread.Message ID: ***@***.***>
|
that's why I'm thinking it will be plug and play. Got all the models for AniPortrait - |
yes,maybe just a combination of existing methods and modules. I'll take a look later.
…------------------ 原始邮件 ------------------
发件人: "johndpope/Emote-hack" ***@***.***>;
发送时间: 2024年3月27日(星期三) 上午10:56
***@***.***>;
***@***.******@***.***>;
主题: Re: [johndpope/Emote-hack] take a look at this (Issue #31)
that's why I'm thinking it will be plug and play. Got all the models for AniPortrait -
check this helper out
xmu-xiaoma666/External-Attention-pytorch#115
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you authored the thread.Message ID: ***@***.***>
|
the aniportrait is good. I thought the ControlNetMediaPipeFace maybe best solution @chris-crucible - it seems that they have enhanced the lips from the base media pipeline
I get that there's still expression issues here. but result is quite good. The head rotations maybe nice branch to get the emotion into video. |
Have you run through the entire process? Congratulations! Let me take a look at the repo and code!
------------------ 原始邮件 ------------------
发件人: "johndpope/Emote-hack" ***@***.***>;
发送时间: 2024年3月29日(星期五) 中午1:36
***@***.***>;
***@***.******@***.***>;
主题: Re: [johndpope/Emote-hack] take a look at this (Issue #31)
Screenshot.from.2024-03-29.16-32-04.png (view on web)
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you authored the thread.Message ID: ***@***.***>
|
python ./scripts/vid2vid.py --config ./configs/prompts/animation_facereenac.yaml -W 512 -H 512 -L 256 animation_facereenac.yaml
there's no speed embedding - so the vanilla image to video will hold the face in video mostly - but because they're using the animateanyone framework - they get video2video out of the box - allowing this |
hey @fenghe12 i had some success with megaportraits and now attempting to integrate into VASA on this branch. |
VASA adopts DiT as backbone denoising network,but it lacks more details about how to integrate conditons into dit. I attempted to replace the unet in Moore animateanyone with DIT (Latte, a video generation model), but the results were not satisfactory. We are now trying to train a talking face video generation model based on OpenSora-plan. |
I guess DiT will be mainstream video generation architecture,because of SORA |
Maybe I can offer some help
|
i attempt to port matmulfree for llm to pytorch https://github.com/ridgerchu/matmulfreellm i send you a link - could be more exciting if i can get cuda code working. |
sorry but i can't click your invitation link.It told me 404 error
…------------------ 原始邮件 ------------------
发件人: "John D. ***@***.***>;
发送时间: 2024年6月12日(星期三) 晚上7:16
收件人: ***@***.***>;
抄送: ***@***.***>; ***@***.***>;
主题: Re: [johndpope/Emote-hack] take a look at this (Issue #31)
i attempt to port matmulfree for llm to pytorch
https://github.com/ridgerchu/matmulfreellm
i send you a link - could be more excited if i can get cuda code working.
—
Reply to this email directly, view it on GitHub, or unsubscribe.
You are receiving this because you were mentioned.Message ID: ***@***.***>
|
sorry that project was 3 days down the gurgler - learned how to compile cuda code. not sure how to handle the audio stuff - wav2vec - |
how can i help you? |
https://github.com/Zejun-Yang/AniPortrait a new open source talking head generation repo.It seems it's very similar to Emote
The text was updated successfully, but these errors were encountered: