Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

LauraTTS模型的训练花了多长时间? #36

Open
Dinxin opened this issue Mar 29, 2024 · 5 comments
Open

LauraTTS模型的训练花了多长时间? #36

Dinxin opened this issue Mar 29, 2024 · 5 comments

Comments

@Dinxin
Copy link

Dinxin commented Mar 29, 2024

No description provided.

@ZhihaoDU
Copy link
Collaborator

ZhihaoDU commented Mar 29, 2024

The model is trained about 1.5 day on LibriTTS clean subset with an A800 GPU, and the batch size is 10240 tokens.

@Dinxin
Copy link
Author

Dinxin commented Mar 29, 2024

on 8 A100 GPUs? The total duration is 6000 hours ?

@ZhihaoDU
Copy link
Collaborator

Only one A800 GPU. I think the duration of LibriTTS clean subset is about 244 hours

@a897456
Copy link

a897456 commented Apr 12, 2024

The model is trained about 1.5 day on LibriTTS clean subset with an A800 GPU, and the batch size is 10240 tokens.

Hi @ZhihaoDU
10240 tokens?How do you calculate it?
speech_max_length / encoder_hop_length * batch_size = 40960 / 320 * 16 = 2048?

@a897456
Copy link

a897456 commented Apr 12, 2024

@ZhihaoDU please
数据集132000个文件,如果设置batch_size=8, 两者相除等于16250,这是不是意味着如果还是保持num_iters_per_epoch=10000,这是不合适的。你可以分享一下,batch_size、num_iters_per_epoch、num_works怎么计算才比较合理?
另外你设置的input_size=1,也参与计算吗?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants