Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

It surprisingly takes around 20minutes to generate features by dump_data. #180

Open
andylida opened this issue Feb 24, 2022 · 1 comment

Comments

@andylida
Copy link

Hi there. I followed the README file and export CFLAGS to boost the dumpdata process. I use librosa to downsample LJspeech and use sox to create PCM files.
It takes me around 20min to process a 7s long wav file. And also comparing to the size of my wav which is only 302kb, the processed feature sum up to a 4G files.
Is it working properly? Can anyone gives some suggestion? Thx!

@jmvalin
Copy link
Member

jmvalin commented Mar 21, 2022

if your training file is too small, dump_data will iterate over it to generate enough augmented data. In this case though 7s is just too short, but that's why it takes a long time.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants