We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
In order to avoid ulimit issues, it would be useful to have an option that limits the number of open files. Maybe open files lazily?
ulimit
cc. @VibhuJawa
The text was updated successfully, but these errors were encountered:
For context, I have seen this error the most while writing partitioned datasets. Don't know how it impacts there.
dask_df.to_parquet(partition_on=["xyz"])
And if #410 helps in that case too.
Sorry, something went wrong.
Should help with issues mentioned in the comments here:
NVIDIA/NeMo-Curator#157 (comment)
No branches or pull requests
In order to avoid
ulimit
issues, it would be useful to have an option that limits the number of open files.Maybe open files lazily?
cc. @VibhuJawa
The text was updated successfully, but these errors were encountered: