-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Write an actual worker script #2
Comments
We should try to make the test work with plain git and with git-lfs as well. It needs to detect if git-annex is in use before running |
I've done some preliminary explorations tonight to optimize I noticed that gitea has a folder I grabbed a dataset I had and uploaded it to my test instance, putting it here:
Everything is singly-linked (notice:
I picked out this file to examine the hardlink status of under different conditions:
I wrote this test script to see how different combinations of flags would result in space usage:
The difference between
--shared, annex.hardlink => +397MiB
--shared, annex.thin => +397MiB
--shared, both => +397MiB
--shared, neither => +397MiB
--local, annex.hardlink => +397MiB
--local, annex.thin => +397MiB
--local, both => +397MiB
--local, neither => +738MiB
And if we use /tmp instead of data/tmp/ then --shared, annex.hardlink => +727MiB
--shared, annex.thin => +387MiB
--local, annex.hardlink => error
--local, annex.thin => error
So it is impossible to get a zero-copy dataset, at least not with |
I tried to see if I could get zero-copies by using
Maybe there's a way to trick it but this seems like a dead-end. And a risky dead-end too, because I'm pretty sure using
so that's not good. |
In retrospect, it seems that the best option is to stick with And to further simplify, we can default to using /tmp (but have a way for sysadmins to configure that for themselves if they need more working storage than their system's RAM). |
One that:
ansifilter --html --fragment --line-numbers --anchors=self
See the current worker script for a skeleton.
The text was updated successfully, but these errors were encountered: