Replies: 1 comment 4 replies
-
If you're opening a single non-chunked netcdf file for example, then the whole variable's array does still need to be loaded into memory before your computation can proceed. It's likely that the IO cost of loading this data into memory is what's taking the majority of the time, so that only applying the processing step to a small slice doesn't lead to much speedup. |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
I have two files; one is bigger simply due to its "time" dimension being longer, about 5 times longer. Thus, on average the computations I apply to the larger file take about ~5 times more than the smaller file. However, when I slice the time dimension of both files into a smaller number, the computation time is still very different with the sliced data from the larger file taking ~5 times longer even though all the dims are equal.
I am guessing even after slicing the larger files contains more stuff in the memory?
Once the file is read by xr.open_dataset (graf_data), the operations are like below:
Beta Was this translation helpful? Give feedback.
All reactions