-
-
Notifications
You must be signed in to change notification settings - Fork 11
Speed up data loading #24
Comments
the next code (as of the evening of Fri 2021-05-14), can get ~ 50 it/s and only 24 G of RAM usage without NWP loading (with 4 workers). |
The interpolation (even linear) takes a while. Replacing the linear interpolation for |
hmmm, maybe the issue is that
|
…Ps are still a little slow though (#24). About to try resampling 'step' in load_single_chunk()
I've implemented option 2, and it's helped a lot! |
More things to try:
|
So, we know that including NWP data slows training down by a factor of more the 2x.
|
Profiling each line in
|
oooh... looks like it's possible to significantly speed up the selection based on 'step' by first transposing so that 'step' is the first dimension. This gets the runtime down to 1.73 ms if always using the first init_time. Need to see if this speed up holds when using multiple init times based on t0. |
Nope, doesn't look like transposing gives us the same performance increase when selecting multiple init times. But, better news: I noticed that, when using NWPs, the code is pretty constantly loading from disk when To really speed things up, I think we perhaps need to re-create the NWP Zarr, so the data is stored more efficiently on disk (#26). |
Swapping back to the 'old', more thorough way of getting NWPs, gives us 47.8 it/s |
Can't launch sub-processes from the worker processes: daemonic child processes aren't allowed to have children :) |
Uh oh!
There was an error while loading. Please reload this page.
The text was updated successfully, but these errors were encountered: