Open garyttierney opened 3 months ago
At this point, I'm unsure if the write_npz method needs to be included in the library. Would it be sufficient to implement this method within the benchmark code, or could you extend it to handle more general cases, such as accepting a dataset name and other parameters?
Additionally, I'm curious about the proper compression options for npz files. Will using custom options cause any compatibility issues? This option is currently used.
At this point, I'm unsure if the write_npz method needs to be included in the library. Would it be sufficient to implement this method within the benchmark code, or could you extend it to handle more general cases, such as accepting a dataset name and other parameters?
Yeah I think it would make more sense to move it to the benchmark code, although we code make the to_npz
code in recorder
call this too and make sure it can handle the options.
Additionally, I'm curious about the proper compression options for npz files. Will using custom options cause any compatibility issues? This option is currently used.
Those should be the same options used as defaults, although I'll explicitly add them to the benchmark code.
One thing I'd like to get in before this is ready is a memory-mapped reader and compare throughput vs the additional syscalls + buffer copy.
This is a WIP and just some groundwork to get some figures on the recent conversations around improving backtest performance.
cargo bench
will produce a report on the command-line with a comparison against the previous run if available: