Open jason-fries opened 5 years ago
Bumping this, as currently saving/logging models with pickle
will fail for any models > 4GB: https://stackoverflow.com/questions/29704139/pickle-in-python3-doesnt-work-for-large-data-saving. This is especially problematic with high-dimensional outputs (e.g. label models that spit out ~1000 different types of labels)
Currently load and save operate directly over pickles. This causes issues when trying to load models across devices (GPU->CPU). These calls should wrap
torch.load
andtorch. load_state_dict
in some configuration based on whatuse_cuda
flag is provided to the model. See. https://pytorch.org/tutorials/beginner/saving_loading_models.html#saving-loading-model-across-devices