Open RK-BAKU opened 2 years ago
Loading model with torch.hub.load() always loads model as fp16 which is not supported on cuda thus results in slower inference. Is there any option in kwargs to disable half() during loading? Thanks in advance!
@RK-BAKU how do you load the model with torch.hub.load()? For me it doesn't even load
Loading model with torch.hub.load() always loads model as fp16 which is not supported on cuda thus results in slower inference. Is there any option in kwargs to disable half() during loading? Thanks in advance!