Open yumianhuli1 opened 8 months ago
Does inference currently only support torch_dtype=torch.float16?int8_float16、int8 will be supported?
Does inference currently only support torch_dtype=torch.float16?int8_float16、int8 will be supported?