Open mengyao00 opened 10 months ago
Hi @mengyao00 ,
Currently we do not support using DeepSpeed for inference. I think the main advantage of DeepSpeed is about reducing training GPU memory usage. For inference, traditional data parallel is enough.
Yes, in fact, our evaluation results are all based on fp16.
Describe Model I am using (UniLM, MiniLM, LayoutLM ...):e5 Can I use deepspeed for evaluating beir benchmark? Or does it support other strategies? Can I use fp16? Would it impact results?