Open vv12kant opened 1 year ago
I guess it results from transformers/modeling_bloom and it misses a type adaption.
I am seeing the same issue with llama 2. Were you able to get this working?
I guess it results from transformers/modeling_bloom and it misses a type adaption.
hi guys. i encounter same issue and i guess this is not because of specific modeling_xxx.py
.
in my case, when i casting model with bfloa16 and use single GPU, activations of all layers were bf16, but when i use multiple GPU with zeropp i saw activations were fp16.
so i think it is deepspeed's issue.
Does DeepSpeed supports bf16 now?
i want to use this function but can not 😥
Describe the bug When using both the Zero++ and BFloat16 features simultaneously. Sometimes the gathered param is Float15 dtype,but the intermediate result are still BFloat16 dtype.
To Reproduce Steps to reproduce the behavior:
set deepspeed configuration,enable zero++ and bf16.
Create a Bloom HF model,and initialize model engine by ds_config.
run inference.
See error.