hltcoe / sandle

Run a large language modeling SANDbox in your Local Environment
Other
7 stars 1 forks source link

DeepSpeed #88

Open ccmaymay opened 1 year ago

ccmaymay commented 1 year ago

w/ bf16

ccmaymay commented 1 year ago

https://github.com/microsoft/DeepSpeed-MII#deploying-mii-public

Has grpc server, seems to resemble our own https://github.com/microsoft/DeepSpeed-MII/blob/main/mii/server.py for example:

        # XXX: fetch model provider based on model name in a more general way
        if model_name == "gpt-neox":
            provider = mii.constants.MODEL_PROVIDER_NAME_EA
        elif ("bigscience/bloom" == model_name) or ("microsoft/bloom" in model_name):
            provider = mii.constants.MODEL_PROVIDER_NAME_HF_LLM
        elif self.task == mii.Tasks.TEXT2IMG:
            provider = mii.constants.MODEL_PROVIDER_NAME_DIFFUSERS
        else:
            provider = mii.constants.MODEL_PROVIDER_NAME_HF
        server_args_str += f" --provider {provider}"
ccmaymay commented 1 year ago

Branch status: event loop not found at query time

ccmaymay commented 1 year ago

Looks like DeepSpeed-MII recently removed the Bloom-specific LLM functionality and replaced it with a generalized "meta tensor" functionality: https://github.com/microsoft/DeepSpeed-MII/commit/60a85dc3da5bac3bcefa8824175f8646a0f12203

may be worth trying out again