Closed carlini closed 6 months ago
Thanks for your attention!
For the first part, thank you for finding the typo and we will fix it.
For the second part, BunnyPhiConfig
is defined here. If you install this project as INSTALL in README, it shouldn't be a KeyError
. Could you please share more information?
For the third part,
python -m bunny.serve.model_worker \
--host 0.0.0.0 \
--controller http://localhost:10000 \
--port 40000 \
--worker http://localhost:40000 \
--model-path /path/to/bunny-phi-2-siglip-lora \
--model-base /path/to/microsoft/phi-2 \
--model-type phi-2
Ah okay the third point fixes both the second and third.
If you don't pass --model-base to the phi-2 path then you get the error, and if you do pass it then it works correctly. Thanks for the help.
I'm trying to spin up the server so I can run this for inference as described in the README and I've hit a few issues.
First: demo_3.png and demo_4.png don't exist. This is easy to fix here:
https://github.com/BAAI-DCAI/Bunny/blob/516437ebf250b99e3967b0fb1ad5cb6bc8b43953/bunny/serve/gradio_web_server.py#L348-L349
should be edited to example_1.png and example_2.png.
Second: (and why this isn't just a PR) I can't get the service to start with the model to load phi-2. I'm just trying to get the inference demo working.
If I run the
model_worker
service with--model-type phi-2
then I get a crashKeyError: 'BunnyPhiConfig'
when it tries to load the tokenizer. It looks like you try to configure this config somewhere but it doesn't get added to the huggingface transformers list of known configs for some reason.Are there other steps required (e.g., modifying the huggingface code)?
Third: I don't understand what set of model paths I should be passing to run the service if I don't want to fine-tune anything. Could you give an example for what
model-path
should be? I've downloadedbunny-phi-2-siglip-lora
and I'm passing this as the path, but I can't test this because of the prior crash.I'm pretty sure I have the correct versions of everything installed. Have you tried following the readme on a clean machine install to verify it runs as expected?