-
### Bug Description
I use this command to bring vllmserverr up
```
python -m vllm.entrypoints.openai.api_server --model /hy-tmp/hub/models--Phind--Phind-CodeLlama-34B-v2/snapshots/949f61e203f91b4…
-
I was running the script https://github.com/noahshinn/reflexion/blob/main/programming_runs/run_reflexion_codellama_multi.sh
with CodeLLaMA model, simply change the `codellama` to `codellama-7b`
```b…
-
Currently the models need to be specified as `llama7b` for example, but what if one wants to use `codellama/CodeLlama-7b-hf` or `meta-llama/Llama-2-7b-hf` (non chat version), etc.?
A more flexible me…
-
is it possible to run the subject?
-
With a simple prompt like 'give me an image of a sunrise' it shows:
Here is an image of a sunrise:
![Sunrise](https://i.pinimg.com/originals/09/32/1c/09321c45876f2d8b0e0a203993160992.jpg)
How…
-
exllama/model.py", line 45, in __init__
self.pad_token_id = read_config["pad_token_id"]
KeyError: 'pad_token_id'
-
I was trying the script with CodeLLaMA
Sometimes I found the script just got killed, without showing any error. Any intuition?
![image](https://github.com/noahshinn/reflexion/assets/3351187/77bb63…
-
### What is the issue?
i'm trying to use codellama. i get inconsistent experience depending on the question and i have no idea what is causing it. if i ask simple questions like "what is the capita…
-
I just used base script from the TheBloke repo and got an error during loading the model:
OSError: libcudart.so.12: cannot open shared object file: No such file or directory
my script is :
from…
-
When I use codellama or codellama-python to finish the continuation of a prompt, a lot of '\n' are outputed in the end until it reaches the max_gen_len. Is there any way to let it stop early except li…