-
I am curious what is required to apply this method to the 70B parameter version of the llama2 model?
On reddit, noticed you mention: "For training, these models barely fit in 128 80GB A100s using Dee…
ghost updated
10 months ago
-
Running into the same error on the 13b and 70b chat models. Using a h100 80GB card. The 7b chat model works fine.
Command (13b):
`torchrun --nproc_per_node 2 example_chat_completion.py --ckp…
-
These problems can be resolved by:
1.
> Further testing shows that quantization with llm_attacks is possible by using:
>
> * transformers==4.31.0
> * fschat==2.20.0
> pip will y…
-
When I run the python mainly the first agent works fine but when its time for the next agent do its task i come up with this error
python3 main.py …
-
How many shots do you use to test advGLUE?
-
When I try to run the llama2 model card that I trained, with chat-ui in Space using Nvidia A10G small, I get no response.
I see the following error in the container log:
"""
07:43:00 3|index …
-
Because I heard that CUDA is not actually needed when CRAG is actually running. Is that so?
> there is no NIVIDA CUDA on the Mac Apple Silicon series computers.
errors:
```
Preparing metadat…
-
I noticed `components/resources/ai_chat_prompts.grdp` has all the strings as translateable="false" . This is probably because Llama2 has not great support for non-english text.
I assume Anthropic i…
-
I am using a single GPU(A10) to run Bloom-560m model fine-tune, error, how to solve? I found similar problems in other projects, but I didn't know how to solve the problems in alpaca
https://github.c…
-
参考 https://soulteary.com/2023/07/23/build-llama2-chinese-large-model-that-can-run-on-cpu.html
使用 Apple M2, 用最后的 docker `soulteary/llama2:runtime` 运行 `Chinese-Llama-2-7b-ggml-q4.bin`
```bash
main:…