-
https://huggingface.co/anon8231489123/vicuna-13b-GPTQ-4bit-128g/discussions/36#6441d80a55a16ae60fa3f2b6
Could the same thing be done for vicuña gptq 4 bit model? I am very new to this stuff, so not …
-
Which LLaMA 7b model was used?
Is it available on Hugging face?
Thank you. I am trying to reproduce your success
-
不支持 n_gpu_layers 参数控制装载的层数吗?多实例环境对推理速度要求不太高的场合,哪怕每个实例少装载 4~5 层也能节省很多 GPU
-
### Description
clicking start a task on an agent, the run status and the text log of what is happening never appears.
I can see the log in the dockerlog that seems the backend is making requests …
-
Supporting a vector database like ChromaDB would have numerous benefits including a longer memory.
-
as for the quesiton:
Requirements
GPU is not used and is not required.
I have gpu and I expect to run model faster, but your code is only for cpu? how to change it?
-
### Describe the issue
Issue: It keep telling, "NETWORK ERROR DUE TO HIGH TRAFFIC. PLEASE REGENERATE OR REFRESH THIS PAGE." when click send with the example image & prompt. No detailed error is shown…
-
OS: Ubuntu 23.10
GPU: Nvidia 3070
Nvidia NVCC version 12.0
Build: cuda_12.0.r12.0/compiler.32267302_0
nvidia-smi is working.
gcc version: 13.2.0
g++ version: 13.2.0
References:
1. h…
-
### 相关问题
_No response_
### 可能的解决办法
您好,我是 RWKV 的作者,介绍见:https://zhuanlan.zhihu.com/p/626083366
目前支持 RWKV 的中英文界面有:
* 闻达:https://github.com/l15y/wenda
* Gradio DEMO: https://huggingface.co/spaces/…
-
### Describe the bug
It appears that the issue lies with the most recent iterations of Oobabooga. Notably, examining the two images provided below reveals that the latest version fails to accurately …