ViperX7 / Alpaca-Turbo

Web UI to run alpaca model locally
GNU Affero General Public License v3.0
876 stars 92 forks source link

can't not load model.. #34

Open melodysdreamj opened 1 year ago

melodysdreamj commented 1 year ago

스크린샷 2023-04-05 오후 9 12 20 스크린샷 2023-04-05 오후 9 12 25

wait 1 hours, but did not load model..

in windows10

JosephSBoyle commented 1 year ago

I'm having a similar problem. Win10, ggml Alpaca 7B downloaded from huggingface: https://huggingface.co/Pi3141

ViperX7 commented 1 year ago

can you try the alpaca-native-enhanced model and can you tell me how many cores are there in your system

Teragron commented 1 year ago

I've had the same problem with the "ggml-model-q4_0.bin" but had no issue with "ggml-model-q4_0_unfiltered.bin". So it's probably about the model.

llkj11 commented 1 year ago

Same issue here with: gpt4all-lora-quantized.bin and ggml-alpaca-7b-q4.bin

OzGoober commented 1 year ago

Using 2 files from Pi3141, gpt4-x-alpaca-native-13B-ggml worked, alpaca-native-7B-ggml didn't load

gpt4-x-alpaca-native-13B-ggml: 'models\ggml-model-q4_1.bin', Working... ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 100% 0:00:00

alpaca-native-7B-ggml: 'models\ggml-model-q4_0.bin', ERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoR

andresdelcampo commented 1 year ago

Similar, won't load any of my 3 quantized 7B and 13B alpaca variants that worked in dalai or alpaca.cpp.

... '-m', 'models\ggml-model-q4_0.bin', '--interactive-first' ] ERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoRERRoR

boehm-e commented 1 year ago

Same here ggml-alpaca-7b-q4.bin not loading

AndreasKunar commented 1 year ago

Same here, llama7B, llama13B, alpaca,... - all working locally with llama.cpp on the commandline. All hanging on load.

Parameters for invoking llama.cpp commandline seem right and commandline status shows apparent completion: Working... ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 100% 0:00:00

But web-ui dialogue hangs with loading.

EDIT: running on macOS/Apple Silicon via current git-clone + copying templates folder from ZIP.

sjackp commented 1 year ago

Same here, endless -loading model "ggml-model-q4_0.bin"

AndreasKunar commented 1 year ago

Updated / totally edited for better clarification.

Cause of model-hang for me: Alpaca-Turbo expects a certain string "RSTsr" to be displayed by it's invocation of llama.cpp once its finished loading (set in alpaca_turbo.py line 46). The current llama.cpp + llama 7B e.g. does not show this string in its invocation. So the load hangs. For me it seems, it's not how the model is, it's more how Alpaca-Turbo was constructed. The current llama.cpp+models produces very different responses - I'm trying to understand alpaca_turbo.py, so that maybe I can fix it (at least for me).

Hope this clarifies/helps.

AndreasKunar commented 1 year ago

Update - I got it to work (most of the time) on my Mac by changing alpaca_turbo.py quite a bit. But I don't think it is mergeable into a pull-request, because my solution seems to be one-off just for my situation.