Closed katopz closed 5 months ago
Seems the rustls plugin is not required. Please try the following command to reinstall wasmedge with ggml
+ rustls
plugins:
curl -sSf https://raw.githubusercontent.com/WasmEdge/WasmEdge/master/utils/install.sh | bash -s -- -v 0.13.5 --plugins wasi_nn-ggml wasmedge_rustls
I reinstalled wasmedge with ggml
+ rustls
plugins, rebuilt rag-api-server.wasm
, and ran
wasmedge rag-api-server.wasm -h
,
but got
[2024-05-14 23:07:43.871] [error] instantiation failed: module name conflict, Code: 0x60
[2024-05-14 23:07:43.871] [error] At AST node: module
LlamaEdge-RAG API Server
Usage: rag-api-server.wasm [OPTIONS] --model-name <MODEL_NAME> --prompt-template <PROMPT_TEMPLATE>
Options:
-m, --model-name <MODEL_NAME>
Sets names for chat and embedding models. The names are separated by comma without space, for example, '--model-name Llama-2-7b,all-minilm'
-a, --model-alias <MODEL_ALIAS>
Model aliases for chat and embedding models [default: default,embedding]
-c, --ctx-size <CTX_SIZE>
Sets context sizes for chat and embedding models. The sizes are separated by comma without space, for example, '--ctx-size 4096,384'. The first value is for the chat model, and the second is for the embedding model [default: 4096,384]
-p, --prompt-template <PROMPT_TEMPLATE>
Prompt template [possible values: llama-2-chat, llama-3-chat, mistral-instruct, mistrallite, openchat, codellama-instruct, codellama-super-instruct, human-assistant, vicuna-1.0-chat, vicuna-1.1-chat, vicuna-llava, chatml, baichuan-2, wizard-coder, zephyr, stablelm-zephyr, intel-neural, deepseek-chat, deepseek-coder, solar-instruct, phi-2-chat, phi-2-instruct, phi-3-chat, phi-3-instruct, gemma-instruct, octopus]
-r, --reverse-prompt <REVERSE_PROMPT>
Halt generation at PROMPT, return control
-b, --batch-size <BATCH_SIZE>
Batch size for prompt processing [default: 512]
--rag-prompt <RAG_PROMPT>
Custom rag prompt
--rag-policy <POLICY>
Strategy for merging RAG context into chat messages [default: system-message] [possible values: system-message, last-user-message]
--qdrant-url <QDRANT_URL>
URL of Qdrant REST Service [default: http://localhost:6333]
--qdrant-collection-name <QDRANT_COLLECTION_NAME>
Name of Qdrant collection [default: default]
--qdrant-limit <QDRANT_LIMIT>
Max number of retrieved result (no less than 1) [default: 5]
--qdrant-score-threshold <QDRANT_SCORE_THRESHOLD>
Minimal score threshold for the search result [default: 0.4]
--chunk-capacity <CHUNK_CAPACITY>
Maximum number of tokens each chunk contains [default: 100]
--log-prompts
Print prompt strings to stdout
--log-stat
Print statistics to stdout
--log-all
Print all log information to stdout
--socket-addr <SOCKET_ADDR>
Socket address of LlamaEdge API Server instance [default: 0.0.0.0:8080]
--web-ui <WEB_UI>
Root path for the Web UI files [default: chatbot-ui]
-h, --help
Print help (see more with '--help')
-V, --version
Print version
I tried to rename mod backend
, error
, utils
, ggml
with their corresponding file names, then rebuilt the program, but unfortunately that did not solve the problem.
@ChloeWKY The error message at the beginning can be ignored. The coming WasmEdge v0.14.0
will fix it.
Seems the rustls plugin is not required. Please try the following command to reinstall wasmedge with
ggml
+rustls
plugins:curl -sSf https://raw.githubusercontent.com/WasmEdge/WasmEdge/master/utils/install.sh | bash -s -- -v 0.13.5 --plugins wasi_nn-ggml wasmedge_rustls
Thanks, working now.
@ChloeWKY The error message at the beginning can be ignored. The coming
WasmEdge v0.14.0
will fix it.
So this is related? https://github.com/LlamaEdge/Example-LlamaEdge-RAG/issues/1 Seem like an example is also outdated?
Yeah. The example repo was created at the very early stage of rag-api-server project. After that, a lot of changes were introduced into rag-api-server. However, the example repo failed to keep the same pace with the project. We'll catch up asap.
Yeah. The example repo was created at the very early stage of rag-api-server project. After that, a lot of changes were introduced into rag-api-server. However, the example repo failed to keep the same pace with the project. We'll catch up asap.
Many thanks! Hope this will sort out soon. I will talk about LlamaEdge
RAG at AWS Thailand events 11th next month (+30 dev). No pressure 😅 just really want to make WasmEdge
look pretty at that time. 🤗
run
got
Not sure what I miss here 🤔