intel-analytics / text-generation-webui

A Gradio Web UI for running local LLM on Intel GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max) using IPEX-LLM.
GNU Affero General Public License v3.0
14 stars 8 forks source link

Add bigdl-llm loader to bigdl-upstream #17

Closed chtanch closed 5 months ago

chtanch commented 6 months ago

Description:

Add BigDL-LLM loader to upstream main

Environment setup

pip install --pre --upgrade bigdl-llm[all] cd \<text-generation-webui> pip install -r requirements_cpu_only.txt

Run WebUI

python server.py --load-in-low-bit sym_int4 --share

Note:

Tested models

Successful - chatglm3-6b, mistral-7B-v0.1, qwen-7b-chat, falcon-7b-instruct-with-patch Fail - llama2-7b-chat-hf. Reason is upstream main's requirements*.txt has updated transformers version to 4.38.2, which is unsupported.

shane-huang commented 5 months ago

update bigdl-llm to ipex-llm.

hkvision commented 5 months ago

@jason-dai @shane-huang The current version as the first commit for upstream seems good, merge it first.