-
```-- 4522 -- 20230930 23:34:53 FTUNE /model load 20853.9ms
downloads-refact-1 | -- 4522 -- FILTER explanation: initial loss too big calculated on a single file, threshold is 3.500. Likely
downl…
-
### 请提出你的问题
在做nlp 文档抽取 finetune 过程中,在几个轮次过后,会自动中断训练,中断的时机不确定
训练命令:
python3.7 finetune.py --device cpu --logging_steps 5 --save_steps 100 --eval_steps 100 --seed 42 --model_name_or_path uie-x-base …
-
within the docker (IMAGE: nvidia/cuda:12.1.0-devel-ubuntu22.04)
GPU: A100 40GB
TensorRT-LLM version: 0.10.0
flash-attn 2.5.9.post1
I quantize the phi3 model(phi-3-medium-128k-instrcut/), wi…
-
Dear all,
executing finetuning.py consistantly leads for me to the above error. I would be very thankful for a hint that leads me to the right direction. The log was taken from Ubuntu 22.04 machin…
-
Is there support for llama3.2 with TensorRT-LLM? I tried engine build but got a rope error? Maybe it is related to the context length? Thanks.
-
您好,我在进行paddle检测的infer时候,
在运行
outs = exe.run(infer_prog,
feed=feeder.feed(data),
fetch_list=values,
ret…
-
(HPT) D:\HPT>python demo/demo.py --image_path demo/einstein.jpg --text 'Question: What is unusual about this image?\nAnswer:' --model hpt-air-demo
usage: demo.py [-h] --image_path IMAGE_PATH [IMAGE…
-
I try to use **trusted-bigdl-llm** to run the Fastchat.
When I build the base image,I got the patching failed error:
```
patching file /usr/local/lib/python3.9/dist-packages/bigdl/llm/utils/utils.p…
-
### Reminder
- [X] I have read the README and searched the existing issues.
### Reproduction
CUDA_VISIBLE_DEVICES=1 llamafactory-cli example/......
below is the yaml file:
# model
model_name_or_…
-
First of all, thank you for great work.
## System info
autoawq==0.1.8
## Details
While I tried to quantize GPT NeoX model, encountered the error below.
```
>>> from awq import AutoAWQForCa…