-
**Describe the bug**
What the bug is, and how to reproduce, better with screenshots(描述bug以及复现过程,最好有截图)
![图片](https://github.com/modelscope/swift/assets/77217949/fbaf7a1f-9270-4c32-bcae-56938fc7c5f4)…
-
Hi, finetuning on flrorence-ft model, the model gets forgetting in old knowledge, (the way we using is not use florence directly, we training it and then adopt the vision encoder to large LLM instead)…
-
release date:
- postpone till further notice, as anti-virus are not happy on things starting 2024-04 b6 of September 1st
- can't continue till eventual clear-up, and removed all downloads associate…
-
### Describe the bug
Versioning diffusers to 0.29.0.dev0 did not solve the problem.
### Reproduction
pip install diffusers
from diffusers import StableDiffusionPipeline
pipe = StableDiffusionPipe…
-
RuntimeError: Failed to import swift.tuners.base because of the following error (look up to see its traceback): No module named 'transformers.models.gemma'
-
`import os
os.environ['CUDA_VISIBLE_DEVICES'] = '3'
from modelscope import Model, AutoModelForSequenceClassification, AutoTokenizer, MsDataset
from swift import Swift, LoRAConfig, AdapterConfig, Tr…
-
**Describe the bug**
What the bug is, and how to reproduce, better with screenshots(描述bug以及复现过程,最好有截图)
将vllm版本从0.3.1升级到0.4.0后,使用swift部署模型,在相同模型、相同prompt的情况下,服务请求时间明显变长(2倍以上),server部署命令参数没有做任何修改
C…
-
模型:Qwen1.5-110B-Chat-AWQ
执行命令
CUDA_VISIBLE_DEVICES=1,6 swift infer --model_type qwen1half-110b-chat-awq --infer_backend vllm --max_model_len 8192 --model_id_or_path /share/models/Qwen1.5-110B-Cha…
-
CUDA_VISIBLE_DEVICES=0 \
NPROC_PER_NODE=1 \
nproc_per_node=1 \
swift infer \
--ckpt_dir "output_llava/llava1d6-mistral-7b-instruct/v32-20240524-165418/checkpoint-2003" \
--custom_val_dataset_path…
-
RuntimeError: Input type (torch.cuda.ByteTensor) and weight type (CUDABFloat16Type) should be the same:
Details is :
Traceback (most recent call last):
File "/datadisk/ai-prj/swift/examples…
hyhzl updated
6 months ago