CUDA_VISIBLE_DEVICES=0 swift infer --model_type got-ocr2 --model_id_or_path stepfun-ai/GOT-OCR2_0
<<< <image>OCR:
Input an image path or URL <<< https://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/ocr.png
简介 SWIFT支持250+LLM和35+MLLM(多模态大模型)的训练、推理、 评测和部署。开发者可以直接将我们的框架应用到自己的Research和 生产环境中,实现模型训练评测到应用的完整链路。我们除支持了 PEFT提供的轻量训练方案外,也提供了一个完整的Adapters库以支持 最新的训练技术,如NEFTune、LoRA+、LLaMA-PRO等,这个适配器 库可以脱离训练脚本直接使用在自己的自定流程中。 为方便不熟悉深度学习的用户使用,我们提供了一个Gradio的web-ui用 于控制训练和推理,并提供了配套的深度学习课程和最佳实践供新手入 门。 此外,我们也在拓展其他模态的能力,目前我们支持了AnimateDiff的 全参数训练和LoRA训练。 SWIFT具有丰富的文档体系,如有使用问题请请查看这里 可以在Huggingfacespace和ModelScope创空间中体验SWIFTweb ui功能了。
--------------------------------------------------
<<< clear
<<< <image>OCR:
Input an image path or URL <<< https://modelscope-open.oss-cn-hangzhou.aliyuncs.com/images/ocr_en.png
Introduction
SWIFT supports training, inference, evaluation and deployment of 250+ LLMs
and 35+ MLLMs (multimodal large models). Developers can directly apply our
framework to their own research and production environments to realize the
complete workflow from model training and evaluation to application. In addition
to supporting the lightweight training solutions provided by PEFT, we also
provide a complete Adapters library to support the latest training techniques
such as NEFTune, LoRA+, LLaMA-PRO, etc. This adapter library can be used
directly in your own custom workflow without our training scripts.
To facilitate use by users unfamiliar with deep learning, we provide a Gradio
web-ui for controlling training and inference, as well as accompanying deep
learning courses and best practices for beginners.
Additionally, we are expanding capabilities for other modalities. Currently, we
support full-parameter training and LoRA training for AnimateDiff.
SWIFT has rich documentations for users, please check here.
SWIFT web-ui is available both on Huggingface space and ModelScope studio,
please feel free to try!
Inference:
fine-tuning:
inference after fine-tuning