-
run on Mac M3 Max 128GB
run this code
```
from transformers import AutoModel, AutoTokenizer
MAX_LENGTH = 128
model = AutoModel.from_pretrained("unsloth/Meta-Llama-3.1-405B-Instruct-bnb-4b…
-
https://github.com/PaddlePaddle/PaddleSlim/tree/develop/example/auto_compression/ocr
该方案更换ICDAR2015数据集,采用预训练ResNet50模型(更改模型配置即可)可以成功运行,其精度基本不变,速度减少为1/4,获得Inference模型。此时的模型在转为ONNX时报错,缺少量化配置文件(cali…
-
Hi, i trained a model using 16/8 configuration (attach the configuration JSON i used) and everything was fine during AIMET optimization. When I try to deploy the model the the DSP I use the commands m…
-
Since my training environment could not connect to the internet, I download the model and dataset and save them in the local disk.
The arguments:
**model path**: ModelArguments(base_model_revision=N…
-
### System Info
```shell
The examples provided do not work correctly, I think there has been updates in the intel neural compressor toolkit, which is now 3.0. and the habana quantization toolkit, and…
-
Please add Qwen2 support
```
EETQ_CAUSAL_LM_MODEL_MAP = {
"llama": LlamaEETQForCausalLM,
"baichuan": BaichuanEETQForCausalLM,
"gemma": GemmaEETQForCausalLM
}
```
-
### Your current environment
The output of `python collect_env.py`
```text
C:\Users\bobni\OneDrive\Desktop\Projects\p2pIssue>bash
training@Training:/mnt/c/Users/bobni/OneDrive/Desktop/Projects…
-
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking argument for argument mat2 in method wrapper_CUDA_bmm)
```python
from aw…
-
# Todo
https://github.com/OCA/maintainer-tools/wiki/Migration-to-version-15.0
# Modules to migrate
- [ ] delivery_package_default_shipping_weight
- [x] delivery_procurement_group_carrier - B…
-
设备为两台linux,每台2张A100 40G显卡:A100(40G) * 2
训练命令如下:主节点命令为CUDA_VISIBLE_DEVICES=0,1 NNODES=2 NODE_RANK=0 NPROC_PER_NODE=2 MASTER_ADDR=127.0.0.1 swift sft --model_type qwen1half-7b-chat --model_id_or_path /…