-
My scripts:
```
import torch
from mmpretrain import get_model
import torch_pruning as tp
import torch.nn as nn
from typing import Sequence
from mmpretrain.models.utils import PatchMerging, Wind…
-
****Version****
pyabsa==2.3.1
torch==1.13.0
transformers==4.29.0
****Describe the bug****
I used to load my custom state_dict from my windows system and the loading procedure worked.
However, …
-
我自己也尝试用deeplabv3+架构测试transformer 但是完全不如官方的upernet架构代码。我看你的visdom截图MIOU也只有0.7左右,官方的代码能到0.83左右,而且用你的代码和我自己移植的代码都出现了loss严重波动的问题。请问这个问题是因为什么呢?我搜索到的其他非官方swin tranformer作为encoder的代码结果精度都很低,是什么原因造成的呢
-
> Now that `train.py` supports both training and inference, I think we should close this PR. I think the best path forward is to modify `predict_step` in our trainers to store and stitch together pred…
-
golang.org/x/text/encoding/unicode's documentation is a bit unclear to the newbie like myself that just wants to read files that Windows folks send them without having to become an expert in UTF-16. …
-
**Is your feature request related to a problem? Please describe.**
I'm able to use the `onnxruntime.transformers` codebase to optimize Tranformer-based model using self-attention, however it's not …
-
### Describe the bug
For some reason tranformers is throwing an Exception when loading OpenELM from Apple although they state in their own README to use transformers.
I think the config is missing i…
-
Thank you for great repo! Will be great to have somekind of acceleration for ex. TensorRT.
-
### Feature request
Is there any chance we coukd get this 4bit adam optimizer added to tranformers?
It has nearly the same performance as 32bit adam with significant drop in vram overhead.
[repo…
-
Hi @Qubitium . Since the CPU path is already in gptqmodel, when do you plan to replace auto_gptq to gptqmodel in HuggingFace/optimum? I think we can start an issue in Optimum to let the maintainer kno…