-
I am trying to run this code
```python
from optimum.quanto import quantize, freeze, qint8
import torch
import torch.nn as nn
class Model(nn.Module):
def __init__(self):
super().__…
-
hello @VainF,
Can you please check the following? Thanks.
deit_small_patch16_224 pruning throws the following error:
```ruby
File "/home/viplab/Deepak/Torch_Pruning_v137/examples/transf…
-
I am trying to lower GPT2 model to linalgIR but I am running into errors. I have built torch_mlir from source and have installed transformers with the latest version: pip install git+https://github.co…
-
```python
try:
import transformers
except ImportError:
pass
from ctranslate2.specs import (
transformer_spec,
)
from ctranslate2.converters.transformers import TransformersConver…
-
I found this project being discussed in local llama subreddit.
I read the paper but had questions.
One of the questions that came up that is gnawing at me... Why Transformer++ as your basis of co…
-
By saving the model and reloading it I managed to get the model working, both with quantized and full precision (it still uses 10gb max of gpu ram).
However, the model generates random characters. He…
-
**Is your feature request related to a problem? Please describe.**
Your Seq2SeqSharp project already support LSTMs. Please consider to implement the RWKV large language "linear attention" idea into y…
-
# URL
- https://arxiv.org/pdf/2002.05202
# Affiliations
- Noam Shazeer, N/A
# Abstract
- Gated Linear Units (arXiv:1612.08083) consist of the component-wise productof two linear projections, one o…
-
self.tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False,trust_remote_code=True)
self.model = AutoModelForCausalLM.from_pretrained(model_path, device_map="auto",torch_dtype=t…
-
### 复现方式
`xtuner train llava_internlm2_chat_20b_clip_vit_large_p14_336_e1_gpu8_pretrain.py`
### 配置文件
仅改动数据集及模型位置
### 运行日志
```
Map (num_proc=32): 100%|████████████████████████████████████████…