-
- [ ] `TransformerDecoder.forward`: where does `self.training` come from?
https://github.com/asyml/texar-pytorch/blob/d17d502b50da1d95cb70435ed21c6603370ce76d/texar/torch/modules/decoders/transforme…
-
working fine previously however recently, Im getting no ouputs with message -> Cannot close object, library is destroyed. This may cause a memory leak!, im using a colab notebook
-
Hi, congratulations with this nice model
I am wondering if it is possible to run the model using CPU only and if this has been tested?
-
Model generates only garbage.
Sample: https://github.com/aws-neuron/aws-neuron-samples/blob/master/torch-neuronx/transformers-neuronx/inference/llama-3-8b-32k-sampling.ipynb
NeuronSDK2.19 PyTorc…
-
### Describe the bug
passing Different Height or Width other than 1024 by 1024 leads to an error
### Reproduction
```
from diffusers import AuraFlowPipeline
import torch
pipeline = AuraFlowPip…
-
PEER looks like an interesting approach and thanks for implementing so cleanly! I do have a quick question though about recommended usage with x-transformers. Would something like this be a good way o…
-
**Description**
Please consider adding Core ML model package format support to utilize Apple Silicone Nural Engine + GPU.
**Success Criteria**
Utilize both ANE & GPU, not just GPU on Apple Sili…
-
I want to use pre-trained model from PyTorch to train a faster-rcnn. And I see:
> if you want to use pytorch pre-trained models, please remember to transpose images from BGR to RGB, and also use the…
-
Hello. I wanted to use gritlm to a open-source embedding model —— gte-qwen2-7b-instruct, but I encountered some problems:
```
[rank1]: Traceback (most recent call last):
[rank1]: File "/code/xx…
-
## 🚀 Feature
I propose adding instead of batch size a dictionary with batch size per GPU, for example {"cuda0":4, "cuda1", 6}
### Motivation
I have gv100 (32gb) and 3090 (24gb). using the cur…