-
llama_model_loader: loaded meta data with 32 key-value pairs and 219 tensors from /data/huggingface/hub/models--city96--t5-v1_1-xxl-encoder-gguf/snapshots/005a6ea51a7d0b84d677b3e633bb52a8c85a83d9/./t5…
-
Hello, thank you for providing the community with a great framework. I did some experiments with the T5-small model on the C4 Vietnamese dataset, and I would like to get some feedback from you:
1. …
-
cc @mattdangerw
-
Great project! Thanks for the release.
In your paper you mentioned to employ T5 model to extract language embedding for K-means clustering. I noticed this repo seems not to contain the code of this…
-
Hi,
I am trying to use this framework on causal models such as llama based models and other LLMs. For my case, I use Tinyllama and Pythia to replace the T5 model in the original pipeline (TinyLlam…
-
Hi,
Can you please provide the support for T5 model inference. I see that only decoder models are supported https://github.com/ray-project/aviary/tree/master/models/static_batching
Thanks
-
Whu it changes pytorch version and installs different cuda on the system?
This would break most peoples's environments actually, because there can be only one cuda version on the Ubuntu, and it has…
-
### Feature request
Currently TGI NeuronX loads the artifacts with NeuronModelForCausalLM class, which gives error when loading Flan T5.
```
Unrecognized configuration class for this kind of Aut…
-
i am stuck at this errors:
```.bash
(deforum_xflux_env) PS D:\_AI\ComfyUI_windows_portable\ComfyUI\custom_nodes\deforum-x-flux> python.exe run.py
Loading checkpoint shards: 100%|█████████████████…
-
### System Info
Hi Team,
First of all huge thanks for all the great work you are doing.
Recently, I was benchmarking inference for T5 model on AWS EC2 ( G6E machine with L40 GPU) for batch sizes…