-
I'm trying to use the using_t5.py script, but i get the above error.
I'm running on WIn10, python 3.8.
Transformers 4.3.3.
Any idea?
Thanks,
Dorit
-
Thanks for the great work. When I pull the current version and test it, I get the following error. If I fix only the part where the error occurs, the error continues to occur in other parts. I would b…
-
Hello
I am running in the following machine.
CPU: 12th Gen Intel(R) Core(TM) i7-12700
RAM: 32GB, speed: 4400MT/s
NVIDIA RTX A2000 12GB
model is:
llama-2-7b-chat.Q6_K.gguf
And it takes a…
-
## Status-quo
At the moment, there are different ways to interact with the hub. This RFC aims to look at a summary of them, and try to find ways to enhance end users' and developers' experience.
…
-
### Model description
VMamba is a visual foundation model proposed in https://arxiv.org/pdf/2401.10166.pdf.
It is inspired by the recent advances in state stace models and in particular Mamba. T…
dmus updated
9 months ago
-
### Describe the issue
The following problem occurred when I optimized Babelscape/mrebel-large:
warnings.warn(
Some non-default generation parameters are set in the model config. These should go …
-
### Feature request
Hi! I’ve been researching LLM quantization recently ([this paper](https://arxiv.org/abs/2405.14852)), and noticed a potentially improtant issue that arises when using LLMs with 1-…
-
Thank you for your excellent work! The code CoT_with_ChatGPT.py does not provide the knowledge_base.pkl saving code, but only provides the text output content. Can you provide the specific generation …
-
### Issues Policy acknowledgement
- [X] I have read and agree to submit bug reports in accordance with the [issues policy](https://www.github.com/mlflow/mlflow/blob/master/ISSUE_POLICY.md)
### W…
-
### Feature request
Seems that there is no config for DeBERTa v1-2-3 as decoder (while there are configs for BERT/RoBERTa et similia models)... This is needed in order to perform TSDAE unsupervised…