Closed AayushSameerShah closed 1 year ago
Hi, thanks for reporting! @prabod I think this might be related to your feature?
Hie @prabod ! Any update on this? I have found other models too which are causing this issue 😢
Thanks!
@AayushSameerShah I tried to reproduce this issue, but I cannot reproduce it on Colab and locally via the GPU.
You can check this notebook:
and it uses the GPU (just in case, so no issue here)
https://colab.research.google.com/drive/1XyZ6ibezz275QCo9zRlbHeHd7fLqfuWU?usp=sharing
I am afraid we might need the actual issue template
with all the details and required versions so we can try to reproduce it. (you chose a documentation
template which does not really ask for extra information)
issue template: https://github.com/JohnSnowLabs/spark-nlp/issues/new?assignees=maziyarpanahi&labels=question&projects=&template=bug_report.yml
Hie @maziyarpanahi 👋🏻
Thanks for the code and solution... I have tried the notebook that you've provided.
There are some problems still, on the first run the inference can be done, but when we change the parameters like DoSample
it gives the same collectionError
.
I know I should have used the "issue" template, I apologize for this. Here I have created a colab that is linearly runnable and is able to reproduce the error. It is annotated so that should be easy to follow around.
Link: https://colab.research.google.com/drive/1xsYGjHcPxnh4e5UZxYj-hLT7BCFYP_0H?usp=sharing
Please let me know if it doesn't work. Thank you so much! 🤗
@maziyarpanahi Hope you were able to catch that error 🤗
Hi @AayushSameerShah yes, your notebook was very helpful and we are debugging it currently.
Hello!
Thank you for addressing this issue... but I am unsure if I am able to run this properly.
For example, I am still using the same colab as given above in the link. Trying to download the bart_cnn
model but still shows the error.
Thanks.
This code and model that was previously failing now works in 5.0.2:
A few things to keep in mind:
temperature
and the k
is too low it couldn't find anything. So we fixed this by falling back on more deterministic approachHie, I am not really sure if this is regarding because of the "large model". Because this time I tried to use the https://sparknlp.org/2023/01/30/t5_efficient_large_dl2_en.html model which is around 700MB and... I suppose it is sufficient for the 12GB RAM instance.
In the image above, I have tried to download the model first and then trying to load it... expecting it as a workaround...
😕
Hi @AayushSameerShah
The initial bug report was for BART annotator, this seems to be T5. (I also see Java side is empty
which means you have issue with either starting your SparkSession or it was started and it was killed due to OOM)
In any case, if you have a notebook that can be reproduced (without loading an offline model), please create a new issue regarding T5 annotator.
Yeah, true the original issue was regarding the BART model... but the general error seems to be the same. In both cases, the same error was being thrown. Sure, I will open the issue for the T5 model with reproducible code.
Thanks.
Hello, Sorry for posting this under the "documentation" label, but I thought it is more appropriate than the "bug". The problem I am facing is currently of two types.
1️⃣ Model is not downloading at all 2️⃣ Model can be downloaded but cannot do the inference.
Let me brief you about them.
Before going through any of the problems, I would provide some code which I use as a starter to drive the context:
Now, let me show the error-prone code.
1️⃣ Model is not downloading at all
📝 The model page here: Page
Model Name:
bart_large_cnn
Throws this error:
(error in short)
(error in long)
2️⃣ Model is downloaded but can't run inference
I have seen this problem in these two models. 📝 Model-1 page here: Page-1 📝 Model-2 page here: Page-2
The code:
After the successful download I run the inference like below:
And it gives this error:
It's a long error... but it will give the context. Seems like a problem in
.collect()
and also I tried using the.select("summaries").first()
and still gives some other related error.Thank you 🙏🏻