-
Hi authors,
Thank you for the nice works and release of the datasets. Could you add corresponding licenses to the released datasets, namely https://huggingface.co/datasets/princeton-nlp/llama3-ultraf…
-
Hello! Not an issue but rather a question regarding features; are there plans to extend this to NLP datasets? :) Thanks!!
-
For the Instruct setup, why do different models require different training datasets? Can the same dataset be used?
-
**Is your feature request related to a problem? Please describe.**
When using the `automl.text_classification()` method to train an NLP task, I would like to track the log_loss over time to check how…
-
python3.8和python3.9 win11x64都试了全报错,
pip install paddlepaddle-gpu==2.5.1
pip install paddlehub==2.4.0
````pythonimport paddlehub as hub
import cv2
import sys
import os
if __name__ == '__main…
-
Hi Everyone,
I'm trying to run model pre-training using a large dataset +150GB.
And looked around for any reference to do that using the APIs of the library but found nothing saddly.
Any ide…
-
Would be great to have the datasets in JMTEB (https://github.com/sbintuitions/JMTEB) integrated into MTEB for those which aren't yet already, so we can also add a Japanese leaderboard sometime 😊 cc @l…
-
Thank you for providing these excellent datasets. I am currently using the "Vitamin and Supplements", "Beyaz Perde All Movies", and "Beyaz Perde Best Movies" datasets from this repository for a sentim…
-
Hi, thanks a lot for the great work!
It seems that the data module is missing, and there is module import related to this.
https://github.com/bowang-lab/scGPT/blob/bc8939504fc62dd617360618a840ca9b67…
-
Hi, thanks for your work.
I'm trying to test out the result of your work but found some difficulties on reproducing similar accuracy results.
Below is the Environment that I created:
channels:
…